Inspiration

Wanted to experiment with taking notes using speech to text software. Some times its easier to understand something when you can hear the person's tone when they are saying it.

What it does

Allows users to record and upload audio clips which are then transcribed, saved, and made text-searchable. User's can also play back the audio that they find.

How I built it

Wrote several cloud functions to handle things such as file format conversion, speech to text transcription, google cloud storage, and synchronization between Algolia database, Firestore database, and google cloud storage.

The audio files are saved to google cloud storage, and the transcription data + audio file location within google cloud storage are saved in Firestore. Algolia contains the username, the text transcription, and the fire object ID that links it to the firestore object. This way, whenever we update or delete a Firestore audio clip, it can be deleted from both the google cloud storage, as well as Algolia.

Algolia was used to create the fully text searchable database.

Google auth ensures that the app is secure and the APIs can only be accessed by authorized users.

Challenges I ran into

I spent a long time trying to convert audio files to .wav from the browser, before I decided to just make a cloud function to handle the conversion.

What's next for sound-bite

We would like to add additional natural language processing features so that we can sort sound bites by sentiment and show the positivity/negativity associated with certain key-words.

We would also like to add the use of a Maps API so that users can associate certain sound bites with certain geographic landmarks such that those sound bites will only be visible if you are nearby that location.

Built With

Share this project:

Updates