You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+16
Original file line number
Diff line number
Diff line change
@@ -17,3 +17,19 @@ Reconnect first allows the user to listen to a sound file that contains a senten
17
17
Reconnect uses Microsoft Azure’s Speech-to-Text function to convert the user’s speech input into text. By comparing the text against the sentence provided to the user, Reconnect is able to determine if the user’s pronunciation is adequately correct. Following which, Azure’s Text-to-Speech function is used to generate a separate speech output from the same sentence. These two .wav files will then be processed by Reconnect.
18
18
Reconnect uses the SciPy library to convert the sound files into audio data chunks. By using our self-developed algorithms to process the audio data’s amplitude, frequency, and breaks, Reconnect is able to determine the relative speed of vowel enunciation, and the presence of unnaturally long or short breaks between words and sentences.
19
19
Finally, Reconnect will compile all of these feedback before presenting them to the user. The user will then be given the opportunity to try again. The user can also type a sentence which he or she hopes to practice, and Reconnect will generate a sound file to facilitate the same learning process as mentioned above.
20
+
21
+
# Challenges we ran into
22
+
Since the team comprised of a sophomore and two freshmen having a less technical background, we ran into a lot of difficulties. This was the first time we ever played with APIs and it was difficult to get things working together. In the beginning, we did not go think about the number of channels of the input. Also, for the text comparison, it was necessary to mind the length of the expected text and the received text. While the typed text had to be preprocessed so that it did not contain any characters, the expected text had to be preprocessed so that it omitted some unintended words like “oh”, “umm” etc. Since none of us had enough experience in web development, the significant challenge was getting the input from the user in the form of audio
23
+
- Microphone input
24
+
- Two-channeled audio files and wave comparison algorithms for them
25
+
26
+
# Accomplishments that we're proud of
27
+
Despite all the challenges, we are proud that we successfully built an interactive platform Reconnect where people can practice speaking to reconnect with the world. Helping thousands of people worldwide in transitioning from impaired hearing to speaking effectively is indeed a great satisfaction to our team.
28
+
29
+
# What we learned
30
+
Being new to the hackathons, initially, we were unsure if we should go forward with this idea due to technical complexities. It was a second hackathon for all of us and our first-ever time using any sort of APIs. However, we decided to take up the challenge and finally it worked. Therefore, in addition to learning more about programming, using APIs and developing web-site, we learned to think big and apply the knowledge to have an impact on people’s life
31
+
32
+
# What's next for Reconnect
33
+
- consulting with medical professionals to get effective strategies for speech reconstruction
34
+
- expanding to different languages
35
+
- turning Reconnect into an actual learning platform with the ability to track progress and try different strategies
0 commit comments