Project

# Title Team Members TA Documents Sponsor
70 Podcast Analyzer with AI for Ringer (Sponsored)
Bhuvaneshkumar Radjendirane
Elliot Couvignou
Sai Rajesh
Shuai Tang design_document1.pdf
other1.pdf
proposal1.pdf
his project pitch was given by Prof. Patel on the 2/4 lecture. We have met with Prof. Patel about the details for this project and decided on the following.

Names: Elliot Couvignou (esc2), Sai Rajesh (srajesh2), Bhuvan Radj (br3)

Problem:

For context, Ringr is an app that allows for high-quality recording of multiple clients regardless of their location and compiles the audio into one that sounds like they are in the same room. Their current problem is that analyzing the content of each recorded podcast and picking out the 'good parts' is obviously too time consuming to be done manually so an AI is needed. With this AI they would like to feed in any raw recordings and from that, get an audio recording of only the desired parts. They do want this feature eventually implemented on their app with a reasonable run-time. What is defined as ‘good’ parts is still left up for us to finalize when we meet with Ringr for the first time later this week.

Solution Overview:

Because this is a software only project we will go over the general flow of the AI from input to output and break that down into components. The overall idea of the AI is to break down the audio input into small pieces of audio and then reconstruct phonemes into words and those into word meaning, etc. up until the AI is aware about the recording's semantics and can pick apart the sections that the user wants. At the moment we plan to create the AI model(s) and save them so we don't need to recreate a new model on each input analysis. This should greatly increase performance to be usable as an app feature.

Solution Components: Listed in order from closest to Input (top) to farthest (bot)

Word Recognition: From the input we need to recognize words by breaking down audio into phonemes and discern the types through formant filters. From there we can combine our phonemes together to form words. Some extra touch up is needed here to make sure phoneme words are spelled correctly. From this we should now have a transcript of the input audio which is much easier to work with.

Semantic Recognition: We need to understand what the people are saying in order to know what's 'good'. This is a section that is covered by ECE 448 so we hope to use similar methods here such as word2vec or n-gram for HMM. This is probably the most challenging and intricate part as this is a rough topic to get right in AI.

’Good Part’ Recognition: Now that we know what the audio is talking about we can we look for portions of audio that closely resemble that of what the user wants and use those pieces in our resulting data. We compile each good slice into one and return this as our result.

The type of AI model we hope to use is some combination of RNNs and classifiers as we found this to be the most successful models in recent years. We want this model saved on the app but are aware that the size might grow too large from intelligence and require remote computing (Google does this).

Criterion For Success:

Our model works if it can correctly slice out unwanted segments and keep relevant segments of the recording data. This feature should be able to run on both mobile and desktop in a reasonable amount of time given the input length. If our model ends up using any libraries/API’s, then it still needs to be economical to use (i.e no paid services/licenses).

The Marching Band Assistant

Wynter Chen, Alyssa Louise Licudine, Prashant Shankar

The Marching Band Assistant

Featured Project

NetID/Names

wynterc2 (Wynter Chen), alyssal3 (Alyssa Licudine), shankar7 (Prashant Shankar)

Problem

Drum majors lead and conduct marching bands. One of their main jobs is to maintain tempo for the musicians by moving their hands in specific patterns. However, many drum majors, especially high school students, need to learn how to conduct specific tempos off the top of their head and maintain a consistent tempo without assistance for performances. Even those with musical experience have difficulty knowing for certain what tempo they're conducting without a metronome.

Solution Overview

Our project consists of an arm attachment that aids drum major conducting. The attachment contains an accelerometer that helps determine the tempo in beats per minute via hand movement. A display shows the beats per minute, which allows the drum major to adjust their speed as necessary in real time. The microcontroller data is wirelessly transmitted, and a program can be downloaded that not only visualizes the data in real-time, but provides an option to save recorded data for later. There is also a convenient charging port for the device.

This project is a unique invention that aims to help marching bands. There have been previous projects and inventions that have also digitized the conducting experience, such as the Digital Conducting Baton from Spring 2015. However, these have been in the form of a baton rather than a glove, and are used to alter music files as opposed to providing feedback. Additionally, orchestra conductors use very delicate motions with a baton, while drum majors create large, sharper motions with their arms; thus, we believed that an arm attachment was better suited for marching band usage. Unlike other applications that only integrate digital instruments, this project seeks to assist live performers.

Link to RFA: https://courses.grainger.illinois.edu/ece445/pace/view-topic.asp?id=37939

Project Videos