![Mel-ResNet figure](https://private-user-images.githubusercontent.com/94499717/345984268-e2e3a490-7a09-4014-8d44-653de56840b9.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkwOTc5OTYsIm5iZiI6MTczOTA5NzY5NiwicGF0aCI6Ii85NDQ5OTcxNy8zNDU5ODQyNjgtZTJlM2E0OTAtN2EwOS00MDE0LThkNDQtNjUzZGU1Njg0MGI5LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA5VDEwNDEzNlomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWZjYzg3NjU5ODNiYTdlNTBlMGFjMDE3Y2U1MmFmY2IzMWNkM2E0OWUyZjU3NDc1NGZmZGU4ZWU0NDNiM2RjYzImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.k2Pwp51XSLjjCwRjblg8iqfoR9JgBhv5x8VF7bwC0vQ)
- We developed a generative model that inputs EEG data during Inner Speech (Imagined Speech) and the corresponding mel spectrogram of Spoken Speech (target) into a GAN.
- Through this approach, the mel spectrogram generated from the EEG is fed into a ResNet trained on actual speech mel spectrograms to predict the imagined word.
`Python >= 3.7`
All the codes are written in Python 3.7.
You can install the libraries used in our project by running the following command:
pip install -r requirements.txt
We extracted word utterance recordings for a total of 13 classes using the voices of 5 contributors and TTS technology.
Additionally, to address any data scarcity issues, we applied augmentation techniques such as time stretching, pitch shifting, and adding noise.
- Call
- Camera
- Down
- Left
- Message
- Music
- Off
- On
- Receive
- Right
- Turn
- Up
- Volume
ResNet-50
![](https://private-user-images.githubusercontent.com/94499717/355336270-d7e55c25-702e-48f5-8ecb-00b02cf92b85.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkwOTc5OTYsIm5iZiI6MTczOTA5NzY5NiwicGF0aCI6Ii85NDQ5OTcxNy8zNTUzMzYyNzAtZDdlNTVjMjUtNzAyZS00OGY1LThlY2ItMDBiMDJjZjkyYjg1LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA5VDEwNDEzNlomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTg5ZmI4YzRiNDZiNDg1MDZjMGI5NjFlZjZjNjY4ZGM1Njk4OWZhYjUyYTcxNzRlODkyMmVhOTVhYTU3YTAzYzImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.w9fNOVQ_gbD1svvTVSKJ0OwNd86K7DbdDxUmnSV1CmU)
(We are continuously collecting data and will need to undergo a hyperparameter tuning process after training.)
Performance metrics : Accuracy, F1 score
(learning curve and metrics will be here)