diff --git a/content/en/project/AD-Audio-Classifier-Regression/AcousticVsSemantic.png b/content/en/project/AD-Audio-Classifier-Regression/AcousticVsSemantic.png new file mode 100644 index 00000000..49e53560 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/AcousticVsSemantic.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/AudioPreproc.png b/content/en/project/AD-Audio-Classifier-Regression/AudioPreproc.png new file mode 100644 index 00000000..1d70b197 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/AudioPreproc.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/ClassifierResults.png b/content/en/project/AD-Audio-Classifier-Regression/ClassifierResults.png new file mode 100644 index 00000000..9a5a4445 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/ClassifierResults.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/CookeTheftEnhance.png b/content/en/project/AD-Audio-Classifier-Regression/CookeTheftEnhance.png new file mode 100644 index 00000000..5c41e745 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/CookeTheftEnhance.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/Interspeech2021.png b/content/en/project/AD-Audio-Classifier-Regression/Interspeech2021.png new file mode 100644 index 00000000..2858c2b0 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/Interspeech2021.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/R2Results.png b/content/en/project/AD-Audio-Classifier-Regression/R2Results.png new file mode 100644 index 00000000..0afaba10 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/R2Results.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/VoiceHackGag.png b/content/en/project/AD-Audio-Classifier-Regression/VoiceHackGag.png new file mode 100644 index 00000000..fa18e109 Binary files /dev/null and b/content/en/project/AD-Audio-Classifier-Regression/VoiceHackGag.png differ diff --git a/content/en/project/AD-Audio-Classifier-Regression/index.md b/content/en/project/AD-Audio-Classifier-Regression/index.md new file mode 100644 index 00000000..9ab8b0b9 --- /dev/null +++ b/content/en/project/AD-Audio-Classifier-Regression/index.md @@ -0,0 +1,96 @@ +--- +type: "project" # DON'T TOUCH THIS ! :) +date: "2024-06-14" # Date you first upload your project. +# Title of your project (we like creative title) +title: "Detection of Alzheimer's through Acoustic and Semantic Markers" + +# List the names of the collaborators within the [ ]. If alone, simple put your name within [] +names: [Matías Caccia, Jeremías Inchauspe, Paloma Georgopulos] + +# Your project GitHub repository URL +github_repo: https://github.com/MatiasCaccia/BrainHack2024 + +# If you are working on a project that has website, indicate the full url including "https://" below or leave it empty. +website: https://udesa.edu.ar/cnc + +# List +- 4 keywords that best describe your project within []. Note that the project summary also involves a number of key words. Those are listed on top of the [github repository](https://github.com/brainhack-school2020/project_template), click `manage topics`. +# Please only lowercase letters +tags: [alzheimer, speech-analysis, acoustic-markers, semantic-markers] + +# Summarize your project in < ~75 words. This description will appear at the top of your page and on the list page with other projects.. +summary: "Our project aims to identify acoustic and semantic markers from the speech of Alzheimer's patients to detect the disease and estimate MMSE scores using machine learning models. This approach offers a scalable and cost-effective method for early diagnosis." + +# If you want to add a cover image (listpage and image in the right), add it to your directory and indicate the name +# below with the extension. +image: "AcousticVsSemantic.png" +--- + + +## Project definition + +### Background + +Inspired by the need for scalable and cost-effective diagnostic methods for Alzheimer's disease (AD), our project at BrainHack 2024 focuses on identifying acoustic and semantic markers in the speech of patients. Alzheimer's disease is characterized by a gradual decline in cognitive functions, and early detection is crucial for managing the disease. Speech analysis offers a non-invasive and reliable data source for this purpose. + +### Objectives + +1. **Extract acoustic and semantic markers from patient speech** to detect the presence of Alzheimer's. +2. **Estimate MMSE scores** based on voice characteristics and semantic indicators from conversations. +3. **Corroborate cognitive decline** using the most relevant linguistic features. + +### Tools + +The project utilizes several tools and technologies: + * **Speech pre-processing**: Noise reduction, volume normalization. + * **Feature extraction**: Acoustic features like pitch, formants, loudness, and MFCC; Semantic analysis using Sentence-BERT and KeyBERT. + * **Machine learning models**: XGBoost, Random Forest, SVM, and Logistic Regression for classification and regression tasks. + +### Data + +The dataset used is the **Pitt Corpus from DementiaBank**, employed in the ADReSS Challenge at Interspeech 2021. It includes: +- 156 low-quality interview audios. +- Transcriptions in CHAT format. +- Metadata such as age, gender, condition, and MMSE scores. +- Audios are from interviews of patients describig the "Cookie Theft Picture" + +image: "CookeTheftEnhance.png" +### Deliverables + +At the end of this project, we will have: + - A detailed report on the methodology and findings. + - Visualizations of data and results, including bar graphs, scatter plots, and network graphs. + - Code and documentation for reproducibility, available on GitHub. + +## Results + +### Progress overview + +Our project demonstrated that acoustic features are critical for classifying Alzheimer's, with Logistic Regression achieving 79% accuracy. The regression of MMSE scores was more challenging due to limited data, with the best model being SVM. + +### Tools I learned during this project + + * **Speech processing techniques** for noise reduction and feature extraction. + * **Machine learning models** for both classification and regression tasks. + * **Semantic analysis** using advanced NLP models like Sentence-BERT and KeyBERT. + +### Results + +#### Classification + +image: "ClassifierResults.png" + +- **Best model**: Logistic Regression with 79% accuracy. +- **Key features**: Duration and number of pauses, MFCC, and spectral flux. + +#### Regression + +image: "R2Results.png" + +- **Best model**: SVM with R² of 0.32, MSE of 25.76, MAE of 3.82, and RMSE of 5.07. +- **Key features**: Proportion of adverbs and nouns, text similarity score. + +## Conclusion and Acknowledgement + +This study highlights the potential of speech analysis for early detection and cognitive assessment in Alzheimer's patients. The approach using acoustic and semantic features combined with machine learning models shows promising results. We acknowledge the contributions of our team and the support from BrainHack 2024. + +