Neural encoding of acoustic and semantic features during speech and music perception: Matlab to Python code translation
Bruno Giordano (INT) : https://github.com/brungio/bhack_td https://twitter.com/brungio https://framateam.org/blgnatsou/messages/@bruno.giordano
Giorgio Marinato (INT) : https://github.com/neurogima https://twitter.com/neurogima mattermost: @neurogima
Benjamin Morillon (CR INS - https://github.com/DCP-INS) Nadège Marin (IE, INS) Arnaud Zalta (PhD student, INS)
Brainhack Marseille
In everyday life, humans are particularly attuned to listening to two particular types of sound: speech and music. We apply a novel analysis method to shed light on how the brain is almost effortlessly able to use acoustic features to assign meaning to sounds. To do so, we use an original cross-validated Representational Similarity Analyses (RSA) approach implemented in Matlab to estimate the similarity between acoustic or semantic features of an auditory stream (speech, music) and neural activity (here intracranial EEG recordings decomposed into frequency bands).
https://github.com/brungio/bhack_td
The main goal of this project is to translate the Matlab code into Python:
python coding Representational Similarity Analyses (RSA) Cross-validation (train-test-validate) intracranial EEG signal processing sharing data analyses ideas
to analyse intracranial EEG data to perform cross-validation procedures to estimate the neural encoding of different acoustic features
We will provide a sample of a dataset of iEEG recordings.
from 3 to 10
Acknowledgment in the code and in the Github repo.
Leave this text if you don’t have an image yet.
pipeline_development
1_basic structure
neural_decoding
MNE, RSAtoolbox (github.com/rsagroup/rsatoolbox), Scikit Learn
Python, Matlab
intracranial EEG
Basic git workflow: fork, branches, commit and pull request