[P]Building an End-to-End Music Genre Classifier: My first deep dive into Audio Processing and ML.

Building an End-to-End Music Genre Classifier: My first deep dive into Audio Processing and ML.

Hi everyone, ​I’m a 2nd-year Electrical and Electronics Engineering student, and I just finished my first end-to-end project in the intersection of Audio Processing and Machine Learning. ​As someone who is passionate about metal music and embedded systems, I wanted to understand how machines “hear” and categorize different genres. I built a Music Genre Classifier using Python, and it was a great learning experience in what some people call “Vibe Coding”—using LLMs to prototype rapidly while focusing on the underlying engineering logic. ​What I did: ​Data Processing: Used Librosa for feature extraction (MFCCs, Spectrograms, and Mel-scale). ​The Model: Built a classification model (CNN/SVM) to recognize various genres. ​The Workflow: I used AI as a collaborative partner to handle boilerplate code and debugging, which allowed me to focus on the signal processing theory (Fourier Transforms, etc.). ​I’m looking for feedback on: ​Code Architecture: How can I make my Python scripts more modular for future embedded integration? ​Optimization: Are there more efficient ways to handle real-time audio features? ​General Advice: As an EEE student aiming for a master’s in AI/Robotics, what should be my next step to level up this project? ​GitHub Repository: https://github.com/Baturalpbyg/music-genre-classification

submitted by /u/dreamcull
[link] [comments]

Liked Liked