- Basics of Machine Learning
- The Learning Problems
- Feasibility of Learning
- Experiments with Perceptron Learning Algorithm
- Theory of Generalization & Decision Stumps
- Linear Models
- Beyond Gradient Descent
- Experiments with Decision Stumps
- Linear Models and More
- Playing with Regularization
- Virtual Examples and Regularization
- Experiments with Linear and Nonlinear Models
- More about Regularization
- Validation
- Support Vector Machine
- Experiments with Regularized Logistic Regression
- Support Vector Machines
- Bagging and Boosting
- Experiments with Soft-Margin SVM and AdaBoost
The project explores different machine learning approaches for predicting the danceability of music tracks. The focus is on evaluating the performance of multiple models by preprocessing the data and experimenting with various techniques. The final goal is to achieve a balance between model complexity and accuracy.
- Dealing with Missing Values:
- Fill missing values with the median to avoid outliers’ influence.
- Text Features (Artist, Composer, etc.):
- Calculated mean danceability for each artist/composer. Sparse data were treated as missing values.
- Standardization:
- Numerical features were standardized to balance their influence and enhance model performance.
- Correlation Analysis: Examined correlation between features like energy, liveness, and danceability.
- Experimentation: Selected features based on performance, testing combinations of numerical values, artist, composer, and other attributes.
- Linear Regression:
- Simple and efficient, performed well after feature selection.
- Ridge Regression:
- Applied regularization, but improvement was minimal.
- Support Vector Regression (SVR):
- Significant improvement with the RBF kernel and fine-tuned regularization parameters.
- Random Forest:
- Tended to overfit; required careful tuning of depth and sample leaf size.
- Neural Networks:
- Best performance overall, capturing complex relationships in the data.
- Cross-validation and AdaBoost were applied, but with limited success in reducing the gap between training and test errors.
- The Neural Network Model outperformed others with an
Eout = 1.84
, proving the ability to model nonlinear relationships more effectively.
- Blending: Combined predictions from multiple models, leading to the optimal result of
Eout = 1.83
.
Preprocessing text features and blending models were critical for improving performance. While simpler models like Linear Regression provided reasonable results, the Neural Network was the best at capturing complex patterns.