Kaggle Competition Submission
π
Ranked #44 on the Public Leaderboard
This repository contains my solution for the Chatbot Arena Preference Prediction Kaggle competition.
The challenge was to predict which chatbot response a human judge preferred, given a prompt and two responses.
The submission ranked 44th on the public leaderboard out of hundreds of global teams π.
π notebooks/ # Exploratory data analysis and experiments
π README.md # Project overview (this file)- β¨ Fine-tuned transformer-based models on human preference data.
- π Analyzed semantic similarities between responses.
- βοΈ Normalized token lengths and prompt alignment.
- π Applied ensemble strategies to combine model strengths.
- π§ͺ Used stratified validation to handle subjectivity and avoid leakage.
| Tool | Purpose |
|---|---|
| π Python | Programming Language |
| π€ Transformers | Pretrained NLP Models |
| π₯ PyTorch | Deep Learning Framework |
| π Scikit-learn | Metrics & Utilities |
| π Pandas | Data Manipulation |
| π Matplotlib | Visualization |
| Metric | Value |
|---|---|
| Leaderboard Rank | π₯ #44 |
| Final Score | [Insert final score] |
| Total Teams | [Insert number] |
- Human preferences in NLP are highly nuanced and often subjective.
- Even small model tweaks (like input formatting or length balancing) had large effects on performance.
- Ensembling and careful validation strategy were critical to climb the leaderboard.
Huge thanks to:
- The Kaggle community for insightful discussions and open-source notebooks.
- Competition organizers for an exciting and innovative challenge.
- Open-source contributors to libraries like Hugging Face & PyTorch.
If you're interested in discussing the project or collaborating, please reach out at bhandeystruck@gmail.com