Kaggle Winners Walkthroughs: Child Mind Institute —Problematic Internet Use with Team Lennart Haupts
From Kaggle
The presentation outlines the solution developed by Nar House for a Kaggle competition on problematic internet use, emphasizing his background in economics and machine learning. Key points include his strategy of predicting underlying scores for better accuracy, using an ensemble of models like LightGBM and XGBoost, as well as highlighting significant features such as age, daily internet use, and sleep disturbance scores.
Key Takeaways
- Winning isn't just about predicting; it's about optimizing your strategy—like predicting scores over labels.
- In machine learning, the noise is sometimes louder than the signal; PCA to the rescue, but with caution!
- Feature selection is more art than science; eyeballing and manual selection can yield gold alongside the metrics.
- Overfitting to leaderboards? Always a risk. Stepping back can foster resilience and better results in competitions.
- Sometimes your model's choices reveal surprising truths—who knew sleep disturbances would be such a pivotal feature?