Telstra - Prediction of Severity of Faults on Telecommunication Network
02 Mar 2016Telstra is Australia’s largest telecommunications, media, and network technology company, offering a full range of communications services. They conducted their first-ever “recruitment” competition in Kaggle to predict the severity of service disruptions on their network - whether it is a momentary glitch or a total interruption of connectivity. The algorithm and model developed for this competition will help Telstra to predict the severity of service disruptions and provide better service to its customers.
The competition ran from 25-Nov-2015 to 29-Feb-2016 and there were 974 individuals who participated across the globe.
The dataset was comprised of multiple files each containing different features extracted from Log files, collected from various locations at different times. The target feature, Fault severity with 3 categories (0: No Fault, 1: Few Faults, and 2: Many Faults) was the one that has to be predicted from the given datasets. Again, the prediction should be the probability of each severity type (multi-class) for the given test dataset.
In this competition, Multi-class logarithmic loss was used to evaluate the performance of predicted multi-class probabilities.
I generated the following set of models to submit them individually and then to ensemble:
- Generalized Boosted Regression Model (GBM) with Out-of-Bag (OOB) estimator,
repeatedcv
, and 5 separate 10-fold cross-validations
- Random Forest Model with Out-of-Bag (OOB) estimator,
repeatedcv
, and 8 separate 10-fold cross-validations
- Stacking (Meta-Ensembling) of Random Forest and GBM Models
After creating the above models, I took the arithmetic mean of all the predictions. Then I ensembled by applying different weightage to each model’s predictions for submission.
Refer below for first few rows of submission file:
id | predict_0 | predict_1 | predict_2 |
---|---|---|---|
2 | 0.991082892 | 0.007185119 | 0.001731983 |
3 | 0.276082613 | 0.609547675 | 0.114369705 |
4 | 0.951721534 | 0.044502856 | 0.003775634 |
7 | 0.235435281 | 0.231292959 | 0.533271745 |
9 | 0.071940752 | 0.187302526 | 0.74075672 |
10 | 0.839680135 | 0.126305137 | 0.034014717 |
11 | 0.564171374 | 0.428022489 | 0.007806152 |
12 | 0.882261053 | 0.110307094 | 0.007431867 |
14 | 0.457473055 | 0.532052293 | 0.010474668 |
15 | 0.78682138 | 0.207976922 | 0.005201698 |