Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
Categories
Tags
We have previously added a set of company identity-agnostic predictors, such as the number of drivers a company employs, or the number of vehicles in the fleet with a hydraulic lift, and so on. we took this approach, rather than having each company as a unique predictor, so that the addition of a new contractor would not (necessarily) confuse our model.
I understand that this means company history, specific vehicle history, and so on, are not included. Nevertheless, we want to see how much information is in the new predictors.
Data exploration
in_csv <- "../output/intermediate/ii_times.csv" ii_times <- read_csv(in_csv) %>% filter(reported_before_resolved == 1) %>% #we kept these flags for exploration, but they are troublesome for our real prediction. If we know a case has already been resolved, our model looks better than it is at predicting delays. mutate(reported_before_resolved = NULL) #first we look at the correlation of the predictors with the time_delayed: corrr_analysis <- ii_times %>% select(-Busbreakdown_ID) %>% correlate() %>% focus(time_delayed) %>% rename(feature = rowname) %>% arrange(desc(abs(time_delayed))) %>% mutate(feature = as_factor(feature)) corrr_analysis %>% print(n=61) ## # A tibble: 45 x 2 ## feature time_delayed ## <fct> <dbl> ## 1 Boro_Bronx -0.166 ## 2 vehicle_total_max_riders -0.159 ## 3 vehicle_total_with_attendants 0.159 ## 4 vehicle_total_reg_seats -0.157 ## 5 drivers_total_attendant 0.147 ## 6 Has_Contractor_Notified_Parents 0.140 ## 7 service_type_d2d 0.140 ## 8 Number_Of_Students_On_The_Bus -0.139 ## 9 Boro_Manhattan 0.125 ## 10 Boro_StatenIsland -0.106 ## 11 Reason_HeavyTraffic -0.106 ## 12 vehicle_total_with_lifts 0.106 ## 13 Reason_MechanicalProblem 0.100 ## 14 vehicle_total_ambulatory_seats -0.0960 ## 15 drivers_staff_servSchool 0.0716 ## 16 rush_min_from_peak 0.0648 ## 17 Reason_FlatTire 0.0582 ## 18 Boro_Queens 0.0550 ## 19 Reason_Accident 0.0503 ## 20 drivers_num_servPreK -0.0478 ## 21 Reason_DelayedbySchool -0.0471 ## 22 drivers_staff_servPreK -0.0468 ## 23 School_Age 0.0441 ## 24 drivers_numServ_prek -0.0425 ## 25 Boro_Westchester 0.0400 ## 26 drivers_numServ_school 0.0390 ## 27 rush_within -0.0385 ## 28 Reason_WontStart 0.0383 ## 29 Boro_Brooklyn 0.0378 ## 30 Reason_WeatherConditions 0.0367 ## 31 Boro_NassauCounty 0.0345 ## 32 Has_Contractor_Notified_Schools 0.0334 ## 33 Boro_NewJersey 0.0265 ## 34 time_am -0.0257 ## 35 rush_between 0.0254 ## 36 Active_Vehicles -0.0164 ## 37 rush_outside 0.0152 ## 38 Reason_ProblemRun 0.0115 ## 39 Have_You_Alerted_OPT -0.0113 ## 40 Reason_LatereturnfromFieldTrip 0.0109 ## 41 drivers_total_driver -0.00878 ## 42 Boro_RocklandCounty 0.00809 ## 43 Boro_Connecticut -0.00787 ## 44 vehicle_total_disabled_seats 0.00572 ## 45 drivers_num_servSchool -0.00182 #NR: there are some NAs here. I need to have a look at them. These are the no-information columns. corrr_analysis %>% filter(!is.na(time_delayed)) %>% ggplot(aes(x = time_delayed, y = fct_reorder(feature, desc(time_delayed)))) + geom_point() + # Positive Correlations - Contribute to churn geom_segment(aes(xend = 0, yend = feature), color = magma(4)[[3]], data = corrr_analysis %>% filter(time_delayed > 0)) + geom_point(color = magma(4)[[3]], data = corrr_analysis %>% filter(time_delayed > 0)) + # Negative Correlations - Prevent churn geom_segment(aes(xend = 0, yend = feature), color = magma(4)[[1]], data = corrr_analysis %>% filter(time_delayed < 0)) + geom_point(color = magma(4)[[1]], data = corrr_analysis %>% filter(time_delayed < 0)) + # Vertical lines geom_vline(xintercept = 0, color = plasma(3)[[2]], size = 1, linetype = 2) + geom_vline(xintercept = -0.25, color = plasma(3)[[2]], size = 1, linetype = 2) + geom_vline(xintercept = 0.25, color = plasma(3)[[2]], size = 1, linetype = 2) + # Aesthetics theme_bw() + labs(title = "Delay Correlation Analysis", subtitle = paste("Positive => delay", "; negative => reduce delay"), x = "delay correlation", y = "Feature") + theme(text=element_text(family="Arial", size=16), axis.text.y = element_text(size=7))
There are already some interesting insights from these data. Drivers appear more likely to alert OPT for delays which end up shorter. Companies with a lot of ambulatory seats in their fleet suffer shorter delays. Services closer to peak hour suffer lower delays, something of a surprise. Companies with a lot of attendants and hydraulic lifts suffer longer delays. Having a high number of pre-K services disposes a company towards less delay; having a high number of school services disposes them towards more delay. And so on. We also see that there is structure from the original data, including different delays for different boroughs, and for different reasons.
One of the first questions we might ask of all these predictors is: are they correlated with each other? If they are not, then perhaps we could, at some stage, run a naive Bayesian classification. Let's take a look at the cross-correlations between some of the predictors. We include 'time_delayed' for reference:
corrr_high <- corrr_analysis %>% mutate(feature = as.character(feature)) %>% filter(feature != "reported_before_resolved") %>% filter(abs(time_delayed) >= quantile(abs(corrr_analysis$time_delayed), 0.8) ) %>% dplyr::select(feature) %>% as.data.frame() ii_times %>% select(time_delayed, corrr_high[[1]]) %>% ggcorr(nbreaks = 8, low = "#924A51", mid = "#70A5D4", high = "#FE9F14", max_size = 20, min_size = 2, size = 3, hjust = 1, vjust = 0.5, angle = 0, layout.exp = 4)
Even in this reduced set of 'more important' predictors, we can see that many predictors correlate more strongly with each other than with the target variable (time_delayed). This seems pretty bad news for the independence assumption of naive Bayes. The data set also might be a candidate for dimensionality reduction, but at this stage I wanted to keep things as human-transparent as possible.
Building a Cubist regression model with the extra predictors
This next part is straightforward. We go through essentially the same process that we did before, but this time we are armed with the extra predictors. If these predictors are useful, our RMSE should be lower and our \(R^2\) should be higher.
halfFold <- createFolds(ii_times$time_delayed, k = 5, list=TRUE) ii_small <- ii_times[halfFold[[2]],] ii_test <-ii_times[halfFold[[3]],] ii_train <- ii_times %>% anti_join(ii_test, by = "Busbreakdown_ID") rctrl_manual <- trainControl(method = "none", returnResamp = "all") rctrl_repcv <- trainControl(method = "repeatedcv", number = 2, repeats = 5, returnResamp = "all") #---There is actually a problem here. If I do not use the reported_before_resolved flag, I should remove those points from the data set. Do this and retry. #---cubist cubist_grd <- expand.grid( committees = 75, neighbors = 0 ) set.seed(849) cubist_cv <- train(time_delayed ~ . -Busbreakdown_ID, data = ii_small, method = "cubist", metric="RMSE", na.action = na.pass, trControl = rctrl_repcv, tuneGrid = cubist_grd, control = Cubist::cubistControl(unbiased = T, rules = 30, sample = 0) ) cubist_cv ## Cubist ## ## 40665 samples ## 46 predictor ## ## No pre-processing ## Resampling: Cross-Validated (2 fold, repeated 5 times) ## Summary of sample sizes: 20333, 20332, 20333, 20332, 20331, 20334, ... ## Resampling results: ## ## RMSE Rsquared MAE ## 11.74162 0.3502251 8.273603 ## ## Tuning parameter 'committees' was held constant at a value of 75 ## ## Tuning parameter 'neighbors' was held constant at a value of 0 #comm75, rules30: 0.355, RMSE 11.72 cubist_cv$finalModel$splits %>% glimpse ## Observations: 8,341 ## Variables: 8 ## $ committee <dbl> 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... ## $ rule <dbl> 1, 1, 1, 1, 2, 2, 2, 3, 3, 3, 3, 3, 4, 4, 4, 4, 4, ... ## $ variable <fct> Number_Of_Students_On_The_Bus, vehicle_total_with_l... ## $ dir <fct> >, <=, <=, <=, >, <=, >, >, >, <=, >, <=, >, <=, >,... ## $ value <dbl> 4, 4, 225, 5449, 0, 389, 601, 7, 131, 5449, 4, 82, ... ## $ category <fct> , , , , , , , , , , , , , , , , , , , , , , , , ## $ type <chr> "type2", "type2", "type2", "type2", "type2", "type2... ## $ percentile <dbl> 0.8056068, 0.3884421, 0.4687077, 0.5004795, 0.94132... cubist_cv$finalModel$coefficients %>% glimpse ## Observations: 2,234 ## Variables: 48 ## $ `(Intercept)` <dbl> 26.4, 59.3, 11.1, 19.1, 17.7, ... ## $ Has_Contractor_Notified_Schools <dbl> NA, 1.5, NA, 0.7, NA, NA, 3.3,... ## $ Has_Contractor_Notified_Parents <dbl> 0.6, 1.3, 1.5, 0.5, NA, 2.5, N... ## $ Have_You_Alerted_OPT <dbl> 1.4, -0.6, -4.5, -2.4, 0.4, -1... ## $ Number_Of_Students_On_The_Bus <dbl> -0.14, -0.07, -0.14, -0.08, -0... ## $ School_Age <dbl> NA, NA, NA, NA, NA, NA, NA, NA... ## $ Reason_Accident <dbl> 4, 6, NA, 2, NA, NA, NA, NA, 7... ## $ Reason_DelayedbySchool <dbl> -2, NA, NA, NA, NA, NA, NA, NA... ## $ Reason_FlatTire <dbl> 2, 4, NA, NA, NA, NA, NA, NA, ... ## $ Reason_HeavyTraffic <dbl> -2.4, -1.1, NA, -0.4, -3.7, NA... ## $ Reason_LatereturnfromFieldTrip <dbl> NA, 1.1, NA, NA, NA, NA, NA, 1... ## $ Reason_MechanicalProblem <dbl> 2.2, 2.5, NA, 0.8, NA, NA, NA,... ## $ Reason_ProblemRun <dbl> NA, 2, NA, NA, NA, NA, NA, NA,... ## $ Reason_WeatherConditions <dbl> NA, 3.9, NA, NA, 5.0, NA, NA, ... ## $ Reason_WontStart <dbl> 2, 2, NA, NA, NA, NA, NA, NA, ... ## $ Boro_Bronx <dbl> 0.6, 0.6, -1.1, -1.3, 4.2, -2.... ## $ Boro_Brooklyn <dbl> 2.4, -0.4, NA, 2.5, 0.6, 1.4, ... ## $ Boro_Connecticut <dbl> NA, NA, NA, NA, NA, NA, NA, NA... ## $ Boro_Manhattan <dbl> 2.3, 1.5, -1.0, -0.4, 0.6, NA,... ## $ Boro_NassauCounty <dbl> 4, 2, NA, NA, NA, NA, NA, NA, ... ## $ Boro_NewJersey <dbl> NA, NA, NA, NA, NA, NA, NA, NA... ## $ Boro_Queens <dbl> 3.9, 0.7, NA, NA, 0.8, NA, -4.... ## $ Boro_RocklandCounty <dbl> NA, NA, NA, NA, NA, NA, NA, NA... ## $ Boro_StatenIsland <dbl> NA, 1.9, NA, NA, NA, NA, NA, N... ## $ Boro_Westchester <dbl> 2.5, NA, NA, NA, NA, NA, NA, N... ## $ drivers_numServ_school <dbl> -2.0, NA, NA, NA, -0.3, NA, NA... ## $ drivers_numServ_prek <dbl> -2.4, NA, NA, NA, NA, NA, NA, ... ## $ drivers_total_driver <dbl> 0.0330, -0.4627, 0.1162, 0.523... ## $ drivers_total_attendant <dbl> -0.1228, -0.4552, 0.2071, 0.29... ## $ drivers_num_servPreK <dbl> -0.592, NA, -0.463, -0.872, -0... ## $ drivers_num_servSchool <dbl> NA, NA, NA, NA, -1.7894, NA, 1... ## $ drivers_staff_servPreK <dbl> 0.303, NA, NA, NA, -0.982, 0.0... ## $ drivers_staff_servSchool <dbl> 0.0372, 0.3984, -0.2191, -0.42... ## $ Active_Vehicles <dbl> 0.4074, 0.0722, 0.0733, 0.1984... ## $ vehicle_total_max_riders <dbl> -0.00330, 0.00047, 0.02326, 0.... ## $ vehicle_total_reg_seats <dbl> -0.00571, -0.00160, -0.01979, ... ## $ vehicle_total_disabled_seats <dbl> NA, -0.0183, -0.0368, -0.0802,... ## $ vehicle_total_ambulatory_seats <dbl> NA, NA, NA, NA, NA, NA, NA, 0.... ## $ vehicle_total_with_lifts <dbl> 14.131, 0.130, 0.120, 0.460, 1... ## $ vehicle_total_with_attendants <dbl> -0.2296, -0.0095, -0.0022, -0.... ## $ service_type_d2d <dbl> -1.6, 0.5, NA, -1.8, -0.6, -0.... ## $ time_am <dbl> -5.0, -0.5, NA, NA, NA, NA, 4.... ## $ rush_within <dbl> -2.3, 1.3, NA, NA, -1.3, -2.8,... ## $ rush_between <dbl> -4.8, -1.2, NA, -1.3, -3.2, -5... ## $ rush_outside <dbl> -1.6, 0.6, NA, -0.4, -3.5, -2.... ## $ rush_min_from_peak <dbl> NA, 0.019, NA, 0.006, NA, 0.02... ## $ committee <chr> "1", "1", "1", "1", "1", "1", ... ## $ rule <chr> "1", "2", "3", "4", "5", "6", ... cubist_cv$finalModel$usage %>% head ## Conditions Model Variable ## 1 48 86 vehicle_total_reg_seats ## 2 41 20 vehicle_total_ambulatory_seats ## 3 37 75 vehicle_total_with_lifts ## 4 35 49 Reason_HeavyTraffic ## 5 28 90 vehicle_total_max_riders ## 6 24 86 vehicle_total_with_attendants cubist_man <- train(time_delayed ~ . -Busbreakdown_ID, data = ii_train, method = "cubist", metric="RMSE", na.action = na.pass, trControl = rctrl_manual, tuneGrid = cubist_grd, control = Cubist::cubistControl(unbiased = T, rules = 30, sample = 0) ) ii_withpred <- ii_test %>% mutate( time_predicted = predict(cubist_man, newdata = ii_test) ) ii_withpred %>% summarise( rms = (mean((time_predicted-time_delayed)^2))^0.5, sum_sq = sum((time_predicted-time_delayed)^2), tot_sum_sq = sum((time_delayed - mean(time_delayed))^2), r_sq = 1- sum_sq / tot_sum_sq ) #0.371, RMSE 11.6 ## # A tibble: 1 x 4 ## rms sum_sq tot_sum_sq r_sq ## <dbl> <dbl> <dbl> <dbl> ## 1 11.5 5422077. 8556551. 0.366 ii_overfit <- ii_train %>% mutate( time_predicted = predict(cubist_man, newdata = ii_train) ) ii_overfit %>% summarise( rms = (mean((time_predicted-time_delayed)^2))^0.5, sum_sq = sum((time_predicted-time_delayed)^2), tot_sum_sq = sum((time_delayed - mean(time_delayed))^2), r_sq = 1- sum_sq / tot_sum_sq ) #0.374, RMSE 11.5: the overfit is not too bad. ## # A tibble: 1 x 4 ## rms sum_sq tot_sum_sq r_sq ## <dbl> <dbl> <dbl> <dbl> ## 1 11.5 21412743. 34357934. 0.377 protec <- ii_withpred %>% mutate(Reason_Other = 1L) %>% #dummy flag to get the name in select(-contains("Reason")) %>% colnames() pal <- brewer.pal(10, "Paired") ii_withpred %>% mutate(Reason_Other = ifelse(Reason_Accident + Reason_DelayedbySchool + Reason_FlatTire + Reason_HeavyTraffic + Reason_LatereturnfromFieldTrip + Reason_MechanicalProblem + Reason_ProblemRun + Reason_WeatherConditions + Reason_WontStart == 1, 0L, 1L)) %>% gather(key = Reason, value = dummy, -one_of(protec)) %>% filter(dummy == 1) %>% mutate(dummy = NULL) %>% filter(row_number() %% 10 == 0) %>% ggplot(aes(x = time_predicted, y = time_delayed, fill = Reason, color = Reason)) + geom_point(alpha = 0.75, shape = 21) + scale_fill_manual(values = pal) + geom_abline(intercept = 0, slope = 1, color = "red") + coord_cartesian(expand = c(0,0), x = c(0,60)) + theme_classic() + labs(x="time_predicted [min]",y="time_delayed [min]") + theme(text=element_text(family="Arial", size=16)) + theme(axis.line = element_line(colour = 'black', size = 1)) + theme(axis.ticks = element_line(colour = "black", size = 1)) + ggtitle("Cubist with extra predictors: delayed vs predicted")
We have more than doubled our \(R^2\) value, although we are still below 0.5. Our time_delayed vs time_predicted is starting to cluster more around the line of perfect prediction. We are still mispredicting a lot of delay times, but we are certainly doing better than before.
While there is possibly more to be done to improve our predictors, we can guess that we might not get too much further with this approach. We know that our data have a problem related to how delays have been reported, and it isn't going to go away. So perhaps we ought to try the problem as a classification: how well can we predict if delay time will be above some arbitrary threshold? This is the approach we take next.
Related Post
- NYC buses: simple Cubist regression
- Visualization of NYC bus delays with R
- Explaining Black-Box Machine Learning Models – Code Part 2: Text classification with LIME
- Image clustering with Keras and k-Means
- ‘How do neural nets learn?’ A step by step explanation using the H2O Deep Learning algorithm.
R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.