Predicting pneumonia outcomes: Modelling via DataRobot API

This post is a supplementary material for an assignment. The assignment is part of the Augmented Machine Learning unit for a Specialised Diploma in Data Science for Business. The aim of this project is to classify if patients with Community Acquired Pneumonia (CAP) became better after seeing a doctor or became worse despite seeing a doctor. Previously, EDA (part1, part 2)and feature enginnering was done in R. The dataset was then imported into DataRobot for predictive modelling (the aim of the assigment was to use DataRobot).
The post here will outline basic modelling done in DataRobot and how I used R to conduct advanced feature selection to enhance the performance in subsequent modelling.

Modelling (basic)

Predictors for modelling

When the data is imported into DataRobot, the raw features will go through “a reasonableness” check for useful information (e.g., are not duplicates or reference ids and not a constant value)". Predictors that pass this screener will form Informative Features which DataRobot deems to be more appropriate for modelling.
DataRobot deemed all variables informative including the case identifier, Pt_CaseNumber. The uniqueness of the case identifier must have been disrupted when the dataset was split up between unseen data and data for DataRobot. A custom Feature List, L1,included all variables except Pt_CaseNumber.

Modelling options

The number of K-fold was increased from the default 5 to 10 folds. The increase in K-folds helps to reduce biasness and therefore reduces the risk of underfitting. Based on the concept of bias and variance trade off, the decrease of biasness will increase the variance. Nonetheless, the sample size of the training set is large enough (1840 rows) to accommodate this increase in folds without adversely compromising variance. Additionally, 10-fold cross validation is the default number of folds for some modelling programs; for example, R’s modern modelling meta-package, Tidymodels.
DataRobot was also permitted to search for Interactions among the variables. There is support in identifying interactions as the combined effect of two or more variables maybe different than the impact of the individual constituent variables.
Down-sampling was introduced to address the imbalanced dataset (DataRobot only allows down-sampling). The proportion of positive and negative outcome of the binary classification was imbalanced which [could affect model performance as most models work best when the classes are about equal)( There were 6.25 more Better outcomes than Worse outcomes. DataRobot’s Down-sampling was used to reduce the disparity. Under-sampling can help to address imbalance dataset. The number of Better outcomes were reduced by 20%. Post down-sampling, there was 5.22 Better outcomes than Worse outcomes.

Performance Metric

The performance metric for the binary classification was Area Under the Curve (AUC) which is common in the medical literature. In DataRobot, when down-sampling is used, the performance metric AUC becomes a weighted AUC.

1st Quick Mode

The experiment ran on Quick Mode (DataRobot has 3 modelling modes: Autopilot, Quick, Manual. The student version which I used only has Quick and Manual Mode).

Modelling (2nd run)

The results from the first round of Quick Mode were used for advanced feature selection. A smaller but possibly more influential subset of variables were used in a second round of Quick Mode. The advanced feature selection used results from the first round of Quick Mode to create 2 new Feature Lists. The creation of these 2 new Feature List required RStudio to be connected to DataRobot via its API.

DataRobot API


Connecting to DataRobot

DataRobot was connected to RStudio via datarobot::ConnectToDataRobot. The endpoint for cloud based DataRobot is The token can be found under Developer Tools option.

ConnectToDataRobot(endpoint = "", 
                   token = "API token key")
## [1] "Authentication token saved"

Saving the relevant project ID

projects_inDR<-data.frame(name=ListProjects()$projectName, ID=ListProjects()$projectId)

project_ID<-projects_inDR %>% filter(name=="Classify CAP") %>% pull(ID)
## # A tibble: 11 x 2
##    name                        ID                      
##    <chr>                       <chr>                   
##  1 Classify CAP                5f54dce5e5e3f924b99408ee
##  2 CAP Experiment 3            5f3678ff735d1f59f99534ed
##  3 CAP Experiment 2            5f33f9899071ae44ded6de84
##  4 CAP Experiment 1            5f269f896c6a7a1cefaff780
##  5 CAP informative ft          5f2694a9d6e13f1dc8ae17d4
##  6 CleanDR OG cv10             5f2655a8d6e13f1c26ae17c7
##  7 CleanDR OG cv5              5f263758ce243f1bfac853eb
##  8 P04_Readmission(1).csv      5f16dd3b99c1372d672bdde4
##  9 P03_Advertising(1).csv      5f0da6861c29b917239402dc
## 10 P02_LendingClubData2(1).csv 5f047ce3c40c123f3487346b
## 11 P01_TitanicData.csv         5efb44cc60882464b616c03b

Saving the models

List_Model3<-ListModels(project_ID) %>% = F) 

List_Model3 %>% select(modelType, expandedModel, featurelistName, `Weighted AUC.crossValidation`) %>% arrange(-`Weighted AUC.crossValidation`)DT::datatable(rownames = F, options = list(searchHighlight = TRUE, paging= T))

In the 1st round of Quick Mode, 13 models were created. The best model was Extreme Gradient Boosted Tree Classifier, with a 10 fold cross validation weighted AUC of 0.9299.

1. Advanced Feature Selection (DR Reduced Features)

During the first round of Quick Mode, DataRobot created other Feature Lists including DR Reduced Features. These feature lists contained curated variables consisting of most important features based on Feature Impact scores from models.

Ft_DR<-GetFeaturelist(project_ID, "5f54e0f611a5fc3a03274070") %>% as_tibble()
Ft_DR$description %>% unqiue()
## [1] "The most important features based on Feature Impact scores from a particular model."

3/13 models used DR Reduced Features in the 1st round of Quick Mode.

List_Model3  %>% count(featurelistName)
## # A tibble: 3 x 2
##   featurelistName            n
##   <chr>                  <int>
## 1 DR Reduced Features M8     3
## 2 L1                         9
## 3 Multiple featurelists      1

Originally 70 variables were used for Quick Mode. 33 variables were identified for DR Reduced Features M8 (32 predictors + 1 outcome).

## [1] 33

2. Advanced Feature Selection (Top 32 Mean Feature Impact Variables)

The number of predictors were kept the same as the number of predictors in DR Reduced Features to see which technique achieved better performance. The steps for identifying these variables were as follows:

  1. All the cross-validation scores for all models were enabled on DataRobot GUI (by default, DataRobot calculates the cross-validation score for the top few models based on the 1st cross-validation fold score). The top 80% of the models based on cross validation AUC were selected. These selected models also had to have sufficient sample size of >64%. 9/13 models passed this screener.
# Extract models 
#extract models (df format)
models_df<, simple = F)
## top 80%  cv AUC
top80AUC<-quantile(models_df$`Weighted AUC.crossValidation`, probs = seq(0,1,.1), na.rm = T)[["20%"]]
# remove small sample size % models
models_listdetails_big <- Filter(function(m) m$metrics$`Weighted AUC`$crossValidation>top80AUC & m$samplePct >= 64 , models_listdetails)

Ft_mean_screened<-Ft_mean_screenModels(project_ID), simple = F) %>% select(expandedModel, featurelistName, `Weighted AUC.crossValidation`) %>% arrange(-`Weighted AUC.crossValidation`) DT::datatable(rownames = F, options = list(searchHighlight = TRUE, paging= T))
  1. The unnormalized feature impact score for all the variables in these models were averaged out. Variables in the top 32 mean feature impact score were used to create a new set of Feature List.
Ft_mean_var<-function(Ft_mean_screened_results, ID_of_project){
# feature impact of selected models 
all_impact<- NULL
for(i in 1:length(Ft_mean_screened_results)) {  
     featureImpact <- GetFeatureImpact(Ft_mean_screened_results[[i]])
    featureImpact$modelname <- Ft_mean_screened_results[[i]]$modelType
    all_impact <- rbind(all_impact,featureImpact)

#mean impact
mean_impact<-all_impact %>% group_by(featureName) %>% 
    summarise(avg_impact = mean(impactUnnormalized), .groups="drop")

#top X mean feature impact scores 
## number of predictors for DR
no_of_DRft<-ListFeaturelists(ID_of_project) %>% %>% filter(str_detect(name, "DR Reduced")) %>% nrow()-1 # minus 1 as one of the variables is the target 

##top X mean feature impact scores 
top_mean_names<-mean_impact %>% slice_max(order_by = avg_impact, n=no_of_DRft ) %>% pull(featureName)

# multiple output
output<-list(all_impact, top_mean_names)
return (output)

Ft_mean_varOutput<- Ft_mean_var(Ft_mean_screened, project_ID)

The calculation can be seen on DataRobot’s GUI.

The top 32 variables based on mean feature impact score and their respective feature impact score for each model is plotted here.

  filter(featureName %in% Ft_mean_varOutput[[2]]) %>% 
  mutate(featureName= fct_reorder(featureName, impactUnnormalized, .fun =mean)) %>%
     ggplot(aes(featureName, impactUnnormalized, colour=featureName))+ geom_point(alpha=.5) + coord_flip() + theme(legend.position="none") + labs(title = "Top 32 variables based on mean feature impact score", subtitle = "each dot is the score for a specific model")

New Feature List using the top 32 mean feature impact variables was created.

Feature_id_percent_rank = CreateFeaturelist(project_ID, 
                                            listName= "mean ft impact" , 
## [1] "Featurelist mean ft impact created"

DR Features vs Top 32 mean Feature Impact variables

Both Feature Lists had these predictors in common

intersect(Ft_mean_varOutput[[2]], Ft_DR$features)
##  [1] "Abx_AmoxicillinSulbactamOral" "Abx_Duration"                
##  [3] "Pt_Age"                       "PE_AMS"                      
##  [5] "Lab_RBC"                      "PE_HR"                       
##  [7] "Care_daysUnfit"               "Care_admit"                  
##  [9] "Lab_Hb"                       "Lab_urea"                    
## [11] "Care_BPSupport"               "Lab_Neu"                     
## [13] "Lab_WBC"                      "Lab_plt"                     
## [15] "Care_GP/OutptVisit"           "Care_breathingAid"           
## [17] "PE_temp"                      "Lab_Sugar"                   
## [19] "SS_daysOfRespSymp"            "Hx_comorbidities"            
## [21] "Lab_Cr"                       "Abx_LowFreq"                 
## [23] "PE_RR"                        "PE_BP_MAP"                   
## [25] "SS_breathing"                 "Social_smoke"                
## [27] "Abx_Ceftriaxone"              "Abx_ClarithromycinIV"        
## [29] "HCAP_IVAbx"                   "Abx_no"                      
## [31] "Hx_brainMental"

The only difference was that the top 32 mean Feature Impact variables contained medical history on the type of heart disease the patient had.

setdiff(Ft_mean_varOutput[[2]], Ft_DR$features)
## [1] "Hx_heart_type"

Running 2nd Quick Mode

Another round of Quick Mode was ran with both of the new Feature Lists created with advanced feature selection. Again, cross validation scores for all models was calculated. 8 additional models were generated with DR Reduced features (the other 3 was previously generated in the 1st round of Quick Mode). 10 additional models were generated with the Top 32 mean Feature Impact score variables.

#Get all the models from 1st and 2nd Quick Mode 
List_Model322<-ListModels(project_ID) %>% = F) 

List_Model322 %>% count(featurelistName)
## # A tibble: 4 x 2
##   featurelistName            n
##   <chr>                  <int>
## 1 DR Reduced Features M8    11
## 2 L1                         9
## 3 mean ft impact            10
## 4 Multiple featurelists      1

In the next post, the results from all the models from both runs of Quick Mode will be analysed.