memory efficient prediction with randomForest in R
问题 TL;DR I want to know memory efficient ways of performing a batch prediction with randomForest models built on large datasets (hundreds of features, 10's of thousands of rows). Details : I'm working with a large data-set (over 3GB, in memory) and want to do a simple binary classification using randomForest . Since my data is proprietary, I cannot share it, but lets say the following code runs library(randomForest) library(data.table) myData <- fread("largeDataset.tsv") myFeatures <- myData[,