com.salesforce.op.stages.impl.classification
Default model types and model parameters for problem type
Default model types and model parameters for problem type
Default model types and model parameters for problem type
Creates a new Binary Classification Model Selector with a Cross Validation
Create the model selector for specified interface
Create the model selector for specified interface
training split of cross validator
data prep class
evaluation to do on data
list of models to use
sequence of models and parameters to explore
default model types and model parameters for problem type
model selector with these settings
Creates a new Binary Classification Model Selector with a Cross Validation
Creates a new Binary Classification Model Selector with a Cross Validation
instance that will balance and split the data
number of folds for cross validation (>= 2)
metric name in evaluation: AuROC or AuPR
List of evaluators applied on training + holdout data for evaluation. Default is the standard OpBinaryClassificationEvaluator and OpBinScoreEvaluator.
random seed
whether or not stratify cross validation. Caution : setting that param to true might impact the runtime
level of parallelism used to schedule a number of models to be trained/evaluated so that the jobs can be run concurrently
list of model types to run grid search on must from supported types in BinaryClassificationModelsToTry (OpLogisticRegression, OpRandomForestClassifier, OpGBTClassifier, OpLinearSVC, OpDecisionTreeClassifier, OpNaiveBayes)
pass in an explicit list pairs of estimators and the accompanying hyperparameters to for model selection Seq[(EstimatorType, Array[ParamMap])] where Estimator type must be an Estimator that takes in a label (RealNN) and features (OPVector) and returns a prediction (Prediction)
maximum allowable time to wait for a model to finish running (default is 1 day)
Classification Model Selector with a Cross Validation
Creates a new Binary Classification Model Selector with a Train Validation Split
Creates a new Binary Classification Model Selector with a Train Validation Split
instance that will balance and split the data
ratio between training set and validation set (>= 0 && <= 1)
metric name in evaluation: AuROC or AuPR
List of evaluators applied on training + holdout data for evaluation. Default is the standard OpBinaryClassificationEvaluator and OpBinScoreEvaluator.
random seed
whether or not stratify train validation split. Caution : setting that param to true might impact the runtime
level of parallelism used to schedule a number of models to be trained/evaluated so that the jobs can be run concurrently
list of model types to run grid search on must from supported types in BinaryClassificationModelsToTry (OpLogisticRegression, OpRandomForestClassifier, OpGBTClassifier, OpLinearSVC, OpDecisionTreeClassifier, OpNaiveBayes)
pass in an explicit list pairs of estimators and the accompanying hyperparameters to for model selection Seq[(EstimatorType, Array[ParamMap])] where Estimator type must be an Estimator that takes in a label (RealNN) and features (OPVector) and returns a prediction (Prediction)
maximum allowable time to wait for a model to finish running (default is 1 day)
Classification Model Selector with a Train Validation Split
A factory for Binary Classification Model Selector