case class BoostingStrategy(treeStrategy: Strategy, loss: Loss, numIterations: Int = 100, learningRate: Double = 0.1, validationTol: Double = 0.001) extends Serializable with Product
Configuration options for org.apache.spark.mllib.tree.GradientBoostedTrees.
- treeStrategy
Parameters for the tree algorithm. We support regression and binary classification for boosting. Impurity setting will be ignored.
- loss
Loss function used for minimization during gradient boosting.
- numIterations
Number of iterations of boosting. In other words, the number of weak hypotheses used in the final model.
- learningRate
Learning rate for shrinking the contribution of each estimator. The learning rate should be between in the interval (0, 1]
- validationTol
validationTol is a condition which decides iteration termination when runWithValidation is used. The end of iteration is decided based on below logic: If the current loss on the validation set is greater than 0.01, the diff of validation error is compared to relative tolerance which is validationTol * (current loss on the validation set). If the current loss on the validation set is less than or equal to 0.01, the diff of validation error is compared to absolute tolerance which is validationTol * 0.01. Ignored when
org.apache.spark.mllib.tree.GradientBoostedTrees.run()
is used.
- Annotations
- @Since("1.2.0")
- Source
- BoostingStrategy.scala
- Alphabetic
- By Inheritance
- BoostingStrategy
- Product
- Equals
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Instance Constructors
- new BoostingStrategy(treeStrategy: Strategy, loss: Loss, numIterations: Int = 100, learningRate: Double = 0.1, validationTol: Double = 0.001)
- treeStrategy
Parameters for the tree algorithm. We support regression and binary classification for boosting. Impurity setting will be ignored.
- loss
Loss function used for minimization during gradient boosting.
- numIterations
Number of iterations of boosting. In other words, the number of weak hypotheses used in the final model.
- learningRate
Learning rate for shrinking the contribution of each estimator. The learning rate should be between in the interval (0, 1]
- validationTol
validationTol is a condition which decides iteration termination when runWithValidation is used. The end of iteration is decided based on below logic: If the current loss on the validation set is greater than 0.01, the diff of validation error is compared to relative tolerance which is validationTol * (current loss on the validation set). If the current loss on the validation set is less than or equal to 0.01, the diff of validation error is compared to absolute tolerance which is validationTol * 0.01. Ignored when
org.apache.spark.mllib.tree.GradientBoostedTrees.run()
is used.
- Annotations
- @Since("1.4.0")
Value Members
- def getLearningRate(): Double
- Annotations
- @Since("1.2.0") @BeanProperty()
- def getLoss(): Loss
- Annotations
- @Since("1.2.0") @BeanProperty()
- def getNumIterations(): Int
- Annotations
- @Since("1.2.0") @BeanProperty()
- def getTreeStrategy(): Strategy
- Annotations
- @Since("1.2.0") @BeanProperty()
- def getValidationTol(): Double
- Annotations
- @Since("1.4.0") @BeanProperty()
- var learningRate: Double
- Annotations
- @Since("1.2.0")
- var loss: Loss
- Annotations
- @Since("1.2.0")
- var numIterations: Int
- Annotations
- @Since("1.2.0")
- def productElementNames: Iterator[String]
- Definition Classes
- Product
- def setLearningRate(arg0: Double): Unit
- Annotations
- @Since("1.2.0") @BeanProperty()
- def setLoss(arg0: Loss): Unit
- Annotations
- @Since("1.2.0") @BeanProperty()
- def setNumIterations(arg0: Int): Unit
- Annotations
- @Since("1.2.0") @BeanProperty()
- def setTreeStrategy(arg0: Strategy): Unit
- Annotations
- @Since("1.2.0") @BeanProperty()
- def setValidationTol(arg0: Double): Unit
- Annotations
- @Since("1.4.0") @BeanProperty()
- var treeStrategy: Strategy
- Annotations
- @Since("1.2.0")
- var validationTol: Double
- Annotations
- @Since("1.4.0")