final class EMLDAOptimizer extends LDAOptimizer
Optimizer for EM algorithm which stores data + parameter graph, plus algorithm parameters.
Currently, the underlying implementation uses Expectation-Maximization (EM), implemented according to the Asuncion et al. (2009) paper referenced below.
References:
- Original LDA paper (journal version):
Blei, Ng, and Jordan. "Latent Dirichlet Allocation." JMLR, 2003.
- This class implements their "smoothed" LDA model.
- Paper which clearly explains several algorithms, including EM: Asuncion, Welling, Smyth, and Teh. "On Smoothing and Inference for Topic Models." UAI, 2009.
- Annotations
- @Since( "1.4.0" )
- Source
- LDAOptimizer.scala
- Alphabetic
- By Inheritance
- EMLDAOptimizer
- LDAOptimizer
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
- new EMLDAOptimizer()
Value Members
-
def
getKeepLastCheckpoint: Boolean
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
- Annotations
- @Since( "2.0.0" )
-
def
setKeepLastCheckpoint(keepLastCheckpoint: Boolean): EMLDAOptimizer.this.type
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up). Deleting the checkpoint can cause failures if a data partition is lost, so set this bit with care.
Default: true
- Annotations
- @Since( "2.0.0" )
- Note
Checkpoints will be cleaned up via reference counting, regardless.