public abstract class UnaryTransformer<IN,OUT,T extends UnaryTransformer<IN,OUT,T>> extends Transformer implements HasInputCol, HasOutputCol, Logging
Constructor and Description |
---|
UnaryTransformer() |
Modifier and Type | Method and Description |
---|---|
T |
setInputCol(String value) |
T |
setOutputCol(String value) |
DataFrame |
transform(DataFrame dataset,
ParamMap paramMap)
Transforms the dataset with provided parameter map as additional parameters.
|
org.apache.spark.sql.types.StructType |
transformSchema(org.apache.spark.sql.types.StructType schema,
ParamMap paramMap)
:: DeveloperAPI ::
|
transform, transform
equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
getInputCol, inputCol
getOutputCol, outputCol
addOutputColumn, checkInputColumn, explainParams, get, getParam, isSet, paramMap, params, set, set, validate, validate
uid
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public T setInputCol(String value)
public T setOutputCol(String value)
public org.apache.spark.sql.types.StructType transformSchema(org.apache.spark.sql.types.StructType schema, ParamMap paramMap)
PipelineStage
Derives the output schema from the input schema and parameters. The schema describes the columns and types of the data.
transformSchema
in class PipelineStage
schema
- Input schema to this stageparamMap
- Parameters passed to this stagepublic DataFrame transform(DataFrame dataset, ParamMap paramMap)
Transformer
transform
in class Transformer
dataset
- input datasetparamMap
- additional parameters, overwrite embedded params