Tokenizer (Spark 3.5.5 JavaDoc) (original) (raw)


public class Tokenizer
extends UnaryTransformer<String,scala.collection.Seq,Tokenizer>
implements DefaultParamsWritable
A tokenizer that converts the input string to lowercase and then splits it by white spaces.
See Also:
RegexTokenizer, Serialized Form