RegexTokenizer (Spark 3.5.5 JavaDoc) (original) (raw)


public class RegexTokenizer
extends UnaryTransformer<String,scala.collection.Seq,RegexTokenizer>
implements DefaultParamsWritable
A regex based tokenizer that extracts tokens either by using the provided regex pattern to split the text (default) or repeatedly matching the regex (if gaps is false). Optional parameters also allow filtering tokens using a minimal length. It returns an array of strings that can be empty.
See Also:
Serialized Form