All nonnegative word counts are set to 1 when set to true
Indicates whether a case sensitive comparison is performed on stop words.
Specify the language to use for stop word removal.
Minimum number of documents in which a term should appear.
Minumum token length; must be 0 or greater.
The size of the Ngrams
Set the number of features to hash each document to
The words to be filtered out.
Indicates whether to convert all characters to lowercase before tokenizing.
Indicates whether the regex splits on gaps (true) or matches tokens (false)
Regex pattern used to match delimiters if gaps (true) or tokens (false)
Scale the Term Frequencies by IDF when set to true
Enumerate N grams when set
Indicates whether to remove stop words from tokenized data.
Tokenize the input when set to true