AutoJapaneseTokenizer
AutoJapaneseTokenizer.from_pretrained()
AutoJapaneseTokenizer.save_pretrained()
AutoJapaneseTokenizer.tokenize()
tokenizer_dict()
BaseTokenizer
BaseTokenizer.fit()
BaseTokenizer.from_pretrained()
BaseTokenizer.save_pretrained()
BaseTokenizer.tokenize()
BaseTokenizer.tokenizer_config_file
JumanppTokenizer
JumanppTokenizer.from_pretrained()
JumanppTokenizer.save_pretrained()
JumanppTokenizer.tokenize()
MecabTokenizer
MecabTokenizer.from_pretrained()
MecabTokenizer.save_pretrained()
MecabTokenizer.setup_tagger()
MecabTokenizer.tokenize()
SpacyGinzaTokenizer
SpacyGinzaTokenizer.from_pretrained()
SpacyGinzaTokenizer.save_pretrained()
SpacyGinzaTokenizer.tokenize()
SentencePieceTokenizer
SentencePieceTokenizer.decode()
SentencePieceTokenizer.encode()
SentencePieceTokenizer.fit()
SentencePieceTokenizer.from_pretrained()
SentencePieceTokenizer.model_cache_dir
SentencePieceTokenizer.save_pretrained()
SentencePieceTokenizer.tokenize()
SentencePieceTokenizer.tokenizer_file_prefix
SudachiTokenizer
SudachiTokenizer.dict_type
SudachiTokenizer.normalize
SudachiTokenizer.tokenizer
SudachiTokenizer.from_pretrained()
SudachiTokenizer.save_pretrained()
SudachiTokenizer.tokenize()
download_file()
run_sh()
unbz_file()
untar_file()
unzip_file()
CountVectorizer
CountVectorizer.from_pretrained()
ModelCard
ModelCard.description
ModelCard.model_computed_fields
ModelCard.model_config
ModelCard.model_fields
ModelCard.name
ModelCard.tokenizer_name
ModelCard.url
W2VSWEMVectorizer
W2VSWEMVectorizer.w2v_dir
W2VSWEMVectorizer.fit()
W2VSWEMVectorizer.from_pretrained()
W2VSWEMVectorizer.get_token_vectors()
W2VSWEMVectorizer.save_pretrained()
W2VSWEMVectorizer.setup_w2v_if_not_installed()
W2VSWEMVectorizer.transform()
TfidfVectorizer
TfidfVectorizer.from_pretrained()
ModelCard.model_name
TransformerEmbeddingVectorizer
TransformerEmbeddingVectorizer.from_pretrained()
TransformerEmbeddingVectorizer.get_token_vectors()
TransformerEmbeddingVectorizer.transform()