You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
tokenizer_catalog.create_text_analyzer(name TEXT, config TEXT): Create a text analyzer.
tokenizer_catalog.drop_text_analyzer(name TEXT): Drop a text analyzer.
tokenizer_catalog.apply_text_analyzer(text TEXT, text_analyzer_name TEXT) RETURNS TEXT[]: Apply a text analyzer to a text.
tokenizer_catalog.create_stopwords(name TEXT, config TEXT): Create a stopwords dictionary.
tokenizer_catalog.drop_stopwords(name TEXT): Drop a stopwords dictionary.
tokenizer_catalog.create_synonym(name TEXT, config TEXT): Create a synonym dictionary.
tokenizer_catalog.drop_synonym(name TEXT): Drop a synonym dictionary.
Model
tokenizer_catalog.create_custom_model(name TEXT, config TEXT): Create a custom model.
tokenizer_catalog.create_custom_model_tokenizer_and_trigger(tokenizer_name TEXT, model_name TEXT, text_analyzer_name TEXT, table_name TEXT, source_column TEXT, target_column TEXT): Create a custom model tokenizer and trigger to update the target column automatically.
tokenizer_catalog.drop_custom_model(name TEXT): Drop a custom model.
tokenizer_catalog.add_preload_model(name TEXT): Add a model to the preload list.
tokenizer_catalog.remove_preload_model(name TEXT): Remove a model from the preload list.
tokenizer_catalog.list_preload_models() RETURNS TEXT[]: List all preload models.
tokenizer_catalog.create_lindera_model(name TEXT, config TEXT): Create a lindera model.
tokenizer_catalog.drop_lindera_model(name TEXT): Drop a lindera model.
tokenizer_catalog.create_huggingface_model(name TEXT, config TEXT): Create a huggingface model.
tokenizer_catalog.drop_huggingface_model(name TEXT): Drop a huggingface model.
Tokenizer
tokenizer_catalog.create_tokenizer(name TEXT, config TEXT): Create a tokenizer.
tokenizer_catalog.drop_tokenizer(name TEXT): Drop a tokenizer.
tokenizer_catalog.tokenize(text TEXT, tokenizer_name TEXT) RETURNS INT[]: Tokenize a text.
Configuration
We utilize TOML syntax to express all configurations.