[WIP] Minimal Tokenizer Implementation #513
Draft
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Overview
In response to #262 and #263, and building off of #271, I've been working on
a minimal tokenizer for exo. Initially aimed to remove the
transformers
dependency,but discovered it's a transitive dependency of
mlx_lm
, which will require more extensive changes.So far I've tested on:
llama-3.2-1b
qwen-2.5-coder-1.5b
(both with MLX inference engine)
Questions
Jinja2
andtiktoken
dependencies. Since my implementationis currently a wrapper around these packages, I wanted to confirm if removing these
dependencies is still the desired direction before proceeding with changes?
transformers
dependency in a separate PR given itstransitive nature through
mlx_lm
?