Comments (2)
So usually in generate it uses the models default chat template (so it's a chat). You can use the raw prompt (so it's just the text) by specifying --ignore-chat-template
. There is currently no way to do the completion
version in the CLI. But if you use the API you could do it like this:
from mlx_lm import load, generate
model, tokenizer = load("mistralai/Mistral-7B-Instruct-v0.1")
prompt = ""
completion = ""
text = tokenizer.apply_chat_template(
[
{"role": "user", "content": prompt},
{"role": "assistant", "content": completion},
],
tokenize=False,
add_generation_prompt=True,
)
response = generate(model, tokenizer, prompt=text)
from mlx-examples.
I see, thank you!
In case it helps anyone down the line:
My goal was to generate using fine tuned adapters, as well as with a system prompt and user prompt in the chat format.
Based on generate.py and utils.py, I load the model with the adapters like this:
from mlx_lm import load, generate
model_repo = "mlx-community/Meta-Llama-3-8B-Instruct-4bit"
adapter_path = "/adapters"
model, tokenizer = load(model_repo, adapter_path=adapter_path)
Then, I can generate:
system_prompt = "Be a helpful assistant"
prompt = "Hey, tell me about Llama"
text = tokenizer.apply_chat_template(
[
{"role": "system", "content": system_prompt},
{"role": "user", "content": prompt},
],
tokenize=False,
add_generation_prompt=True,
)
response = generate(model, tokenizer, prompt=text, verbose=True)
from mlx-examples.
Related Issues (20)
- Contributing SigLIP to `mlx-examples` HOT 1
- MNIST Example Error: 403 Forbidden HOT 4
- omp_set_nested routine deprecated, please use omp_set_max_active_levels instead HOT 1
- FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. HOT 1
- Llama-3-8B-Instruct-Gradient-1048k-4bit not working? HOT 2
- Generating after LORA training CAN NOT Stop Properly HOT 3
- Issue with Fusing Models - Output is Bad HOT 2
- GatedRepoError: 401 Client Error; "You must be authenticated to access it." HOT 1
- how to merge lora adapter to base model HOT 1
- delete and uninstall HOT 11
- KV Cache can only process more than self.step tokens if offset % step == 0 HOT 2
- Text to Speech MLX model. HOT 1
- SLM Example Code HOT 1
- Enhance load function to support model configuration editing HOT 1
- Support for full set of output formats - e.g. vtt, json and json-full HOT 2
- Whisper stutters HOT 8
- mlx 0.13 very slow with q8 and fp16 HOT 5
- Fine tuned a Mixtral-8x7B-Instruct-v0.1 model and unable to load with AutoModelForCausalLM HOT 1
- Phi-3-mini-4k-instruct : Failing to stop at <|end|> on generating the answer. HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from mlx-examples.