Text generation with Mistral
The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
32k context window (vs 8k context in v0.1)
Rope-theta = 1e6
No Sliding-Window Attention
For full details of this model please read our paper and release blog post.
https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2
The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
32k context window (vs 8k context in v0.1)
Rope-theta = 1e6
No Sliding-Window Attention
For full details of this model please read our paper and release blog post.
https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2