This video installs Codestral Mamba locally which is an open code model based on the Mamba2 architecture.
Code:
conda create -n codestralmamba python=3.11 -y && conda activate codestralmamba
pip install torch huggingface_hub pathlib2
pip install mistral_inference>=1 mamba-ssm causal-conv1d
from huggingface_hub import snapshot_download
from pathlib import Path
mistral_models_path = Path.home().joinpath('mistral_models', 'mamba-codestral-7B-v0.1')
mistral_models_path.mkdir(parents=True, exist_ok=True)
snapshot_download(repo_id="mistralai/mamba-codestral-7B-v0.1", allow_patterns=["params.json", "consolidated.safetensors", "tokenizer.model.v3"], local_dir=mistral_models_path)
mistral-chat $HOME/mistral_models/mamba-codestral-7B-v0.1 --instruct --max_tokens 256
No comments:
Post a Comment