Following are the steps you can use in your Jupyter or AWS Sagemaker notebook or even in Linux instances to run and install Llama 2 on Petals easily.
Prerequisites:
- Subscribe with your email at Meta's website here.
- Login with same email at Hugging face and Submit Request to access the Llama 2 model here.
- Generate Hugging Face Token here.
Then use following commands in order (Make sure to replace your own hugging face token below):
%pip install petals
import torch
from transformers import AutoTokenizer
from petals import AutoDistributedModelForCausalLM
model_name = "meta-llama/Llama-2-70b-hf"
!huggingface-cli login --token <Your huggingface Token>
tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False, add_bos_token=False)
model = AutoDistributedModelForCausalLM.from_pretrained(model_name)
model = model.cuda()
I hope this helps.
No comments:
Post a Comment