Skip to content

Commit 03e5909

Browse files
committed
updates for 70b and gpu process monitor
1 parent 77774d2 commit 03e5909

File tree

3 files changed

+456
-2
lines changed

3 files changed

+456
-2
lines changed

dist_run.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -58,6 +58,7 @@
5858
NAME_TO_DISTRIBUTION_AND_DTYPE = {
5959
"llama2-7b-chat": ("meta-llama/Llama-2-7b-chat-hf", torch.float16),
6060
"llama3": ("meta-llama/Meta-Llama-3-8B-Instruct", torch.bfloat16),
61+
"llama3-70b": ("meta-llama/Meta-Llama-3-70B-Instruct", torch.bfloat16),
6162
}
6263

6364

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
from transformers import AutoModelForCausalLM, AutoTokenizer
22

3-
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct")
4-
model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct")
3+
# tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct")
4+
model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-70B-Instruct")
55
print("Model weights and tokenizer downloaded")

0 commit comments

Comments
 (0)