|
19 | 19 | "mistralai/Mistral-7B-v0.1": "https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/config.json,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/generation_config.json,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/pytorch_model-00001-of-00002.bin,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/pytorch_model-00002-of-00002.bin,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/pytorch_model.bin.index.json,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/special_tokens_map.json,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/tokenizer.json,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/tokenizer.model,https://huggingface.co/mistralai/Mistral-7B-v0.1/resolve/main/tokenizer_config.json",
|
20 | 20 | "mistralai/Mistral-7B-Instruct-v0.1": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/config.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/generation_config.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/pytorch_model-00001-of-00002.bin,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/pytorch_model-00002-of-00002.bin,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/pytorch_model.bin.index.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/special_tokens_map.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/tokenizer.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/tokenizer.model,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/resolve/main/tokenizer_config.json",
|
21 | 21 | "mistralai/Mistral-7B-Instruct-v0.2": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/config.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/generation_config.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/pytorch_model-00001-of-00003.bin,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/pytorch_model-00002-of-00003.bin,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/pytorch_model-00003-of-00003.bin,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/pytorch_model.bin.index.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/special_tokens_map.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/tokenizer.json,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/tokenizer.model,https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/resolve/main/tokenizer_config.json",
|
| 22 | + |
| 23 | + # huggingface-cli prefixed Models will download using the huggingface-cli tool |
| 24 | + # TODO: Convert all of the MODEL_REPOS with a NamedTuple that includes the install_method |
| 25 | + "huggingface-cli/meta-llama/Meta-Llama-3-8B": "", |
22 | 26 | }
|
23 | 27 |
|
24 | 28 | JOB_RUNNERS = {
|
@@ -67,7 +71,7 @@ def model_should_run_on_event(model: str, event: str) -> bool:
|
67 | 71 | elif event == "push":
|
68 | 72 | return model in []
|
69 | 73 | elif event == "periodic":
|
70 |
| - return model in ["openlm-research/open_llama_7b"] |
| 74 | + return model in ["openlm-research/open_llama_7b", "huggingface-cli/meta-llama/Meta-Llama-3-8B"] |
71 | 75 | else:
|
72 | 76 | return False
|
73 | 77 |
|
@@ -105,12 +109,22 @@ def export_models_for_ci() -> dict[str, dict]:
|
105 | 109 | if not model_should_run_on_event(repo_name, event):
|
106 | 110 | continue
|
107 | 111 |
|
| 112 | + # This is mostly temporary to get this finished quickly while |
| 113 | + # doing minimal changes, see TODO at the top of the file to |
| 114 | + # see how this should probably be done |
| 115 | + install_method = "wget" |
| 116 | + final_repo_name = repo_name |
| 117 | + if repo_name.startswith("huggingface-cli"): |
| 118 | + install_method = "huggingface-cli" |
| 119 | + final_repo_name = repo_name.replace("huggingface-cli/", "") |
| 120 | + |
108 | 121 | record = {
|
109 |
| - "repo_name": repo_name, |
110 |
| - "model_name": repo_name.split("/")[-1], |
| 122 | + "repo_name": final_repo_name, |
| 123 | + "model_name": final_repo_name.split("/")[-1], |
111 | 124 | "resources": MODEL_REPOS[repo_name],
|
112 | 125 | "runner": runner[0],
|
113 | 126 | "platform": runner[1],
|
| 127 | + "install_method": install_method, |
114 | 128 | "timeout": 90,
|
115 | 129 | }
|
116 | 130 |
|
|
0 commit comments