|
@@ -248,13 +248,11 @@
|
|
"import transformers\n",
|
|
"import transformers\n",
|
|
"from transformers import AutoModelForCausalLM, AutoTokenizer\n",
|
|
"from transformers import AutoModelForCausalLM, AutoTokenizer\n",
|
|
"\n",
|
|
"\n",
|
|
- "\n",
|
|
|
|
- "model_dir = \"/home/ubuntu/release/Meta-Llama-3-8B-Instruct-HF\"\n",
|
|
|
|
|
|
+ "model_dir = \"${path_the_converted_hf_model}\"\n",
|
|
"model = AutoModelForCausalLM.from_pretrained(\n",
|
|
"model = AutoModelForCausalLM.from_pretrained(\n",
|
|
" model_dir,\n",
|
|
" model_dir,\n",
|
|
" device_map=\"auto\",\n",
|
|
" device_map=\"auto\",\n",
|
|
" )\n",
|
|
" )\n",
|
|
- "# model = LlamaForCausalLM.from_pretrained(model_dir)\n",
|
|
|
|
"tokenizer = AutoTokenizer.from_pretrained(model_dir)\n"
|
|
"tokenizer = AutoTokenizer.from_pretrained(model_dir)\n"
|
|
]
|
|
]
|
|
},
|
|
},
|