|
@@ -16,7 +16,7 @@
|
|
},
|
|
},
|
|
"source": [
|
|
"source": [
|
|
"# **Getting to know Llama 3: Everything you need to start building**\n",
|
|
"# **Getting to know Llama 3: Everything you need to start building**\n",
|
|
- "Our goal in this session is to provide a guided tour of Llama 3, including understanding different Llama 3 models, how and where to access them, Generative AI and Chatbot architectures, prompt engineering, RAG (Retrieval Augmented Generation), Fine-tuning and more. All this is implemented with a starter code for you to take it and use it in your Llama 3 projects."
|
|
|
|
|
|
+ "Our goal in this session is to provide a guided tour of Llama 3 with comparison with Llama 2, including understanding different Llama 3 models, how and where to access them, Generative AI and Chatbot architectures, prompt engineering, RAG (Retrieval Augmented Generation), Fine-tuning and more. All this is implemented with a starter code for you to take it and use it in your Llama 3 projects."
|
|
]
|
|
]
|
|
},
|
|
},
|
|
{
|
|
{
|
|
@@ -25,9 +25,8 @@
|
|
"id": "ioVMNcTesSEk"
|
|
"id": "ioVMNcTesSEk"
|
|
},
|
|
},
|
|
"source": [
|
|
"source": [
|
|
- "##**0 - Prerequisites**\n",
|
|
|
|
|
|
+ "### **0 - Prerequisites**\n",
|
|
"* Basic understanding of Large Language Models\n",
|
|
"* Basic understanding of Large Language Models\n",
|
|
- "\n",
|
|
|
|
"* Basic understanding of Python"
|
|
"* Basic understanding of Python"
|
|
]
|
|
]
|
|
},
|
|
},
|
|
@@ -303,7 +302,9 @@
|
|
"source": [
|
|
"source": [
|
|
"## **2 - Using and Comparing Llama 3 and Llama 2**\n",
|
|
"## **2 - Using and Comparing Llama 3 and Llama 2**\n",
|
|
"\n",
|
|
"\n",
|
|
- "In this notebook, we will use the Llama 2 7b chat and Llama 3 8b instruct models hosted on [Groq](https://console.groq.com/). You'll need to first [sign in](https://console.groq.com/) with your github or gmail account, then get an [API token](https://console.groq.com/keys) to try Groq out for free.\n"
|
|
|
|
|
|
+ "In this notebook, we will use the Llama 2 70b chat and Llama 3 8b and 70b instruct models hosted on [Groq](https://console.groq.com/). You'll need to first [sign in](https://console.groq.com/) with your github or gmail account, then get an [API token](https://console.groq.com/keys) to try Groq out for free. (Groq runs Llama models very fast and they only support one Llama 2 model: the Llama 2 70b chat).\n",
|
|
|
|
+ "\n",
|
|
|
|
+ "**Note: You can also use other Llama hosting providers such as [Replicate](https://replicate.com/blog/run-llama-3-with-an-api?input=python), [Togther](https://docs.together.ai/docs/quickstart). Simply click the links here to see how to run `pip install` and use their freel trial API key with example code to modify the following three cells in 2.1 and 2.2.**\n"
|
|
]
|
|
]
|
|
},
|
|
},
|
|
{
|
|
{
|
|
@@ -1552,46 +1553,11 @@
|
|
"* Types (PEFT, LoRA, QLoRA)\n",
|
|
"* Types (PEFT, LoRA, QLoRA)\n",
|
|
"* Using PyTorch for Pre-Training & Fine-Tuning\n",
|
|
"* Using PyTorch for Pre-Training & Fine-Tuning\n",
|
|
"\n",
|
|
"\n",
|
|
- "* Evals + Quality\n"
|
|
|
|
- ]
|
|
|
|
- },
|
|
|
|
- {
|
|
|
|
- "cell_type": "code",
|
|
|
|
- "execution_count": 7,
|
|
|
|
- "metadata": {
|
|
|
|
- "colab": {
|
|
|
|
- "base_uri": "https://localhost:8080/",
|
|
|
|
- "height": 79
|
|
|
|
- },
|
|
|
|
- "executionInfo": {
|
|
|
|
- "elapsed": 327,
|
|
|
|
- "status": "ok",
|
|
|
|
- "timestamp": 1695832272878,
|
|
|
|
- "user": {
|
|
|
|
- "displayName": "Amit Sangani",
|
|
|
|
- "userId": "11552178012079240149"
|
|
|
|
- },
|
|
|
|
- "user_tz": 420
|
|
|
|
- },
|
|
|
|
- "id": "0a9CvJ8YcTzV",
|
|
|
|
- "outputId": "56a6d573-a195-4e3c-834d-a3b23485186c"
|
|
|
|
- },
|
|
|
|
- "outputs": [
|
|
|
|
- {
|
|
|
|
- "data": {
|
|
|
|
- "text/html": [
|
|
|
|
- "<img src=\"https://mermaid.ink/img/CiAgZ3JhcGggTFI7CiAgICAgIEN1c3RvbV9EYXRhc2V0IC0tPiBQcmUtdHJhaW5lZF9MbGFtYQogICAgICBQcmUtdHJhaW5lZF9MbGFtYSAtLT4gRmluZS10dW5lZF9MbGFtYQogICAgICBGaW5lLXR1bmVkX0xsYW1hIC0tPiBSTEhGCiAgICAgIFJMSEYgLS0+IHxMb3NzOkNyb3NzLUVudHJvcHl8RmluZS10dW5lZF9MbGFtYQogICAgICBjbGFzc0RlZiBkZWZhdWx0IGZpbGw6I0NDRTZGRixzdHJva2U6Izg0QkNGNSx0ZXh0Q29sb3I6IzFDMkIzMyxmb250RmFtaWx5OnRyZWJ1Y2hldCBtczsKICA=\"/>"
|
|
|
|
- ],
|
|
|
|
- "text/plain": [
|
|
|
|
- "<IPython.core.display.Image object>"
|
|
|
|
- ]
|
|
|
|
- },
|
|
|
|
- "metadata": {},
|
|
|
|
- "output_type": "display_data"
|
|
|
|
- }
|
|
|
|
- ],
|
|
|
|
- "source": [
|
|
|
|
- "fine_tuned_arch()"
|
|
|
|
|
|
+ "* Evals + Quality\n",
|
|
|
|
+ "\n",
|
|
|
|
+ "Examples of Fine-Tuning:\n",
|
|
|
|
+ "* [Meta Llama Recipes](https://github.com/meta-llama/llama-recipes/tree/main/recipes/finetuning)\n",
|
|
|
|
+ "* [Hugging Face fine-tuning with Llama 3](https://huggingface.co/blog/llama3#fine-tuning-with-%F0%9F%A4%97-trl)\n"
|
|
]
|
|
]
|
|
},
|
|
},
|
|
{
|
|
{
|