|
@@ -1,12 +1,12 @@
|
|
|
-# Llama-On-Perm-Benchmark
|
|
|
-This folder contains code to run inference benchmark for Llama 2 models on-perm with popular serving frameworks.
|
|
|
+# Llama-On-Prem-Benchmark
|
|
|
+This folder contains code to run inference benchmark for Llama 2 models on-prem with popular serving frameworks.
|
|
|
The benchmark will focus on overall inference **throughput** for running containers on one instance (single or multiple GPUs) that you can acquire from cloud service providers such as Azure and AWS. You can also run this benchmark on local laptop or desktop.
|
|
|
We support benchmark on these serving framework:
|
|
|
* [vLLM](https://github.com/vllm-project/vllm)
|
|
|
|
|
|
|
|
|
# vLLM - Getting Started
|
|
|
-To get started, we first need to deploy containers on-perm as a API host. Follow the guidance [here](https://github.com/facebookresearch/llama-recipes/blob/main/demo_apps/llama-on-prem.md#setting-up-vllm-with-llama-2) to deploy vLLM on-perm.
|
|
|
+To get started, we first need to deploy containers on-prem as a API host. Follow the guidance [here](https://github.com/facebookresearch/llama-recipes/blob/main/demo_apps/llama-on-prem.md#setting-up-vllm-with-llama-2) to deploy vLLM on-prem.
|
|
|
Note that in common scenario which overall throughput is important, we suggest you prioritize deploying as many model replicas as possible to reach higher overall throughput and request-per-second (RPS), comparing to deploy one model container among multiple GPUs for model parallelism. Additionally, as deploying multiple model replicas, there is a need for a higher level wrapper to handle the load balancing which here has been simulated in the benchmark scripts.
|
|
|
For example, we have an instance from Azure that has 8xA100 80G GPUs, and we want to deploy the Llama 2 70B chat model, which is around 140GB with FP16. So for deployment we can do:
|
|
|
* 1x70B model parallel on 8 GPUs, each GPU RAM takes around 17.5GB for loading model weights.
|