Browse Source

Update README.md

sekyonda 1 year ago
parent
commit
80d93b632f
1 changed files with 7 additions and 6 deletions
  1. 7 6
      inference/README.md

+ 7 - 6
inference/README.md

@@ -1,10 +1,11 @@
 # Inference
 
-For inference we have provided an [inference script](inference.py). Depending on the type of finetuning performed during training the [inference script](inference.py) takes different arguments.
-To finetune all model parameters the output dir of the training has to be given as --model_name argument.
-In the case of a parameter efficient method like lora the base model has to be given as --model_name and the output dir of the training has to be given as --peft_model argument.
-Additionally, a prompt for the model in the form of a text file has to be provided. The prompt file can either be piped through standard input or given as --prompt_file parameter.
+So far, we have provided support for three methods of inference:
 
-For other inference options, you can use the [vLLM_inference.py](vLLM_inference.py) script for vLLM or review the [hf-text-generation-inference](hf-text-generation-inference/README.md) folder for TGI.
+1. [inference script](inference.py) script provides support for Hugging Face accelerate and PEFT fine tuned models.
 
-For more information including inference safety checks, examples and other inference options available to you, see the inference documentation [here](../docs/inference.md).
+2. [vLLM_inference.py](vLLM_inference.py) script takes advantage of vLLM's paged attention concept for low latency.
+
+3. The [hf-text-generation-inference](hf-text-generation-inference/README.md) folder contains information on Hugging Face Text Generation Inference (TGI).
+
+For more in depth information on inference including inference safety checks and examples, see the inference documentation [here](../docs/inference.md).