Beto 6f53f26e05 Merge branch 'main' into l3p/llama_guard il y a 11 mois
..
__init__.py 207d2f80e9 Make code-llama and hf-tgi inference runnable as module il y a 1 an
chat_utils.py 6d9d48d619 Use apply_chat_template instead of custom functions il y a 11 mois
checkpoint_converter_fsdp_hf.py ce9501f22c remove relative imports il y a 1 an
llm.py a404c9249c Notebook to demonstrate using llama and llama-guard together using OctoAI il y a 1 an
model_utils.py d51d2cce9c adding sdpa for flash attn il y a 1 an
prompt_format_utils.py bcdb5b31fe Fixing quantization config. Removing prints il y a 11 mois
safety_utils.py f63ba19827 Fixing tokenizer used for llama 3. Changing quantization configs on safety_utils. il y a 11 mois