Commit History

Autor SHA1 Mensaxe Data
  Matthias Reso 1437b917d0 Fix lint error hai 1 ano
  Matthias Reso 34e45490ba Exchange micro_batching_size against gradient_accumulation_steps in docs hai 1 ano
  Matthias Reso 47ae6d0326 Remove print as it breaks progress bar and update progress bar description instead hai 1 ano
  Matthias Reso 9def4fbafd Remove micro_batch_training parameter and replace with gradient_accumulation_steps hai 1 ano
  Geeta Chauhan fbc513ec47 adding notes how to get the HF models (#151) hai 1 ano
  Hamid Shojanazeri bcfafd9a0b adding notes how to get the HF models hai 1 ano
  Geeta Chauhan cfba150311 adding llama code inference (#144) hai 1 ano
  Hamid Shojanazeri 6105a3f886 clarifying the infilling use-case hai 1 ano
  Hamid Shojanazeri 8b0008433c fix typos hai 1 ano
  Hamid Shojanazeri 564ef2f628 remove padding logic hai 1 ano
  Hamid Shojanazeri 277a292fbc adding autotokenizer hai 1 ano
  Hamid Shojanazeri 3f2fb9167e adding notes to model not supporting infilling hai 1 ano
  Hamid Shojanazeri c62428b99c setting defaults of temp and top_p hai 1 ano
  Hamid Shojanazeri c014ae7cb8 setting BT option to true hai 1 ano
  Hamid Shojanazeri 4fa44e16d9 add note for python llama not suited for llama infilling hai 1 ano
  Hamid Shojanazeri b18a186385 removing the option to take prompt from cli hai 1 ano
  Hamid Shojanazeri 75991d8795 fix the extra line added and remove take prompt from cli hai 1 ano
  Hamid Shojanazeri d28fc9898a addressing doc comments hai 1 ano
  Hamid Shojanazeri a234d1fe0c fix typos hai 1 ano
  Hamid Shojanazeri 2d9f4796e8 fixing the output format hai 1 ano
  Hamid Shojanazeri 1e8ea70b26 adding llama code inference hai 1 ano
  Geeta Chauhan 82e05c46e0 fix a bug in the config for use_fast_kernels (#121) hai 1 ano
  Hamid Shojanazeri 971c079aa6 bugfix: remove duplicate load_peft_model (#124) hai 1 ano
  hongbo.mo fcc817e923 bugfix: remove duplicate load_peft_model hai 1 ano
  Brian Vaughan 3faf005226 fix a bug in the config for use_fast_kernels hai 1 ano
  Geeta Chauhan 03faba661f Update paddings (#85) hai 1 ano
  Geeta Chauhan 205e5a4b81 save cpu mem by leveraging FSDP rank0 broadcasting (#77) hai 1 ano
  Hamid Shojanazeri 85a4ed1b65 Merge branch 'main' into update_paddings hai 1 ano
  lchu feaa344af3 resolve conflicts hai 1 ano
  Geeta Chauhan 3f1fef7a00 adding flash attention and xformer memory efficient through PT SDPA (#97) hai 1 ano