Holidays season release! π
This Optimum TPU release comes in with a larger support for models, in particular newer Llamas π¦ for serving and fine-tuning, as well as initial support for the all recent TPU v6e
and few fixes here and there.
What's Changed
- fix(ci): correct TGI_VERSION definition in workflow by @tengomucho in #122
- Fix nightlies again by @tengomucho in #123
- βοΈ Fix Integration Test for TGI by @baptistecolle in #124
- π Fix repetition penalty by @tengomucho in #125
- Allow sharding fine tuned misaligned models by @tengomucho in #126
- π¦ Newer Llamas support by @tengomucho in #129
- π¦ Add llama fine-tuning notebook example by @baptistecolle in #130
- doc(v6e): mention initial v6e support by @tengomucho in #131
- βοΈ Refactor TGI Dockerfile to support Google-Cloud-Containers as a target by @baptistecolle in #127
- π Fix the convergence of loss function for the llama fine tuning example by @baptistecolle in #132
- chore: update version to v0.2.3 by @tengomucho in #133
Full Changelog: v0.2.1...v0.2.3