What's Changed
- Refactor position encoding configuration by @vince62s in #60
- fix update vocab by @vince62s in #63
- bfloat16 support, and an attempt at homogenizing model_dtype & precision by @francoishernandez in #54
- Fix prefix and suffix transforms - avoid adding empty suffix or prefix by @sersh88 in #57
- fix the incorrect dockerimages in the ReadMe by @aaaallleen in #68
- Remove unnecessary optim in convert_HF by @francoishernandez in #71
- Add onmt_config converter to facilitate switch by @francoishernandez in #69
- Update some FAQ sections by @francoishernandez in #74
- Added TER and BLEU for early stopping by @aaaallleen in #73
- [fix] fix normalize and clean transforms config management by @francoishernandez in #87
- [docs] Fix quickstart config and command by @francoishernandez in #90
- add head_dim setting when diff from hidden // heads by @vince62s in #78
- Some MHA and RoPE refactoring, llama-3.1 rope_scaling by @francoishernandez in #91
- Fixed variable referenced before assignment when position_embeddings is None error by @dameikle in #95
- Send src_pad_mask and tgt_pad_mask to decoder in _align_forward by @dameikle in #96
- Fixdistrib by @vince62s in #100
- fix added tokens by @vince62s in #101
- Support mapped tokens eg: <im_start> ==> ⦅im_start⦆in inference.yaml … by @vince62s in #102
- add wmt22 recipes with TowerInstruct and Llama3.1 LLMs by @vince62s in #103
- Remove duplicate sentencepiece requirement by @francoishernandez in #104
- [patch] Adapt some warning behaviours for reduced verbosity by @francoishernandez in #105
- [patch] Update precision to compute_dtype in forgotten places by @francoishernandez in #106
- Inference server, lots of related changes by @francoishernandez in #42
New Contributors
- @sersh88 made their first contribution in #57
- @aaaallleen made their first contribution in #68
- @dameikle made their first contribution in #95
Full Changelog: 0.0.1...0.0.2