Commit History
feat(doc): add pillow to lambda instructions (#445)
919f4ca
unverified
support user defined prompters, pretokenized datasets in config, local parquet, local arrow files (#348)
d2e7f27
unverified
docs(readme): add `cd axolotl` (#440)
d21318d
unverified
use save_strategy from config if available (#434)
b3f5e00
unverified
flash attn pip install (#426)
cf66547
unverified
Fix(docs): Remove gptq+lora and fix xformer compat list (#423)
3d1f203
unverified
hopefully improve the README (#419)
2495909
unverified
Merge pull request #413 from mhenrichsen/chore/update-deepseed-config
f806e86
unverified
Feat(doc): Add lr_quadratic_warmup to readme (#412)
2b990eb
unverified
update path to align with fsdp example
bd8cab4
Fix(config): Update handling of deepspeed config (#404)
c01015f
unverified
Fix(docs): Update flash attn requirements (#409)
72fe3f8
unverified
update docs for tokenizer_legacy (#401)
47961fd
unverified
Feat(doc): Add how to save by epochs (#396)
be294fd
unverified
Feat(doc): Add max_steps to readme (#389)
41ecb45
unverified
Feat(config): Add hub_strategy (#386)
73a0b6e
unverified
Feat(doc): Improve sharegpt doc (#378)
729c299
unverified
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
Add wandb_entity to wandb options, update example configs, update README (#361)
7019509
unverified
Feat: Add rope scaling (#343)
b521206
unverified
Update README.md on pretraining_dataset (#360)
fae6ed8
unverified
Clarify pre-tokenize before multigpu (#359)
94d03c8
unverified
note pattern when using groups
b4d1d22
update comment for group_by_length
9f99104
python 3.10 and 3.11 both work fine, as does pytorch 2.1.0.dev
58d6659
there is no configs folder
cc7e800
update README for updated docker images (#328)
41a4d15
unverified
Merge pull request #306 from ethanhs/xgen
dcdec44
unverified
don't resize embeddings to multiples of 32x by default
1066751
Add XGen info to README and example config
3881143
Ethan Smith
commited on