Skip to content

Releases: oumi-ai/oumi

v0.1.4

03 Feb 21:06
fc3d45e
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.1.3...v0.1.4

v0.1.3

28 Jan 00:44
86124a9
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.1.2...v0.1.3

v0.1.2.3

27 Jan 22:30
5a9564a
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.1.2.2...v0.1.2.3

v0.1.2.0-alpha

17 Jan 17:31
0016ff8
Compare
Choose a tag to compare
v0.1.2.0-alpha Pre-release
Pre-release

What's Changed

Read more

v0.1.1.0-alpha.1

08 Jan 04:27
9ff5132
Compare
Choose a tag to compare
v0.1.1.0-alpha.1 Pre-release
Pre-release

What's Changed

Read more

Initial release

02 Oct 22:03
d14a4db
Compare
Choose a tag to compare
Initial release Pre-release
Pre-release

What's Changed

  • Add python project configs by @oelachqar in #1
  • Add repo skeleton by @oelachqar in #2
  • Export lema entrypoint scripts by @oelachqar in #3
  • Update static type checking config by @oelachqar in #5
  • Add example jupyter / colab notebook by @oelachqar in #4
  • Refactor config parsing to use omegaconf by @oelachqar in #6
  • Updating documentation (Dev Environment Setup) by @kaisopos in #7
  • Add tests and vscode config by @oelachqar in #8
  • Added DPOTrainer example to repo, as well as cuda device cleanup to training loop by @jgreer013 in #9
  • Adding torch as top-level module dependency by @optas in #10
  • Add configs for specific hardware requirements by @jgreer013 in #11
  • Sort pre-commit hooks lexicographically by @xrdaukar in #12
  • Add logging config by @oelachqar in #13
  • Lema inference by @xrdaukar in #14
  • Panos dev by @optas in #16
  • Add job launcher by @oelachqar in #15
  • Making split of data a flexible variable by @optas in #17
  • Configure max file size in precommit hooks by @xrdaukar in #18
  • Minor bugfix and documentation update by @oelachqar in #19
  • adding pynvml to train env by @kaisopos in #20
  • Panos dev by @optas in #22
  • Augmenting Types for training hyperparams by @optas in #23
  • Train refactoring (config file visibility) + a few minor changes by @kaisopos in #21
  • Minimal test for train function by @xrdaukar in #25
  • Fix leftover '_torch_dtype' in 'ModelParams' by @xrdaukar in #26
  • Update GPU types list in the default SkyPilot config by @xrdaukar in #27
  • Add a missing lema-infer command under [project.scripts] by @xrdaukar in #28
  • add basic pytests for evaluate and infer by @xrdaukar in #29
  • Update README and pyproject.toml by @wizeng23 in #30
  • A helper function to print info about available CUDA devices by @xrdaukar in #31
  • Update SkyPilot cconfig to start using torchrun by @xrdaukar in #32
  • Support basic single-node, multi-gpu training by @xrdaukar in #33
  • Run all precommit hooks on the repo by @xrdaukar in #35
  • Add experimental code for llama cpp inference by @jgreer013 in #37
  • Create skeleton of STYLE_GUIDE.md by @xrdaukar in #36
  • Adding support for training custom models (for now just a dummy model). by @kaisopos in #38
  • Fix custom model name in test_train.py by @xrdaukar in #39
  • Configure pyright (static type checker) and resolve existing type errors to make it pass by @xrdaukar in #41
  • fix trailing whitespace warning in STYLE_GUIDE.md by @xrdaukar in #43
  • Configure initial GitHub Actions workflow to run pre-commits and tests by @xrdaukar in #44
  • A variety of proposed extensions to finetune a chat-based model (starting with Zephyr) by @optas in #34
  • Fix syntax error in ultrachat by @xrdaukar in #48
  • Create initial version of CONTRIBUTING.md by @xrdaukar in #46
  • Reduce the number of training steps from 5 to 3 to make test_train.py faster by @xrdaukar in #49
  • Adding registry for custom models. by @kaisopos in #42
  • Add config and streaming args to DataParams by @wizeng23 in #47
  • Update Pre-review Tests to only run on pull_request by @xrdaukar in #50
  • Add training flags to computes tokens-based stats by @xrdaukar in #51
  • reduce test training steps in another test which I missed before by @xrdaukar in #53
  • Rename var names of *Params classes by @wizeng23 in #52
  • Make some NVIDIA-specific dependencies optional by @xrdaukar in #54
  • fix trl version as 0.8.6 by @xrdaukar in #56
  • Remove reference to torch.cuda.clock_rate by @xrdaukar in #57
  • Update inference to support non-interactive batch mode. by @kaisopos in #58
  • Update README.md to include Linux/WSL specific instructions by @xrdaukar in #59
  • Minor formatting improvements in README.md by @xrdaukar in #60
  • Minor: Updating Lora Params by @optas in #55
  • Support dataset packing by @wizeng23 in #63
  • Disallow relative imports in LeMa by @xrdaukar in #65
  • Add text_col param that's required for SFTTrainer by @wizeng23 in #66
  • Refactor common config parsing logic (YAML, arg_list) into a common util by @xrdaukar in #68
  • Standardize test naming convention by @wizeng23 in #69
  • Adding support for a hardcoded evaluation with MMLU. by @kaisopos in #67
  • Minor changes to the default configs/skypilot/sky.yaml config by @xrdaukar in #71
  • Prototype to pass config.model.model_max_length to Trainers by @xrdaukar in #70
  • [Inference] Remove the prepended prompts from model responses. by @kaisopos in #73
  • Add a util to print versioning info by @xrdaukar in #74
  • Switch to tempfile.TemporaryDirectory() in test_train.py by @xrdaukar in #75
  • Update docstring verbs to descriptive form by @wizeng23 in #76
  • Add sample accelerate and fsdp configs by @xrdaukar in #77
  • Refactor code to get device rank and world size into a helper function by @xrdaukar in #79
  • Add a simple util to print model summary e.g., layer names, architecture summary by @xrdaukar in #80
  • Freeze numpy to pre 2.0 version by @xrdaukar in #81
  • Adding inference support for next logit probability. by @kaisopos in #78
  • Create FSDP configs for Phi3 by @xrdaukar in #82
  • Auto-format pyproject.toml with "Even Better TOML" by @xrdaukar in #83
  • Minor cleanup updates to SkyPilot configs by @xrdaukar in #84
  • Mixed Precision Training, Flash-Attention-2, Print-trainable-params by @optas in #85
  • Update README.md to include basic instructions for multi-GPU training (DDP, FSDP) by @xrdaukar in #86
  • Start using $SKYPILOT_NUM_GPUS_PER_NODE in SkyPilot config by @xrdaukar in #90
  • Add configs for FineWeb Llama2 pretraining by @wizeng23 in #89
  • Quantization by @optas in #87
  • Update the default SkyPilot config to print more debug/context info by @xrdaukar in #92
  • Add license by @oelachqar in #93
  • Initial version of SkyPilot config for multi-node training (num_nodes: N) by @xrdaukar in #94
  • MMLU eval refactor. by @kaisopos in #88
  • Remove comparison between LOCAL_RANK and RANK by @xrdaukar in #96
  • Handling the loading of peft adapters and other minor issues (e.g., adding more logging parameters) by @optas in #91
  • Update configs/skypilot/sky_llama2b.yaml to start using sky_init.sh by @xrdaukar in #97
  • Add bool param to resume training from the last known checkpoint (if exists) by @xrdaukar in #99
  • Inference: save/restore probabilities to/from file. by @kaisopos in #98
  • Add support for dataset mixtures during training by @taenin in #95
  • Add train, test, and validation splits to the LeMa config. by @taenin in #101
  • nanoGPT (GPT2) pretraining recipe by @wizeng23 in #103
  • Minor: Updates on Zephyr-Config by @optas in https://githu...
Read more