What's Changed
- Documentation: Judge | Custom Model page by @kaisopos in #1195
- [WIP] Add a notebook for using CNN with custom dataset by @xrdaukar in #1196
- [Cherrypick for launch] Evaluate: return dict of results by @kaisopos in #1197
- Configs Train/Infer/Eval and Llama 3.3v (70b) by @optas in #1200
- Adding an integration test for evaluation fn's output (see PR-1197) by @kaisopos in #1199
- [docs] Add more details and cross-references related to customization by @xrdaukar in #1198
- Define
single_gpu
test marker by @xrdaukar in #1201 - Native inference: Don't set
min_p
,temperature
inGenerationConfig
if sampling is disabled by @xrdaukar in #1202 - Update tests to make them runnable on GCP by @xrdaukar in #1203
- Add newline before
pformat(train_config)
by @xrdaukar in #1204 - GCP tests launcher script changes by @xrdaukar in #1205
- [Evaluation] Bug: serialization by @kaisopos in #1207
- [docs] Add inference snippet for together.ai and DeepSeek APIs by @oelachqar in #1208
- Exclude
multi_gpu
tests from GitHub GPU tests by @xrdaukar in #1210 - Update e2e tests to support multi-GPU machines by @xrdaukar in #1206
- Add wrappers for remote inference engines by @oelachqar in #1209
- Vision-Lang & Inference (including LoRA) by @optas in #1174
- [BugFix] Throw a runtime error for quantized models & inference=VLLM by @kaisopos in #1212
- Fix most job configs by @wizeng23 in #1213
- e2e tests update by @xrdaukar in #1216
- [Notebook] Evaluation with Oumi by @kaisopos in #1218
- gpt2: move
include_performance_metrics
param from script to yaml by @xrdaukar in #1217 - Simplify inference engine API by @oelachqar in #1214
- Move configs to experimental by @wizeng23 in #1215
- [docs] Update index page by @oelachqar in #1220
- Update ConsoleLogger to write to STDOUT by @xrdaukar in #1221
- Set
use_spot
to False in our JobConfigs by @wizeng23 in #1222 - Delete
oumi[optional]
install target by @wizeng23 in #1224 - Scaffolding and the first testcase for e2e evaluation tests by @xrdaukar in #1225
- [docs] Update inference engines doc page by @oelachqar in #1227
- Clean-up inference engine builder by @oelachqar in #1226
- [VLLM Engine] Enabling BitsAndBytes quantization by @kaisopos in #1223
- Add example distillation notebook by @jgreer013 in #1228
- Add a script to pre-download models for
gpu_tests
by @xrdaukar in #1231 - Fix multi-GPU inference integration test by @xrdaukar in #1229
- [tiny][docs] Update PEFT/LoRA content by @optas in #1233
- [BugFix] GGUF does not work with VLLM by @kaisopos in #1232
- Re-enable parallel evaluation for VLM-s by @xrdaukar in #1235
- Add multimodal exemplar dataset in our provided mini-datasets by @optas in #1234
- [Tiny] renaming a field name (
init_lora_weights
) by @optas in #1236 - Add more e2e evaluation tests by @xrdaukar in #1237
- Fix pyright breakage when vllm and llama_cpp are not installed by @taenin in #1240
- Update our oumi launch documentation. by @taenin in #1239
- Update index.md title for "Join the Community!" by @mkoukoumidis in #1242
- Update quickstart.md - nit for Oumi support request by @mkoukoumidis in #1241
- [VLLM Engine] Improve support for GGUF models (incl. auto-download) by @kaisopos in #1238
- Update README.md title to "Join the Community!" by @mkoukoumidis in #1243
- Update quickstart.md by @brragorn in #1251
- Update quickstart.md by @brragorn in #1253
- Update quickstart.md by @brragorn in #1252
- Update quickstart.md by @brragorn in #1250
- [Minor refactor] Moving model caching to
oumi.utils
by @kaisopos in #1246 - Add more details to troubleshooting FAQ by @wizeng23 in #1249
- Update training_methods.md - Change compute requirement suggestions by @mkoukoumidis in #1245
- Update train.md - nit description change by @mkoukoumidis in #1244
- [docs] misc docs feedback by @oelachqar in #1248
- [tiny] Qwen2-VL activate experimental datapipes by @optas in #1247
- Update Oumi - A Tour.ipynb by @brragorn in #1254
- [docs] more docs feedback by @oelachqar in #1255
- Update supported_models.md by @penfever in #1256
- Rename
experimental_use_torch_datapipes
data param by @xrdaukar in #1257 - Add pypi release workflow using testpypi by @oelachqar in #1259
- Update workflow names by @oelachqar in #1262
- Update default idle_minutes_to_autostop to 1 hour. by @taenin in #1264
- update pypi release workflow to use trusted env by @oelachqar in #1265
- Add
padding_side
param to internal model config by @xrdaukar in #1260 - Documentation: Updates on Evaluation/Judge (based on Manos' feedback) by @kaisopos in #1261
- [tiny] less strict requirements by @oelachqar in #1266
- Add Deepseek R1 Distill Llama 8B/70B configs by @wizeng23 in #1263
- Update index.md to highlight beta stage by @mkoukoumidis in #1268
- Update README.md to highlight beta stage by @mkoukoumidis in #1267
- Disable pre-release packages by @oelachqar in #1270
- Update common_workflows.md - Clarify OpenAI is just an example by @mkoukoumidis in #1271
- Documentation: Evaluation page (update to highlight multi-modal) by @kaisopos in #1269
- Update launch.md by @taenin in #1272
- Add pypi release workflow by @oelachqar in #1273
- Documentation: Judge | minor edit (bold) by @kaisopos in #1274
Full Changelog: v0.1.2...v0.1.3