Skip to content

v0.1.3

Compare
Choose a tag to compare
@oelachqar oelachqar released this 28 Jan 00:44
· 227 commits to main since this release
86124a9

What's Changed

  • Documentation: Judge | Custom Model page by @kaisopos in #1195
  • [WIP] Add a notebook for using CNN with custom dataset by @xrdaukar in #1196
  • [Cherrypick for launch] Evaluate: return dict of results by @kaisopos in #1197
  • Configs Train/Infer/Eval and Llama 3.3v (70b) by @optas in #1200
  • Adding an integration test for evaluation fn's output (see PR-1197) by @kaisopos in #1199
  • [docs] Add more details and cross-references related to customization by @xrdaukar in #1198
  • Define single_gpu test marker by @xrdaukar in #1201
  • Native inference: Don't set min_p, temperature in GenerationConfig if sampling is disabled by @xrdaukar in #1202
  • Update tests to make them runnable on GCP by @xrdaukar in #1203
  • Add newline before pformat(train_config) by @xrdaukar in #1204
  • GCP tests launcher script changes by @xrdaukar in #1205
  • [Evaluation] Bug: serialization by @kaisopos in #1207
  • [docs] Add inference snippet for together.ai and DeepSeek APIs by @oelachqar in #1208
  • Exclude multi_gpu tests from GitHub GPU tests by @xrdaukar in #1210
  • Update e2e tests to support multi-GPU machines by @xrdaukar in #1206
  • Add wrappers for remote inference engines by @oelachqar in #1209
  • Vision-Lang & Inference (including LoRA) by @optas in #1174
  • [BugFix] Throw a runtime error for quantized models & inference=VLLM by @kaisopos in #1212
  • Fix most job configs by @wizeng23 in #1213
  • e2e tests update by @xrdaukar in #1216
  • [Notebook] Evaluation with Oumi by @kaisopos in #1218
  • gpt2: move include_performance_metrics param from script to yaml by @xrdaukar in #1217
  • Simplify inference engine API by @oelachqar in #1214
  • Move configs to experimental by @wizeng23 in #1215
  • [docs] Update index page by @oelachqar in #1220
  • Update ConsoleLogger to write to STDOUT by @xrdaukar in #1221
  • Set use_spot to False in our JobConfigs by @wizeng23 in #1222
  • Delete oumi[optional] install target by @wizeng23 in #1224
  • Scaffolding and the first testcase for e2e evaluation tests by @xrdaukar in #1225
  • [docs] Update inference engines doc page by @oelachqar in #1227
  • Clean-up inference engine builder by @oelachqar in #1226
  • [VLLM Engine] Enabling BitsAndBytes quantization by @kaisopos in #1223
  • Add example distillation notebook by @jgreer013 in #1228
  • Add a script to pre-download models for gpu_tests by @xrdaukar in #1231
  • Fix multi-GPU inference integration test by @xrdaukar in #1229
  • [tiny][docs] Update PEFT/LoRA content by @optas in #1233
  • [BugFix] GGUF does not work with VLLM by @kaisopos in #1232
  • Re-enable parallel evaluation for VLM-s by @xrdaukar in #1235
  • Add multimodal exemplar dataset in our provided mini-datasets by @optas in #1234
  • [Tiny] renaming a field name (init_lora_weights) by @optas in #1236
  • Add more e2e evaluation tests by @xrdaukar in #1237
  • Fix pyright breakage when vllm and llama_cpp are not installed by @taenin in #1240
  • Update our oumi launch documentation. by @taenin in #1239
  • Update index.md title for "Join the Community!" by @mkoukoumidis in #1242
  • Update quickstart.md - nit for Oumi support request by @mkoukoumidis in #1241
  • [VLLM Engine] Improve support for GGUF models (incl. auto-download) by @kaisopos in #1238
  • Update README.md title to "Join the Community!" by @mkoukoumidis in #1243
  • Update quickstart.md by @brragorn in #1251
  • Update quickstart.md by @brragorn in #1253
  • Update quickstart.md by @brragorn in #1252
  • Update quickstart.md by @brragorn in #1250
  • [Minor refactor] Moving model caching to oumi.utils by @kaisopos in #1246
  • Add more details to troubleshooting FAQ by @wizeng23 in #1249
  • Update training_methods.md - Change compute requirement suggestions by @mkoukoumidis in #1245
  • Update train.md - nit description change by @mkoukoumidis in #1244
  • [docs] misc docs feedback by @oelachqar in #1248
  • [tiny] Qwen2-VL activate experimental datapipes by @optas in #1247
  • Update Oumi - A Tour.ipynb by @brragorn in #1254
  • [docs] more docs feedback by @oelachqar in #1255
  • Update supported_models.md by @penfever in #1256
  • Rename experimental_use_torch_datapipes data param by @xrdaukar in #1257
  • Add pypi release workflow using testpypi by @oelachqar in #1259
  • Update workflow names by @oelachqar in #1262
  • Update default idle_minutes_to_autostop to 1 hour. by @taenin in #1264
  • update pypi release workflow to use trusted env by @oelachqar in #1265
  • Add padding_side param to internal model config by @xrdaukar in #1260
  • Documentation: Updates on Evaluation/Judge (based on Manos' feedback) by @kaisopos in #1261
  • [tiny] less strict requirements by @oelachqar in #1266
  • Add Deepseek R1 Distill Llama 8B/70B configs by @wizeng23 in #1263
  • Update index.md to highlight beta stage by @mkoukoumidis in #1268
  • Update README.md to highlight beta stage by @mkoukoumidis in #1267
  • Disable pre-release packages by @oelachqar in #1270
  • Update common_workflows.md - Clarify OpenAI is just an example by @mkoukoumidis in #1271
  • Documentation: Evaluation page (update to highlight multi-modal) by @kaisopos in #1269
  • Update launch.md by @taenin in #1272
  • Add pypi release workflow by @oelachqar in #1273
  • Documentation: Judge | minor edit (bold) by @kaisopos in #1274

Full Changelog: v0.1.2...v0.1.3