Skip to main content
Deprecated

LLMForge is being deprecated: The Ray Team is consolidating around open source fine-tuning solutions. Llama Factory and Axolotl provide enhanced functionality (quantization, advanced algorithms) and native Ray support for scaling. See the migration guide for transitioning your workflows.

LLMForge Releases

LLMForge versionImage URIRay versionPython versionChangelog
0.5.10localhost:5555/anyscale/llm-forge:0.5.102.40.03.11.10Link
0.5.9localhost:5555/anyscale/llm-forge:0.5.92.40.03.11.10Link
0.5.8localhost:5555/anyscale/llm-forge:0.5.82.38.03.11.10Link
0.5.7localhost:5555/anyscale/llm-forge:0.5.72.35.03.11.9Link
0.5.6localhost:5555/anyscale/llm-forge:0.5.62.35.03.11.9Link
0.5.5localhost:5555/anyscale/llm-forge:0.5.52.32.03.11.9Link
0.5.4localhost:5555/anyscale/llm-forge:0.5.42.32.03.11.9-
0.5.3localhost:5555/anyscale/llm-forge:0.5.32.32.03.11.9-
0.5.2localhost:5555/anyscale/llm-forge:0.5.22.32.03.11.9-
0.5.1localhost:5555/anyscale/llm-forge:0.5.12.31.03.11.9-
0.5.0.1localhost:5555/anyscale/llm-forge:0.5.0.1-ngmM6BdcEdhWo0nvedP7janPLKS9Cdz22.24.03.10.14-

Changelog

Version 0.5.10

  • Added support for customizing the fine-tuned model's tag with custom_model_tag config option.
  • Added an eval_first flag allowing for running an iteration of the evaluation loop (+ logging metrics) before training
  • The assistant role is now optional in each sample. This is helpful for use cases like continued pretraining.
  • General improvements in logging and bug fixes

Version 0.5.9

  • Added experimental support for vision-language training (supports Pixtral architecture for now)
  • Added new optimizer_config option for custom HuggingFace optimizer configurations [API Reference]
  • Upgraded rayturbo to 2.40
  • General improvements and bug fixes

Version 0.5.8

  • Added new checkpoint_and_evaluation_frequency for customizing checkpointing and evaluation cadence.
  • Upgraded rayturbo to 2.38, which comes with many improvements in Ray Data
  • Added support for torch.compile [API Reference]
  • General improvements in data preprocessing and logging

⚠️ Deprecation notice ⚠️

  • The parameters checkpoint_every_n_epochs and max_num_checkpoints are deprecated and will be removed in version 0.6. Please use checkpoint_and_evaluation_frequency instead.

Version 0.5.7

  • Added support for Llama 3.1 and 3.2 with default configurations
  • Improved checkpoint discovery for serving - storage_uri returned by anyscale.llm.model.get now contains forwarded checkpoint for LoRA. [User guide]
  • Renamed no_gradient_checkpoint to gradient_checkpointing
  • General improvements in configuration management and stability

⚠️ Deprecation notice ⚠️

  • The no_gradient_checkpoint parameter was confusing to use and is now deprecated. This will be removed in version 0.6. Please use gradient_checkpointing instead.

Version 0.5.6

Version 0.5.5