Researchers propose a self-distillation fix for ‘catastrophic forgetting’ in LLMs

A new fine-tuning technique aims to solve “catastrophic forgetting,” a limitation that often complicates repeated model updates in enterprise deployments. Researchers at MIT, the Improbable AI Lab, and ETH Zurich have introduced a fine-tuning method designed to let models learn new tasks while preserving previously acquired capabilities. To prevent degrading existing capabilities, many organizations isolate…

Read More

Databricks adds MemAlign to MLflow to cut cost and latency of LLM evaluation

Databricks’ Mosaic AI Research team has added a new framework, MemAlign, to MLflow, its managed machine learning and generative AI lifecycle development service. MemAlign is designed to help enterprises lower the cost and latency of training LLM-based judges, in turn making AI evaluation scalable and trustworthy enough for production deployments. The new framework, according to…

Read More