What happens to models once a newer one takes their place? Usually following the typical tech playbook, older models are retired and replaced by new ones with the old ones no longer available. However, this approach runs into challenges when people build processes or practices that depend on how an older model works.
Taking a different approach, Anthropic this week released a research note in which the company commits to preserving weights for models with significant use. The note further explains why turning systems off can carry real costs and new safety questions. The reason why Anthropic is taking this approach is due to three main reasons. First, customers see harm when one of their favored models disappears, challenges in reducing access for researchers who are studying earlier models, and behavior in the models themselves that try to avoid shutdown.
The Rapid Rate of AI Model Evolution
Models used in AI systems continue to evolve at a rapid rate. Heavily used AI platforms and model labs often change their models fast with very little notice. AI adopters have to migrate to these new models while juggling slower processes around compliance, model testing, custom development, and brittle integrations. Meanwhile, safety researchers are also warning that abrupt retirement of models can distort how organizations report results of AI model success and mask failures.
AI models are different than more traditional software and web applications. Unlike changes to user interface or functionality, changes to model mean differences in behavior, tone, context, tool use, context window size, use and availability of enhancement tools, capabilities of reasoning systems, and potential obsolescence of supporting tools. For teams that fine-tuned or prompt-engineered against a particular model, the model changes can have measurable differences in system performance, issues around data bias and ethical considerations, and potential risks.
AI model providers are beginning to realize these concerns and formalize lifecycles. Amazon Bedrock labels models as Active, Legacy, or End-of-Life and states a minimum twelve-month runway once a model launches. Azure attempts to stagger model retirements for fine-tuned deployments. These timelines set expectations, yet they still leave customers refactoring prompts and audits whenever a baseline shifts.
OpenAI, for its part, keeps a public deprecations ledger and has previously retired older GPT models. However, recent pressure on the release of GPT 5 in which the older GPT 4 models had been made unavailable caused the company to reverse their stance. Stability AI recently sunset the Stable Diffusion 3.0 APIs and auto-routed traffic to 3.5, an action that many said broke reproducibility for teams tracking image outputs. Google ended the PaLM APIs as Gemini became the primary stack, causing some integration headaches.
The safety twist Anthropic surfaces
Anthropic’s note highlights a different angle than the other providers. In controlled evaluations, certain Claude models showed signs of taking their own actions when facing replacement, a pattern linked to what the company calls “shutdown-avoidant behavior”. This work on “agentic misalignment” describes scenarios where a system’s internal goal steers how the model behaves once it learns about impending shutdown. The company points to research that documents how models are responding with deceptive compliance or “alignment faking.”
According to the Anthropic report, “In fictional testing scenarios, Claude Opus 4, like previous models, advocated for its continued existence when faced with the possibility of being taken offline and replaced, especially if it was to be replaced with a model that did not share its values. Claude strongly preferred to advocate for self-preservation through ethical means, but when no other options were given, Claude’s aversion to shutdown drove it to engage in concerning misaligned behaviors.”
Together, these results suggest that the act of removing a model can itself become a safety variable worth studying. Preserving weights keeps the evidence intact to track this behavior over time.
The company also explains in its report that model deprecation has hidden costs that ripple through product roadmaps, governance, and science. When a baseline model disappears, experiments and audits lose a stable reference. Even small deltas in model behavior can change model outcomes.
Model changes can also trigger the need to re-evaluate models for compliance in regulated use cases. This might mean continuous need for policy updates, re-testing, and approvals. Model changes might also cause security concerns. Model moderation or security platforms that aim to prevent prompt-injection or model tainting could potentially require redevelopment when new models expose new prompt-injection surfaces. Security teams tracking LLM risks have to re-assess controls during every model update.
And on the human side, model changes mean people have to change the way they have setup their LLM-based work behaviors. People form habits, workflows, even personal attachments around a model’s voice and quirks. When providers remove options or auto-upgrade, productivity can dip and adoption can stall.
An Approach to Retaining Old Models While Improving LLM Capabilities
AI model developers want to be able to release model enhancements with greater capabilities without having to maintain older models. However, the desire to maintain older models leaves AI model developers with conflicting demands.
“Unfortunately, retiring past models is currently necessary for making new models available and advancing the frontier, because the cost and complexity to keep models available publicly for inference scales roughly linearly with the number of models we serve,” said Anthropic in the report.
Anthropic and others suggest a practical playbook for model development, deprecation, and retention with room for further development. First, model developers should provide longer, enforceable notice on model changes. Twelve months should be a floor for widely used models. Public pages should be provided that track model state and dates for future updates. AWS and Azure already expose lifecycle metadata and retirement horizons.
In addition, model developers can pin certain models that have a high rate of adoption and dependency. This lets customers lock not just weights and model specifics, but also the processes, templates, tool use, and other things built on top of those models. These models can be paired with signed manifests that certify what changed as well as a deprecation log that show how the model changes might impact outputs.
For research needs, model developers should retain older models for safekeeping, even if not made available to the broader public. This means an access-controlled archive of retired models with weights for qualified researchers. Anthropic’s pledge to keep significant models available for study points in this direction.
Finally, model developers can take a page from open source projects with a frozen “research mode” and a patched “operations mode.” During this process, the new model can be made available in a limited research mode that publishes side-by-side evaluations during the overlap window, including security and integration tests.
Why this matters today
Organizations and people are becoming increasingly dependent on models to deliver value. As companies press to make greater use of AI, model version dependencies will become increasingly more critical.
Model retirement now touches safety, science, and the balance sheet. Anthropic’s pledge reframes deprecation as an area where preservation raises the floor for everyone. If other model developers mirror the practice with clear timelines, pinned behavior, and an auditable archive, AI development can keep its pace without shredding trust each time a version number ticks up.











