T
traeai
RSS登录
返回首页
Microsoft Research Blog

AutoAdapt: Automated domain adaptation for large language models

9.0Score
AutoAdapt: Automated domain adaptation for large language models
AI 深度提炼
  • AutoAdapt 可显著减少手动调整领域适配的工作量。
  • 通过自动化流程提升大语言模型在垂直领域的准确性和效率。
  • 该方法结合了迁移学习和无监督学习的优势。
#大语言模型#领域自适应#微软#AI
打开原文

AutoAdapt: Automated domain adaptation for large language models - Microsoft Research

Skip to main content

![Image 6: Microsoft](https://www.microsoft.com/)

Research

Our research

Resources

Research areas: Intelligence

Research areas: Systems

Research areas: Theory

Research areas: Other Sciences

Programs & events

Academic programsEvents & academic conferencesMicrosoft Research Forum

Connect & learn

Behind the Tech podcastMicrosoft Research blogMicrosoft Research ForumMicrosoft Research podcast

About

People & news

Microsoft Research Labs

Other labs

More

Register: Research Forum

All Microsoft

Tech & innovation

Industries

Partners

Resources

View Sitemap

Cancel

Search

Return to Blog Home

Microsoft Research Blog

AutoAdapt: Automated domain adaptation for large language models

Published April 22, 2026

By Sidharth Sinha, Research Fellow Anson Bastos, Senior Researcher Xuchao Zhang, Principal Research Manager Akshay Nambi, Principal Researcher Rujia Wang, Principal Research Product Manager Chetan Bansal, Senior Principal Research Manager

Share this page

  • [Share on Facebook](https://www.facebook.com/sharer/sharer.php?u=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fblog%2Fautoadapt-automated-domain-adaptation-for-large-language-models%2F "Share on Facebook")
  • [Share on X](https://x.com/intent/tweet?text=AutoAdapt%3A%20Automated%20domain%20adaptation%20for%20large%20language%C2%A0models&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fblog%2Fautoadapt-automated-domain-adaptation-for-large-language-models%2F "Share on X")
  • [Share on LinkedIn](https://www.linkedin.com/shareArticle?mini=true&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fblog%2Fautoadapt-automated-domain-adaptation-for-large-language-models%2F&title=AutoAdapt%3A%20Automated%20domain%20adaptation%20for%20large%20language%C2%A0models&summary=AutoAdapt%3A%20Automated%20domain%20adaptation%20for%20large%20language%C2%A0models&source=Microsoft%20Research "Share on LinkedIn")
  • [Share on Reddit](http://www.reddit.com/submit?title=AutoAdapt%3A%20Automated%20domain%20adaptation%20for%20large%20language%C2%A0models&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fblog%2Fautoadapt-automated-domain-adaptation-for-large-language-models%2F "Share on Reddit")
  • [Subscribe to our RSS feed](https://www.microsoft.com/en-us/research/feed/ "Subscribe to our RSS feed")
Image 7: Three white line icons in a row; a document list, a workflow, and process wheel against a blue and purple gradient background.

At a glance

  • **Problem**:Adapting large language models to specialized, high-stakes domains is slow, expensive, and hard to reproduce.
  • **What we built**:AutoAdapt automates planning, strategy selection(e.g., RAG vs. fine-tuning), and tuning under real deployment constraints.
  • **How it works**: A structured configuration graph maps the full scope of the adaptation process, an agentic planner selects and sequences the right steps, and a budget-aware optimization loop (AutoRefine)refines the process within defined constraints.
  • **Why it matters**:The result is faster,automated,more reliable domain adaptation that turns weeks of manual iteration into repeatable pipelines.

Deploying large language models (LLMs) in real-world, high-stakes settings is harder than it should be. In high-stakes settings like law, medicine, and cloud incident response, performance and reliability can quickly break down because adapting models to domain-specific requirements is a slow and manual process that is difficult to reproduce.

The core challenge is domain adaptation, which entails turning a general-purpose model into one that consistently follows domain rules, draws on the right knowledge, and meets constraints such as latency, privacy, and cost. Today, that process typically involves guesswork, choosing among approaches like retrieval-augmented generation (RAG) and fine-tuning, tuning hyperparameters, and iterating through evaluations with no clear path to a good outcome. An operations team responding to an outage can’t afford a model that drifts from domain requirements or a tuning process that takes weeks with no guarantee of a reproducible result.

To tackle this, we’re pleased to introduce AutoAdapt. In our paper, “AutoAdapt: An Automated Domain Adaptation Framework for Large Language Models,” we describe an end-to-end, constraint-aware framework for domain adaptation. Given a task objective, available domain data, and practical requirements like accuracy, latency, hardware, and budget, AutoAdapt plans a valid adaptation pipeline, selecting among approaches like RAG and multiple fine-tuning methods, and tunes key hyperparameters using a budget-aware refinement loop. The result is an executable, reproducible workflow for building domain-ready models more quickly and consistently, helping make LLMs dependable in real-world settings.

Spotlight: Microsoft research newsletter

![Image 8](https://info.microsoft.com/ww-landing-microsoft-research-newsletter.html)

Microsoft Research Newsletter

Stay connected to the research community at Microsoft.

Subscribe today

Opens in a new tab

How it works

AutoAdapt starts from a practical observation: teams don’t just need a better prompt or more data, they need a decision process that reliably maps a task, its domain data, and real constraints to an approach that works. To do this, AutoAdapt treats domain adaptation as a constrained planning problem. Given an objective provided in natural language, dataset size and format, and limits on latency, hardware, privacy, and cost, it provides an end-to-end pipeline that teams can execute and deploy.

Domain adaptation often feels like trial and error because the design space is large and complex. Teams must choose among approaches such as RAG, supervised fine-tuning, parameter-efficient methods (such as LoRA), and alignment steps, each with many hyperparameters. These choices interact in nonobvious ways, and not all combinations are valid, making it difficult to identify a reliable strategy. The problem is compounded by the high cost of LLM training, which limits how many configurations can be explored.

AutoAdapt addresses this with the Adaptation Configuration Graph (ACG), a structured representation of the system’s configuration space that enables efficient search while guaranteeing valid pipelines.

Building on the ACG, AutoAdapt uses a planning agent to make and justify decisions. It proposes strategies, evaluates them against user requirements, and iterates until the plan is feasible and well-grounded. Rather than optimizing in an unconstrained black box, AutoAdapt roots each decision in best practices and explicit constraints, producing an executable workflow with parameter ranges.

Finally, AutoAdapt introduces AutoRefine, a budget-aware refinement loop that optimizes hyperparameters by strategically selecting which experiments to run next, even under limited feedback. AutoRefine replaces weeks of manual tuning with a more disciplined, reproducible process that is easier to audit and compare across projects. In real-world systems such as healthcare documentation, legal workflows, or incident response, this level of rigor is essential. Figure 1 illustrates the end-to-end workflow.

Image 9: A workflow diagram illustrating how a user’s task description and constraints are processed to automatically produce a deployable language model. User inputs are analyzed and refined through multiple stages, including multi‑agent proposal and critique, best‑practice consultation, and iterative pipeline refinement. These stages evaluate task requirements, data choices, and model configurations while verifying user constraints. The process concludes with an executable plan that generates a final model meeting the specified objectives and constraints.

Figure 1. The AutoAdapt workflow, showing how user inputs flow through planning and refinement to produce a deployable model.

Evaluation

In experiments, AutoAdapt consistently identifies effective adaptation strategies and delivers improvements across a range of benchmark and real-world tasks, including reasoning, question answering, coding, classification, and cloud-incident diagnosis. It uses constraint-aware planning and budgeted refinement to find better-performing configurations with minimal added time and cost, making the process practical for production teams. Figures 2 and 3 show aggregate performance against competitive baselines.

Image 10: Three radar plots compare multiple methods across several datasets using success rate, normalized performance score, and a cumulative metric. In all three plots, the AutoAdapt method consistently exhibits larger coverage across most tasks, indicating stronger overall performance. Baseline methods show more uneven profiles, with strengths limited to specific datasets or metrics. The visualization highlights AutoAdapt’s robust and consistent advantage relative to existing approaches.

Figure 2. Success rate (SR), normalized performance score (NPS), and cumulative score (CS) comparing AutoAdapt with baseline methods across datasets. Higher scores indicate better performance, with AutoAdapt outperforming state-of-the-art baselines.

Image 11: Two bar charts compare time and cost overheads for AutoAdapt relative to a default baseline across multiple datasets. AutoAdapt introduces only a small additional time requirement, averaging around half an hour, while achieving noticeable performance improvements. The cost comparison shows a similarly modest increase, with average extra cost remaining low across tasks. Overall, the figure indicates that AutoAdapt delivers performance gains with minimal additional time and financial overhead.

Figure 3. AutoAdapt achieves performance gains with minimal overhead, approximately 30 minutes of additional time and $4 in additional cost.

Implications and looking forward

The broader significance of AutoAdapt is that domain adaptation can become an engineering discipline, not an ad hoc process. By making key choices explicit—what to adapt, how to adapt it, and which constraints the system must satisfy—AutoAdapt helps teams reach results faster, reproduce them more easily, and audit them more rigorously. This shift is especially important in domains where drift from pretrained knowledge is common and failures are costly. When LLMs are used to draft clinical notes, triage support incidents, or summarize regulatory language, organizations need a clear, repeatable path from data to models that behave predictably under latency, privacy, and budget requirements.

Because domain adaptation is a prerequisite for deploying LLMs in real-world settings, we’re making the AutoAdapt frameworkopen source (opens in new tab) to give teams a concrete starting point. The README (opens in new tab) file provides installation and quick-start instructions.

Opens in a new tab

Related publications

[AutoAdapt: An Automated Domain Adaptation Framework for LLMs](https://www.microsoft.com/en-us/research/publication/autoadapt-an-automated-domain-adaptation-framework-for-llms/)

Meet the authors

Image 12: Portrait of Sidharth Sinha

Sidharth Sinha

Research Fellow

Learn more

Image 13: Portrait of Anson Bastos

Anson Bastos

Senior Researcher

Learn more

Image 14: Portrait of Xuchao Zhang

Xuchao Zhang

Principal Research Manager

Learn more

Image 15: Portrait of Akshay Nambi

Akshay Nambi

Principal Researcher

Learn more

Image 16: Portrait of Rujia Wang

Rujia Wang

Principal Research Product Manager

Learn more

Image 17: Portrait of Chetan Bansal

Chetan Bansal

Senior Principal Research Manager

Learn more

Continue reading

Image 18: The image shows a diagram illustrating the relationship between mathematical statements in natural language and formal language. On the left, there is a blue box labeled "Mathematical statement in natural language." An arrow points from this box to a central section containing four smaller boxes arranged in two rows. The top row contains "Formalization" and "Informalization," while the bottom row contains "Symbolic Equivalence" and "Semantic Consistency." An arrow points from this central section to a purple box on the right labeled "Mathematical statement in formal language." The background of the image transitions from blue on the left to purple on the right.

June 17, 2025

[New methods boost reasoning in small and large language models](https://www.microsoft.com/en-us/research/blog/new-methods-boost-reasoning-in-small-and-large-language-models/)

Image 19: Diagram showing how the dimensions of query source (data-driven vs activity-driven) and query scope (local vs global) create four query classes that span the local-to-global query spectrum: data-local, activity-local, data-global, and activity-global.

June 5, 2025

[BenchmarkQED: Automated benchmarking of RAG systems](https://www.microsoft.com/en-us/research/blog/benchmarkqed-automated-benchmarking-of-rag-systems/)

Image 20: LazyGraphRAG blog hero

November 25, 2024

[LazyGraphRAG: Setting a new standard for quality and cost](https://www.microsoft.com/en-us/research/blog/lazygraphrag-setting-a-new-standard-for-quality-and-cost/)

Image 21: GraphRAG image on blue to green gradient

September 9, 2024

[GraphRAG auto-tuning provides rapid adaptation to new domains](https://www.microsoft.com/en-us/research/blog/graphrag-auto-tuning-provides-rapid-adaptation-to-new-domains/)

See all blog posts

Research Areas

  • ![Image 22 Artificial intelligence](https://www.microsoft.com/en-us/research/research-area/artificial-intelligence/)

Research Groups

Related labs

Follow us:

Share this page:

What's new

Microsoft Store

Education

Business

Developer & IT

Company

Your Privacy Choices

Consumer Health Privacy

Notifications