|
| 1 | +--- |
| 2 | +title: Select models for agents in Microsoft Discovery |
| 3 | +description: Learn how to choose the right OpenAI model for your Microsoft Discovery agents based on use case, output quality, cost, and response time. |
| 4 | +author: leijgao |
| 5 | +ms.author: leijiagao |
| 6 | +ms.service: azure |
| 7 | +ms.topic: how-to |
| 8 | +ms.date: 04/15/2026 |
| 9 | + |
| 10 | +#CustomerIntent: As a researcher or scientist, I want to select the best model for my Discovery agents so that I can balance output quality, cost, and response time. |
| 11 | +--- |
| 12 | + |
| 13 | +# Select models for agents in Microsoft Discovery |
| 14 | + |
| 15 | +Microsoft Discovery is built on [Microsoft Foundry Agent Service](/azure/foundry/agents/concepts/runtime-components). All models available in the [Foundry model catalog](https://ai.azure.com/catalog/models) are accessible for Discovery agents. During public preview, we recommend OpenAI GPT-5.x series models for the best experience with Discovery agents. |
| 16 | + |
| 17 | +This article helps you choose the right model for your agents based on task complexity, output quality, cost, and response time. |
| 18 | + |
| 19 | +## Prerequisites |
| 20 | + |
| 21 | +- An active [Azure subscription](https://azure.microsoft.com/free/). |
| 22 | +- A deployed Microsoft Discovery workspace with at least one project. For setup instructions, see [Get started with Microsoft Discovery infrastructure](quickstart-infrastructure-portal.md). |
| 23 | +- A chat model deployment configured at the workspace level. For details, see [Create a chat model deployment](quickstart-infrastructure-portal.md#5-create-chat-model-deployment). |
| 24 | + |
| 25 | +## Understand available GPT-5.x models |
| 26 | + |
| 27 | +The following table summarizes the OpenAI GPT-5.x models recommended for Discovery agents during public preview. |
| 28 | + |
| 29 | +| Model | Context window | Strength | Relative cost | Response time | |
| 30 | +| --- | --- | --- | --- | --- | |
| 31 | +| **GPT-5.2** | 400,000 tokens | General-purpose reasoning, tool use, structured output | Medium | Medium | |
| 32 | +| **GPT-5.4** | 1,050,000 tokens | Production-grade tasks with large context requirements | Medium-High | Medium | |
| 33 | +| **GPT-5.2-Pro** | 400,000 tokens | Deep reasoning, complex research, advanced code generation | High | Slow | |
| 34 | +| **GPT-5.4-Pro** | 1,050,000 tokens | Maximum reasoning depth with extended context | Highest | Slowest | |
| 35 | +| **GPT-5.2-Chat** | 400,000 tokens | Conversational interactions, Question and Answer, guidance | Medium-Low | Fast | |
| 36 | +| **GPT-5.4-Chat** | 1,050,000 tokens | Conversational interactions with larger context | Medium | Fast | |
| 37 | +| **GPT-5-mini** | 400,000 tokens | High-volume, cost-sensitive workloads | Low | Fast | |
| 38 | +| **GPT-5-nano** | 400,000 tokens | Ultra-low-cost, latency-sensitive workloads | Lowest | Fastest | |
| 39 | + |
| 40 | +## Match models to agent use cases |
| 41 | + |
| 42 | +Different agent tasks have different requirements. Use the following guidance to select the right model for each agent in your project. |
| 43 | + |
| 44 | +### Prompt agents for research and analysis |
| 45 | + |
| 46 | +Research agents handle literature review, data analysis, summarization, and scientific reasoning. These tasks require strong reasoning and accurate output. |
| 47 | + |
| 48 | +**Recommended models:** |
| 49 | + |
| 50 | +- **GPT-5.2** (default recommendation)—Provides reliable reasoning for most research tasks. It handles tool execution, structured output, and multi-step analysis well. Start here for most prompt agents. |
| 51 | +- **GPT-5.4**—Choose this model when your agent processes large documents or needs extended context. The 1,050,000 token context window supports ingesting full papers, datasets, or lengthy experiment logs. |
| 52 | +- **GPT-5.2-Pro / GPT-5.4-Pro—Use Pro variants for agents that perform deep scientific reasoning, complex hypothesis generation, or advanced code synthesis. Pro models allocate more compute per request and produce more thorough outputs. Expect higher cost and slower response times. |
| 53 | + |
| 54 | +### Prompt agents for planning and routing |
| 55 | + |
| 56 | +Planning agents generate research plans, make routing decisions, or coordinate other agents. These tasks need consistent, deterministic behavior rather than creative reasoning. |
| 57 | + |
| 58 | +**Recommended models:** |
| 59 | + |
| 60 | +- **GPT-5.2**—Handles planning and routing reliably. Set temperature to `0` for deterministic behavior. |
| 61 | +- **GPT-5.2-Chat**—A cost-effective alternative for lightweight planning tasks that don't require deep reasoning. Chat models respond faster and cost less per token. |
| 62 | +- **GPT-5-mini**—Suitable for simple routing decisions where the agent selects from a fixed set of options. Offers the best balance of cost and speed for straightforward classification tasks. |
| 63 | + |
| 64 | +### Prompt agents for tool execution |
| 65 | + |
| 66 | +Tool-execution agents invoke Discovery tools, code interpreters, or Model Context Protocol (MCP) tools. They need reliable function calling and structured output generation. |
| 67 | + |
| 68 | +**Recommended models:** |
| 69 | + |
| 70 | +- **GPT-5.2**—Offers the most consistent tool-calling behavior across the model family. It reliably generates structured JSON for function arguments and parses tool responses accurately. |
| 71 | +- **GPT-5.4**—Choose this model when tool operations involve large input payloads or when the agent needs to maintain context across many sequential tool calls. |
| 72 | + |
| 73 | +### Prompt agents for user interaction |
| 74 | + |
| 75 | +Interactive agents handle Question and Answer, onboarding, or exploratory conversations with researchers. They benefit from natural conversational tone and fast response times. |
| 76 | + |
| 77 | +**Recommended models:** |
| 78 | + |
| 79 | +- **GPT-5.2-Chat / GPT-5.4-Chat**—Optimized for conversational interactions. Chat models provide natural, responsive dialogue, with lower latency and cost compared to their base counterparts. |
| 80 | +- **GPT-5-mini**—A strong choice for high-volume interactive scenarios if cost efficiency matters. Delivers good conversational quality at a fraction of the cost. |
| 81 | + |
| 82 | +### Workflow agents |
| 83 | + |
| 84 | +Workflow agents orchestrate multiple prompt agents through action flows. The workflow agent itself doesn't use a model directly. Instead, each prompt agent invoked within the workflow uses its own model deployment. Apply the guidance in the previous sections to each prompt agent in your workflow. |
| 85 | + |
| 86 | +For multi-agent workflows, you can mix models across agents. For example, use GPT-5-mini for a routing agent, GPT-5.2 for a data-processing agent, and GPT-5.2-Pro for a synthesis agent. This approach optimizes cost without sacrificing output quality where it matters. |
| 87 | + |
| 88 | +## Evaluate tradeoffs between quality cost and speed |
| 89 | + |
| 90 | +Use the following decision matrix to guide your model selection. |
| 91 | + |
| 92 | +| Priority | Recommended model | Tradeoff | |
| 93 | +| --- | --- | --- | |
| 94 | +| **Highest output quality** | GPT-5.4-Pro | Slowest response, highest cost | |
| 95 | +| **Best general-purpose balance** | GPT-5.2 | Good quality, moderate cost, moderate speed | |
| 96 | +| **Large context requirements** | GPT-5.4 | Higher cost, supports up to 1,050,000 tokens | |
| 97 | +| **Fast conversational responses** | GPT-5.2-Chat or GPT-5.4-Chat | Reduced reasoning depth | |
| 98 | +| **Cost optimization** | GPT-5-mini | Good quality at lower cost | |
| 99 | +| **Lowest cost and latency** | GPT-5-nano | Reduced output quality, best for simple tasks | |
| 100 | + |
| 101 | +### Tips for optimizing cost |
| 102 | + |
| 103 | +- **Start with GPT-5.2.** It's the recommended default for Discovery agents. Move to a different model only when you have a specific reason. |
| 104 | +- **Use smaller models for simple tasks.** Routing, classification, and formatting tasks don't need Pro-level reasoning. GPT-5-mini or GPT-5-nano reduces cost significantly. |
| 105 | +- **Reserve Pro models for high-value tasks.** Deep research synthesis, complex hypothesis generation, and advanced code analysis justify the higher cost. |
| 106 | +- **Mix models in workflows.** Assign different models to different agents within the same workflow based on each agent's task complexity. |
| 107 | + |
| 108 | +## Configure a model deployment for your agent |
| 109 | + |
| 110 | +You configure model deployments at the workspace level. All agents in a project share these deployments. |
| 111 | + |
| 112 | +1. Deploy a model as an Azure resource managed resource at the workspace level using Azure CLI, Bicep, or ARM templates. See [Create a chat model deployment](quickstart-infrastructure-portal.md#5-create-chat-model-deployment) for detailed steps. |
| 113 | + |
| 114 | +1. In Discovery Studio, create or edit a prompt agent. |
| 115 | + |
| 116 | +1. Under **Chat model**, select the deployment name that corresponds to the model you want to use (for example, `my-gpt-52-deployment`). |
| 117 | + |
| 118 | +1. Adjust **Temperature** and **Top-P** response controls based on your use case: |
| 119 | + - For planning and routing agents, set **Temperature** to `0` for deterministic output. |
| 120 | + - For research and analysis agents, use **Temperature** between `0.3` and `0.7` for balanced creativity and precision. |
| 121 | + - For exploratory or brainstorming agents, set **Temperature** between `0.7` and `1.0`. |
| 122 | + |
| 123 | +1. Save the agent. Each save creates a new immutable version. |
| 124 | + |
| 125 | +You can deploy multiple models in the same workspace and assign different deployments to different agents. Reference deployments by name, not resource ID. |
| 126 | + |
| 127 | +## Related content |
| 128 | + |
| 129 | +- [Agent types in Microsoft Discovery](concept-discovery-agent-types.md) |
| 130 | +- [Create agents in Microsoft Discovery](how-to-author-agents.md) |
| 131 | +- [Microsoft Foundry models overview](/azure/foundry/concepts/foundry-models-overview) |
| 132 | +- [Foundry model catalog](https://ai.azure.com/catalog/models) |
0 commit comments