Skip to content

Commit 7a2211a

Browse files
authored
Merge pull request #60 from cephalin/copilot/sub-pr-59
Address documentation feedback: rename scenario files, remove .gitignore change, add content
2 parents b0a1a1f + 15063b3 commit 7a2211a

10 files changed

Lines changed: 250 additions & 82 deletions

.gitignore

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,4 +158,3 @@ articles/planetary-computer/geocatalog_configs/umbra-sar-ships/stac.json
158158
articles/planetary-computer/geocatalog_configs/umbra-sar-ships/tile-settings.json
159159
articles/planetary-computer/helper-content/collection-config-scraper.py
160160
articles/planetary-computer/helper-content/generate_collection_docs.py
161-
.beads/

articles/app-service/overview-ai-integration.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -31,11 +31,11 @@ Explore these scenarios to find the right approach for your AI-powered applicati
3131

3232
| Scenario | Description | Learn more |
3333
|----------|-------------|------------|
34-
| **Chatbots and RAG applications** | Build intelligent chatbots and RAG-powered web apps with Azure OpenAI (and optional Azure AI Search) for grounded, context-aware responses directly in App Service. | [Get started](scenario-ai-chatbot-rag.md) |
34+
| **Chatbots and RAG applications** | Build intelligent chatbots and RAG-powered web apps with Azure OpenAI (and optional Azure AI Search) for grounded, context-aware responses directly in App Service. | [Get started](scenario-ai-chatbot-retrieval-augmented-generation.md) |
3535
| **Agentic web applications** | Add autonomous, reasoning AI agents to your existing CRUD web apps using frameworks like Semantic Kernel, LangGraph, or Foundry Agent Service to enable planning, multi-step actions, and natural language interactions. | [Get started](scenario-ai-agentic-web-apps.md) |
3636
| **OpenAPI tools for Foundry agents** | Expose your App Service REST APIs as secure, callable tools via OpenAPI specs so Foundry Agent Service agents can discover and invoke them for real-world actions and data access. | [Get started](scenario-ai-openapi-tool.md) |
37-
| **Model Context Protocol servers** | Host your App Service app as an MCP server to extend AI coding assistants like GitHub Copilot Chat, Cursor, and Winsurf with your custom business logic, APIs, and data context via the Model Context Protocol. | [Get started](scenario-ai-mcp-server.md) |
38-
| **Local small language models** | Run small language models (e.g., Phi-3/Phi-4) entirely locally as sidecar containers in App Service for full data privacy, zero-latency inference, offline capability, and no external API calls or dependencies. | [Get started](scenario-ai-local-slm.md) |
37+
| **Model Context Protocol servers** | Host your App Service app as an MCP server to extend AI coding assistants like GitHub Copilot Chat, Cursor, and Winsurf with your custom business logic, APIs, and data context via the Model Context Protocol. | [Get started](scenario-ai-model-context-protocol-server.md) |
38+
| **Local small language models** | Run small language models (e.g., Phi-3/Phi-4) entirely locally as sidecar containers in App Service for full data privacy, zero-latency inference, offline capability, and no external API calls or dependencies. | [Get started](scenario-ai-local-small-language-model.md) |
3939
| **Secure AI applications** | Secure your OpenAPI tools, MCP servers, and AI endpoints in App Service using Microsoft Entra ID authentication, managed identities, and built-in authorization to ensure only authorized users and agents can access them. | [Get started](scenario-ai-authentication.md) |
4040

4141
## More resources

articles/app-service/scenario-ai-agentic-web-apps.md

Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,44 @@ ms.update-cycle: 180-days
1616

1717
Transform your traditional CRUD web applications for the AI era by adding agentic capabilities with frameworks like Microsoft Semantic Kernel, LangGraph, or Foundry Agent Service. Instead of users navigating forms, textboxes, and dropdowns, you can offer a conversational interface that lets users "talk to an agent" that intelligently performs the same operations your app provides. This approach enables your web app to reason, plan, and take actions on behalf of users.
1818

19+
## Overview
20+
21+
Agentic web applications represent a paradigm shift from traditional web interfaces. Rather than requiring users to understand and navigate your application's structure, agentic apps use AI to understand user intent, plan multi-step actions, and execute complex workflows autonomously.
22+
23+
Key characteristics of agentic web applications include:
24+
25+
- **Conversational interfaces**: Users express goals in natural language rather than clicking through forms
26+
- **Autonomous reasoning**: Agents break down complex requests into executable steps
27+
- **Multi-step planning**: Agents chain multiple operations to accomplish sophisticated tasks
28+
- **Tool usage**: Agents call your existing APIs and functions as needed to complete user requests
29+
- **Context awareness**: Agents maintain conversation history and application state across interactions
30+
- **Error handling**: Agents recover from failures and adapt their approach based on results
31+
32+
Frameworks like Semantic Kernel, LangGraph, and Foundry Agent Service provide the orchestration layer that connects large language models with your application's business logic, enabling these agentic capabilities in your App Service applications.
33+
34+
## When to build agentic web applications
35+
36+
Consider adding agentic capabilities when:
37+
38+
- **Complex workflows are common**: Users frequently need to perform multi-step operations that could be simplified through conversation
39+
- **Domain expertise is required**: Your application requires specialized knowledge that an AI agent can learn and apply
40+
- **User experience matters**: You want to reduce training time and make your application more intuitive
41+
- **Data exploration is important**: Users need to query, analyze, and visualize data in flexible ways
42+
- **Task automation is valuable**: Repetitive or time-consuming workflows could benefit from AI-assisted completion
43+
44+
Agentic patterns work especially well for enterprise applications, data analysis tools, content management systems, and administrative interfaces where the combination of natural language understanding and automated workflows provides significant productivity gains.
45+
46+
## Choosing an agent framework
47+
48+
App Service supports any agent framework that runs on your chosen language stack. You have complete flexibility to use the tools and frameworks that best fit your needs. Popular options include:
49+
50+
- **Semantic Kernel**: Microsoft's cross-platform SDK for .NET, Python, and Java, ideal for building custom agents with full control
51+
- **LangGraph**: Python and JavaScript framework for building stateful, multi-agent systems with complex workflows
52+
- **Foundry Agent Service**: Managed service for hosting production-ready agents with built-in monitoring and scalability
53+
- **Custom frameworks**: Any other agentic framework supported by your language (e.g., AutoGen, CrewAI, or proprietary solutions)
54+
55+
## Get started with tutorials
56+
1957
## [.NET](#tab/dotnet)
2058
- [Tutorial: Build an agentic web app in Azure App Service with Microsoft Semantic Kernel or Foundry Agent Service (.NET)](tutorial-ai-agent-web-app-semantic-kernel-foundry-dotnet.md)
2159
- Blog series: Build long-running AI agents with Microsoft Agent Framework

articles/app-service/scenario-ai-chatbot-rag.md renamed to articles/app-service/scenario-ai-chatbot-retrieval-augmented-generation.md

Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,32 @@ ms.update-cycle: 180-days
1616

1717
Build intelligent web apps that use Azure OpenAI for chat or retrieval augmented generation (RAG). These tutorials show you how to integrate Azure OpenAI and (optionally) Azure AI Search to create chatbots and RAG solutions in your preferred language, using managed identities for secure authentication.
1818

19+
## Overview
20+
21+
Chatbots and Retrieval Augmented Generation (RAG) applications represent two of the most popular AI integration patterns in modern web applications. A chatbot uses a large language model to engage in natural conversations with users, while RAG enhances these conversations by grounding responses in your own data sources, reducing hallucinations and providing more accurate, contextual answers.
22+
23+
Azure App Service provides a comprehensive platform for hosting these intelligent applications with built-in support for:
24+
25+
- **Azure OpenAI integration**: Seamless connection to the latest Azure OpenAI models using managed identities
26+
- **Azure AI Search connectivity**: Optional integration with Azure AI Search for vector search and document retrieval
27+
- **Secure authentication**: Built-in support for managed identities eliminates the need for API keys
28+
- **Scalability**: Automatic scaling to handle varying workloads
29+
- **Multiple language support**: Deploy chatbots in .NET, Java, Node.js, or Python
30+
31+
## When to use chatbots and RAG
32+
33+
Consider building a chatbot or RAG application when you want to:
34+
35+
- **Provide conversational interfaces**: Replace traditional form-based UIs with natural language interactions
36+
- **Enable intelligent document search**: Allow users to query large document repositories using natural language
37+
- **Create customer support assistants**: Build AI-powered help desks that understand context and provide accurate responses
38+
- **Develop knowledge bases**: Transform static documentation into interactive Q&A systems
39+
- **Build internal tools**: Create employee-facing assistants that can access and explain company data
40+
41+
RAG is particularly valuable when you need to ensure your AI responses are grounded in specific, up-to-date information from your organization's data sources, such as product catalogs, documentation, policies, or customer records.
42+
43+
## Get started with tutorials
44+
1945
## [.NET](#tab/dotnet)
2046
- [Build a chatbot with Azure OpenAI (Blazor)](tutorial-ai-openai-chatbot-dotnet.md)
2147
- [Build a RAG application with Azure OpenAI and Azure AI Search (.NET)](tutorial-ai-openai-search-dotnet.md)

articles/app-service/scenario-ai-local-slm.md

Lines changed: 0 additions & 34 deletions
This file was deleted.
Lines changed: 64 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,64 @@
1+
---
2+
title: Use local small language models (SLMs) in Azure App Service
3+
description: Deploy a web app with a local small language model (SLM) as a sidecar container to run AI models entirely within your App Service environment. No outbound calls or external AI service dependencies required.
4+
author: cephalin
5+
ms.author: cephalin
6+
ms.service: azure-app-service
7+
ms.topic: how-to
8+
ms.date: 01/29/2026
9+
ms.custom:
10+
- build-2025
11+
ms.collection: ce-skilling-ai-copilot
12+
ms.update-cycle: 180-days
13+
---
14+
15+
# Use a local SLM (sidecar container)
16+
17+
Deploy a web app with a local small language model (SLM) as a sidecar container to run AI models entirely within your App Service environment. No outbound calls or external AI service dependencies required. This approach is ideal if you have strict data privacy or compliance requirements, as all AI processing and data remain local to your app. App Service offers high-performance, memory-optimized pricing tiers needed for running SLMs in sidecars.
18+
19+
## Overview
20+
21+
Small Language Models (SLMs) are compact AI models, such as Microsoft's Phi-3 and Phi-4 series, that can run efficiently with fewer computational resources compared to large language models. By deploying an SLM as a sidecar container alongside your web application in App Service, you can process AI requests entirely locally without making external API calls.
22+
23+
This architecture provides several advantages:
24+
25+
- **Complete data privacy**: All data and AI processing stays within your App Service environment
26+
- **Zero external dependencies**: No reliance on external AI services or internet connectivity
27+
- **Predictable latency**: Responses are consistently fast with no network overhead
28+
- **Cost control**: Pay only for App Service compute resources, with no per-token charges
29+
- **Regulatory compliance**: Meet strict data residency and privacy requirements
30+
31+
## When to use local SLMs
32+
33+
Local SLMs are ideal for scenarios where:
34+
35+
- **Data privacy is critical**: Healthcare, finance, legal, or government applications that cannot send data to external services
36+
- **Offline capability is required**: Applications that must function without internet connectivity
37+
- **Predictable costs are important**: Fixed infrastructure costs instead of variable per-request pricing
38+
- **Low latency is essential**: Sub-100ms response times without network calls
39+
- **Moderate AI capabilities suffice**: Tasks like classification, summarization, entity extraction, or simple Q&A that don't require the most powerful models
40+
41+
While SLMs may not match the capabilities of large models like GPT-4, they excel in focused, domain-specific tasks where smaller models can be fine-tuned for excellent performance with complete control over your data.
42+
43+
## Technical approach
44+
45+
App Service supports running SLMs through sidecar containers that deploy alongside your main application. The sidecar runs the model inference engine (such as ONNX Runtime or llama.cpp) and exposes a local endpoint your app can call. This keeps all processing in-process and maintains isolation while sharing the same compute resources.
46+
47+
## Get started with tutorials
48+
49+
## [.NET](#tab/dotnet)
50+
- [Run a chatbot with a local SLM sidecar extension](tutorial-ai-slm-dotnet.md)
51+
52+
## [Java](#tab/java)
53+
- [Run a chatbot with a local SLM (Spring Boot)](tutorial-ai-slm-spring-boot.md)
54+
55+
## [Node.js](#tab/nodejs)
56+
- [Run a chatbot with a local SLM (Express.js)](tutorial-ai-slm-expressjs.md)
57+
58+
## [Python](#tab/python)
59+
- [Run a chatbot with a local SLM (FastAPI)](tutorial-ai-slm-fastapi.md)
60+
-----
61+
62+
## Related content
63+
64+
- [Integrate AI into your Azure App Service applications](overview-ai-integration.md)

articles/app-service/scenario-ai-mcp-server.md

Lines changed: 0 additions & 37 deletions
This file was deleted.

0 commit comments

Comments
 (0)