Welcome to the AI Multi-Agent Architecture Workshop, designed for organizations seeking to unlock the power of AI-driven intelligent agents. Over this 3-to-5-day interactive workshop, Microsoft architects will guide you step-by-step to build a private, secure AI system tailored to your business needs.
This workshop will teach you how to develop a multi-agent system capable of comprehending diverse datasets across various locations. These intelligent agents can answer questions with detailed explanations and source references, providing your organization with a powerful, ChatGPT-like experience designed for enterprise use.
This hands-on workshop will walk you through creating a Proof of Concept (POC) for a Generative AI Multi-Agent Architecture using Azure Services. By the end of the workshop, you'll have built:
-
A Scalable Backend
Developed with Bot Framework and FastAPI, the backend serves as the engine connecting AI logic to multiple communication channels, including:- Web Chat
- Microsoft Teams
- SMS
- Slack, and more!
-
A User-Friendly Frontend
Build a web application that combines:- A search engine capable of querying your data intelligently.
- A bot UI for seamless conversational experiences.
-
A RAG-Based Multi-Agent Architecture
Leverage Retrieval-Augmented Generation (RAG) to enable your agents to retrieve precise information and generate accurate responses.
- Step-by-Step Guidance: Each module builds upon the previous one, progressively introducing you to real-world AI architecture concepts.
- Custom Enterprise AI: Create intelligent agents that understand your organization’s data while maintaining privacy and security.
- Multi-Channel Capabilities: Deploy your agents across various platforms for broad accessibility.
- Practical Experience: Learn by doing, with notebooks and code samples tailored for an enterprise setting.
By the end of the workshop, you'll have a working knowledge of how to design, build, and deploy AI agents in a multi-agentic architecture. This hands-on experience will help you understand the value of Azure-powered Generative AI in solving real-world business problems.
This is a customer-funded Value-Based Delivery (VBD). Below, you'll find all the assets and resources needed for a successful workshop delivery.
Item | Description | Link |
---|---|---|
VBD SKU Info and Datasheet | CSAM must dispatch it as "Customer Invested" against credits/hours of Unified Support Contract. Customer decides if 3 or 5 days. | ESXP SKU page |
VBD Accreditation for CSAs | Links for CSAs to get the Accreditation needed to deliver the workshop | Link 1 , Link 2 |
VBD 3-5 day POC Asset (IP) | The MVP to be delivered (this GitHub repo) | Azure-Cognitive-Search-Azure-OpenAI-Accelerator |
VBD Workshop Deck | The deck introducing and explaining the workshop | Intro AOAI GPT Azure Smart Search Engine Accelerator.pptx |
CSA Training Video | 2 Hour Training for Microsoft CSA's | POC VBD Training Recording (New video coming soon!) |
- Azure subscription
- Microsoft members preferably to be added as Guests in clients Azure AD. If not possible, then customers can issue corporate IDs to Microsoft members
- A Resource Group (RG) needs to be set for this Workshop POC, in the customer Azure tenant
- The customer team and the Microsoft team must have Contributor permissions to this resource group so they can set everything up 2 weeks prior to the workshop
- Customer Data/Documents must be uploaded to the blob storage account, at least two weeks prior to the workshop date
- A Single-Tenant App Registration (Service Principal) must be created by the customer (save the Client Id and Secret Value).
- Customer must provide the Microsoft Team , 10-20 questions (easy to hard) that they want the Agent/Bot to respond correctly.
- For IDE collaboration and standarization during workshop, AML compute instances with Jupyper Lab will be used, for this, Azure Machine Learning Workspace must be deployed in the RG
- Note: Please ensure you have enough core compute quota in your Azure Machine Learning workspace
- The user asks a question.
- In the backend app, an Agent determines which source to use based on the user input
- Five types of sources are available:
- 3a. Azure SQL Database - contains COVID-related statistics in the US.
- 3b. API Endpoints - RESTful OpenAPI 3.0 API from a online currency broker.
- 3c. Azure Bing Search API - provides access to the internet allowing scenerios like: QnA on public websites .
- 3d. Azure AI Search - contains AI-enriched documents from Blob Storage:
- Transcripts of the dialogue of all the episodes of the TV Show: FRIENDS
- 90,000 Covid publication abstracts
- 4 lenghty PDF books
- 3f. CSV Tabular File - contains COVID-related statistics in the US.
- The Agent retrieves the result from the correct source and crafts the answer.
- The Agent state is saved to CosmosDB as persistent memory and for further analysis.
- The answer is delivered to the user.
https://gptsmartsearch-frontend.azurewebsites.net
- 100% Python.
- Uses Azure Cognitive Services to index and enrich unstructured documents: OCR over images, Chunking and automated vectorization.
- Uses Hybrid Search Capabilities of Azure AI Search to provide the best semantic answer (Text and Vector search combined).
- Uses LangChain as a wrapper for interacting with Azure OpenAI , vector stores, constructing prompts and creating agents.
- Multi-Agentic Architecture using LangGraph.
- Multi-Lingual (ingests, indexes and understand any language)
- Multi-Index -> multiple search indexes
- Multi-modal input and output (text and audio)
- Tabular Data Q&A with CSV files and SQL flavor Databases
- Uses Azure AI Document Intelligence SDK (former Form Recognizer) to parse complex/large PDF documents
- Uses Bing Search API to power internet searches and Q&A over public websites.
- Connects to API Data sources by converting natural language questions to API calls.
- Uses CosmosDB as persistent memory to save user's conversations.
- Uses Streamlit to build the Frontend web application in python.
- Uses Bot Framework and Bot Service to Host the Bot API Backend and to expose it to multiple channels including MS Teams.
- Uses also FastAPI to deploy an alternative backend API with streaming capabilites
Note: (Pre-requisite) You need to have an Azure OpenAI service already created
-
Fork this repo to your Github account.
-
In Azure OpenAI studio, deploy these models (older models than the ones stated below won't work):
- "gpt-4o"
- "gpt-4o-mini"
- "text-embedding-3-large"
- "tts"
- "whisper"
-
Create a Resource Group where all the assets of this accelerator are going to be. Azure OpenAI can be in different RG or a different Subscription.
-
ClICK BELOW to create all the Azure Infrastructure needed to run the Notebooks (Azure AI Search, Cognitive Services, etc):
Note: If you have never created a
Azure AI Services Multi-Service account
before, please create one manually in the azure portal to read and accept the Responsible AI terms. Once this is deployed, delete this and then use the above deployment button. -
Clone your Forked repo to your AML Compute Instance. If your repo is private, see below in Troubleshooting section how to clone a private repo.
-
Make sure you run the notebooks on a Python 3.12 conda enviroment.
-
Install the dependencies on your machine (make sure you do the below pip comand on the same conda environment that you are going to run the notebooks. For example, in AZML compute instance run:
conda create -n GPTSearch python=3.12 conda activate GPTSearch pip install -r ./common/requirements.txt conda install ipykernel python -m ipykernel install --user --name=GPTSearch --display-name "GPTSearch (Python 3.12)"
-
Edit the file
credentials.env
with your own values from the services created in step 4.- For BLOB_SAS_TOKEN and BLOB_CONNECTION_STRING. Go to Storage Account>Security + networking>Shared access signature>Generate SAS
-
Run the Notebooks in order using the "GPTSearch (Python 3.12)" kernel. They build up on top of each other.
Troubleshooting
Steps to clone a private repo:
- On your Terminal, Paste the text below, substituting in your GitHub email address. Generate a new SSH key.
ssh-keygen -t ed25519 -C "your_email@example.com"
- Copy the SSH public key to your clipboard. Add a new SSH key.
cat ~/.ssh/id_ed25519.pub
# Then select and copy the contents of the id_ed25519.pub file
# displayed in the terminal to your clipboard
- On GitHub, go to Settings-> SSH and GPG Keys-> New SSH Key
- In the "Title" field, add a descriptive label for the new key. "AML Compute". In the "Key" field, paste your public key.
- Clone your private repo
git clone git@github.com:YOUR-USERNAME/YOUR-REPOSITORY.git
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.