Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Applies to:
Azure SQL Managed Instance
This article provides an overview of using artificial intelligence (AI) options, such as OpenAI and vectors, to build intelligent applications with Azure SQL Managed Instance.
For samples and examples, visit the SQL AI Samples repository.
Overview
Large language models (LLMs) enable developers to create AI-powered applications with a familiar user experience.
Using LLMs in applications brings greater value and an improved user experience when the models can access the right data, at the right time, from your application's database. This process is known as Retrieval Augmented Generation (RAG) and Azure SQL Managed Instance has many features that support this new pattern, making it a great database to build intelligent applications.
The following links provide sample code of various options to build intelligent applications:
AI Option | Description |
---|---|
Azure OpenAI | Generate embeddings for RAG and integrate with any model supported by Azure OpenAI. |
Vectors | Learn how to store and query vectors in the database. |
Azure AI Search | Use your database together with Azure AI Search to train LLM on your data. |
Intelligent applications | Learn how to create an end-to-end solution using a common pattern that can be replicated in any scenario. |
Key concepts for implementing RAG with Azure OpenAI
This section includes key concepts that are critical to implement RAG with Azure OpenAI in Azure SQL Managed Instance.
Retrieval Augmented Generation (RAG)
RAG is a technique that enhances the LLM's ability to produce relevant and informative responses by retrieving additional data from external sources. For example, RAG can query articles or documents that contain domain-specific knowledge related to the user's question or prompt. The LLM can then use this retrieved data as a reference when generating its response. For example, a simple RAG pattern using Azure SQL Managed Instance could be:
- Insert data into a table.
- Link Azure SQL Managed Instance to Azure AI Search.
- Create an Azure OpenAI GPT4 model and connect it to Azure AI Search.
- Chat and ask questions about your data using the trained Azure OpenAI model from your application and from Azure SQL Managed Instance.
The RAG pattern, with prompt engineering, serves the purpose of enhancing response quality by offering more contextual information to the model. RAG enables the model to apply a broader knowledgebase by incorporating relevant external sources into the generation process, resulting in more comprehensive and informed responses. For more information on grounding LLMs, see Grounding LLMs - Microsoft Community Hub.
Prompts and prompt engineering
A prompt refers to specific text or information that serves as an instruction to an LLM, or as contextual data that the LLM can build upon. A prompt can take various forms, such as a question, a statement, or even a code snippet.
Sample prompts that can be used to generate a response from an LLM:
- Instructions: provide directives to the LLM
- Primary content: gives information to the LLM for processing
- Examples: help condition the model to a particular task or process
- Cues: direct the LLM's output in the right direction
- Supporting content: represents supplemental information the LLM can use to generate output
The process of creating good prompts for a scenario is called prompt engineering. For more information about prompts and best practices for prompt engineering, see Azure OpenAI Service.
Tokens
Tokens are small chunks of text generated by splitting the input text into smaller segments. These segments can either be words or groups of characters, varying in length from a single character to an entire word. For instance, the word hamburger
would be divided into tokens such as ham
, bur
, and ger
while a short and common word like pear
would be considered a single token.
In Azure OpenAI, input text provided to the API is turned into tokens (tokenized). The number of tokens processed in each API request depends on factors such as the length of the input, output, and request parameters. The quantity of tokens being processed also impacts the response time and throughput of the models. There are limits to the number of tokens each model can take in a single request/response from Azure OpenAI. To learn more, see Azure OpenAI Service quotas and limits.
Vectors
Vectors are ordered arrays of numbers (typically floats) that can represent information about some data. For example, an image can be represented as a vector of pixel values, or a string of text can be represented as a vector or ASCII values. The process to turn data into a vector is called vectorization. For more information, see Vectors.
Embeddings
Embeddings are vectors that represent important features of data. Embeddings are often learned by using a deep learning model, and machine learning and AI models utilize them as features. Embeddings can also capture semantic similarity between similar concepts. For example, in generating an embedding for the words person
and human
, we would expect their embeddings (vector representation) to be similar in value since the words are also semantically similar.
Azure OpenAI features models to create embeddings from text data. The service breaks text out into tokens and generates embeddings using models pretrained by OpenAI. To learn more, see Creating embeddings with Azure OpenAI.
Vector search
Vector search refers to the process of finding all vectors in a dataset that are semantically similar to a specific query vector. Therefore, a query vector for the word human
searches the entire dictionary for semantically similar words, and should find the word person
as a close match. This closeness, or distance, is measured using a similarity metric such as cosine similarity. The closer vectors are in similarity, the smaller is the distance between them.
Consider a scenario where you run a query over millions of document to find the most similar documents in your data. You can create embeddings for your data and query documents using Azure OpenAI. Then, you can perform a vector search to find the most similar documents from your dataset. However, performing a vector search across a few examples is trivial. Performing this same search across thousands, or millions, of data points becomes challenging. There are also trade-offs between exhaustive search and approximate nearest neighbor (ANN) search methods including latency, throughput, accuracy, and cost, all of which depends on the requirements of your application.
Vectors in Azure SQL Managed Instance can be efficiently stored and queried, as described in the next sections, allowing exact nearest neighbor search with great performance. You don't have to decide between accuracy and speed: you can have both. Storing vector embeddings alongside the data in an integrated solution minimizes the need to manage data synchronization and accelerates your time-to-market for AI application development.
Azure OpenAI
Embedding is the process of representing the real world as data. Text, images, or sounds can be converted into embeddings. Azure OpenAI models are able to transform real-world information into embeddings. The models are available as REST endpoints and thus can easily be consumed from Azure SQL Managed Instance using the sp_invoke_external_rest_endpoint
system stored procedure:
DECLARE @retval INT, @response NVARCHAR(MAX);
DECLARE @payload NVARCHAR(MAX);
SET @payload = JSON_OBJECT('input': @text);
EXEC @retval = sp_invoke_external_rest_endpoint @url = 'https://<openai-url>/openai/deployments/<model-name>/embeddings?api-version=2023-03-15-preview',
@method = 'POST',
@credential = [https://<openai-url>/openai/deployments/<model-name>],
@payload = @payload,
@response = @response OUTPUT;
SELECT CAST([key] AS INT) AS [vector_value_id],
CAST([value] AS FLOAT) AS [vector_value]
FROM OPENJSON(JSON_QUERY(@response, '$.result.data[0].embedding'));
Using a call to a REST service to get embeddings is just one of the integration options you have when working with SQL Managed Instance and OpenAI. You can let any of the available models access data stored in Azure SQL Managed Instance to create solutions where your users can interact with the data, such as the following example:
For additional examples on using Azure SQL and OpenAI, see the following articles:
- Generate images with Azure OpenAI Service (DALL-E) and Azure SQL
- Using OpenAI REST Endpoints with Azure SQL
Vectors
In April 2025, the vector data type was introduced in Azure SQL Managed Instance.
The dedicated vector type allows for efficient and optimized storing of vector data, and comes with a set of functions to help developers streamline vector and similarity search implementation. Calculating distance between two vectors can be done in one line of code using the new VECTOR_DISTANCE
function. For more information on the vector data type and related functions, see Overview of vectors in the SQL Database Engine.
For example:
CREATE TABLE [dbo].[wikipedia_articles_embeddings_titles_vector]
(
[article_id] [int] NOT NULL,
[embedding] [vector](1536) NOT NULL,
)
GO
SELECT TOP(10)
*
FROM
[dbo].[wikipedia_articles_embeddings_titles_vector]
ORDER BY
VECTOR_DISTANCE('cosine', @my_reference_vector, embedding)
Azure AI Search
Implement RAG-patterns with Azure SQL Managed Instance and Azure AI Search. You can run supported chat models on data stored in Azure SQL Managed Instance, without having to train or fine-tune models, thanks to the integration of Azure AI Search with Azure OpenAI and Azure SQL Managed Instance. Running models on your data enables you to chat on top of, and analyze, your data with greater accuracy and speed.
- Azure OpenAI on your data
- Retrieval Augmented Generation (RAG) in Azure AI Search
- Vector Search with Azure SQL and Azure AI Search
Intelligent applications
Azure SQL Managed Instance can be used to build intelligent applications that include AI features, such as recommenders, and Retrieval Augmented Generation (RAG) as the following diagram demonstrates:
For an end-to-end sample to build an AI-enabled application using sessions abstract as a sample dataset, see:
- How I built a session recommender in 1 hour using OpenAI.
- Using Retrieval Augmented Generation to build a conference session assistant
LangChain integration
LangChain is a well-known framework for developing applications powered by language models. For examples that show how LangChain can be used to create a Chatbot on your own data, see:
- langchain-sqlserver PyPI Package
A few of samples on using Azure SQL with LangChain:
End-to-end examples:
- Build a chatbot on your own data in 1 hour with Azure SQL, Langchain, and Chainlit: Build a chatbot using the RAG pattern on your own data using Langchain for orchestrating LLM calls and Chainlit for the UI.
Semantic Kernel integration
Semantic Kernel is an open-source SDK that lets you easily build agents that can call your existing code. As a highly extensible SDK, you can use Semantic Kernel with models from OpenAI, Azure OpenAI, Hugging Face, and more. By combining your existing C#, Python, and Java code with these models, you can build agents that answer questions and automate processes.
An example of how easily Semantic Kernel helps to build AI-enabled solution is here:
- The ultimate chatbot?: Build a chatbot on your own data using both NL2SQL and RAG patterns for the ultimate user experience.