Skip to content

Commit

Permalink
Merge pull request #1717 from MicrosoftDocs/main
Browse files Browse the repository at this point in the history
11/25/2024 AM Publish
  • Loading branch information
Taojunshen authored Nov 25, 2024
2 parents ff7681e + a81b9e6 commit ff52827
Show file tree
Hide file tree
Showing 89 changed files with 201 additions and 199 deletions.
3 changes: 2 additions & 1 deletion articles/ai-studio/tutorials/copilot-sdk-build-rag.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,12 @@ ms.custom: copilot-learning-hub, ignite-2024

# Tutorial: Part 2 - Build a custom knowledge retrieval (RAG) app with the Azure AI Foundry SDK

In this tutorial, you use the Azure AI Foundry SDK (and other libraries) to build, configure, evaluate, and deploy a chat app for your retail company called Contoso Trek. Your retail company specializes in outdoor camping gear and clothing. The chat app should answer questions about your products and services. For example, the chat app can answer questions such as "which tent is the most waterproof?" or "what is the best sleeping bag for cold weather?".
In this tutorial, you use the Azure AI Foundry SDK (and other libraries) to build, configure, and evaluate a chat app for your retail company called Contoso Trek. Your retail company specializes in outdoor camping gear and clothing. The chat app should answer questions about your products and services. For example, the chat app can answer questions such as "which tent is the most waterproof?" or "what is the best sleeping bag for cold weather?".

This part two shows you how to enhance a basic chat application by adding [retrieval augmented generation (RAG)](../concepts/retrieval-augmented-generation.md) to ground the responses in your custom data. Retrieval Augmented Generation (RAG) is a pattern that uses your data with a large language model (LLM) to generate answers specific to your data. In this part two, you learn how to:

> [!div class="checklist"]
> - Get example data
> - Create a search index of the data for the chat app to use
> - Develop custom RAG code
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ author: sdgilley
# Tutorial: Part 1 - Set up project and development environment to build a custom knowledge retrieval (RAG) app with the Azure AI Foundry SDK


In this tutorial, you use the Azure AI Foundry SDK (and other libraries) to build, configure, evaluate, and deploy a chat app for your retail company called Contoso Trek. Your retail company specializes in outdoor camping gear and clothing. The chat app should answer questions about your products and services. For example, the chat app can answer questions such as "which tent is the most waterproof?" or "what is the best sleeping bag for cold weather?".
In this tutorial, you use the Azure AI Foundry SDK (and other libraries) to build, configure, and evaluate a chat app for your retail company called Contoso Trek. Your retail company specializes in outdoor camping gear and clothing. The chat app should answer questions about your products and services. For example, the chat app can answer questions such as "which tent is the most waterproof?" or "what is the best sleeping bag for cold weather?".

This tutorial is part one of a three-part tutorial. This part one gets you ready to write code in part two and evaluate your chat app in part three. In this part, you:

Expand Down
3 changes: 2 additions & 1 deletion articles/ai-studio/tutorials/copilot-sdk-evaluate.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,11 +16,12 @@ author: sdgilley

# Tutorial: Part 3 - Evaluate a custom chat application with the Azure AI Foundry SDK

In this tutorial, you use the Azure AI SDK (and other libraries) to evaluate and deploy the chat app you built in [Part 2 of the tutorial series](copilot-sdk-build-rag.md). In this part three, you learn how to:
In this tutorial, you use the Azure AI SDK (and other libraries) to evaluate the chat app you built in [Part 2 of the tutorial series](copilot-sdk-build-rag.md). In this part three, you learn how to:

> [!div class="checklist"]
> - Create an evaluation dataset
> - Evaluate the chat app with Azure AI evaluators
> - Iterate and improve your app

This tutorial is part three of a three-part tutorial.
Expand Down
2 changes: 1 addition & 1 deletion articles/machine-learning/breadcrumb/toc.yml
Original file line number Diff line number Diff line change
Expand Up @@ -107,7 +107,7 @@ items:
tocHref: /security/benchmark/azure/
topicHref: /security/benchmark/azure/index

# AI Studio or Azure ML
# AI Foundry or Azure ML
- name: Azure
tocHref: /ai/
topicHref: /azure/index
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ Azure Machine Learning designer components (Designer) allow users to create mach

This reference content provides background on each of the custom components (v2) available in Azure Machine Learning designer.

You can navigate to Custom components in Azure Machine Learning Studio as shown in the following image.
You can navigate to Custom components in Azure Machine Learning studio as shown in the following image.

:::image type="content" source="media/designer-new-pipeline.png" alt-text="Diagram showing the Designer UI for selecting a custom component.":::

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ This section contains implementation details, tips, and answers to frequently as

Vowpal Wabbit has many command-line options for choosing and tuning algorithms. A full discussion of these options is not possible here; we recommend that you view the [Vowpal Wabbit wiki page](https://github.com/JohnLangford/vowpal_wabbit/wiki/Command-line-arguments).

The following parameters are not supported in Azure Machine Learning Studio (classic).
The following parameters are not supported in Azure Machine Learning studio (classic).

- The input/output options specified in [https://github.com/JohnLangford/vowpal_wabbit/wiki/Command-line-arguments](https://github.com/JohnLangford/vowpal_wabbit/wiki/Command-line-arguments)

Expand Down
16 changes: 8 additions & 8 deletions articles/machine-learning/concept-hub-workspace.md
Original file line number Diff line number Diff line change
Expand Up @@ -33,11 +33,11 @@ In the transition from proving feasibility of an idea, to a funded project, many

The goal of hubs is to take away this bottleneck, by letting IT set up a secure, preconfigured, and reusable environment for a team to prototype, build, and operate machine learning models.

## Interoperability between ML studio and AI studio
## Interoperability between ML studio and AI Foundry

Hubs can be used as your team's collaboration environment for both ML studio and [AI studio](/azure/ai-studio/what-is-ai-studio). Use ML Studio for training and operationalizing custom machine learning models. Use AI studio as experience for building and operating AI applications responsibly.
Hubs can be used as your team's collaboration environment for both ML studio and [AI Foundry](/azure/ai-studio/what-is-ai-studio). Use ML Studio for training and operationalizing custom machine learning models. Use AI Foundry as experience for building and operating AI applications responsibly.

| Workspace Kind | ML Studio | AI Studio |
| Workspace Kind | ML Studio | AI Foundry |
| --- | --- | --- |
| Default | Supported | - |
| Hub | Supported | Supported |
Expand All @@ -54,7 +54,7 @@ Project workspaces that are created using a hub obtain the hub's security settin
| Network settings | One [managed virtual network](how-to-managed-network.md) is shared between hub and project workspaces. To access content in the hub and project workspaces, create a single private link endpoint on the hub workspace. |
| Encryption settings | Encryption settings pass down from hub to project. |
| Storage for encrypted data | When you bring your customer-managed keys for encryption, hub and project workspaces share the same managed resource group for storing encrypted service data. |
| Connections | Project workspaces can consume shared connections created on the hub. This feature is currently only supported in [AI studio]() |
| Connections | Project workspaces can consume shared connections created on the hub. This feature is currently only supported in [AI Foundry]() |
| Compute instance | Reuse a compute instance across all project workspaces associated to the same hub. |
| Compute quota | Any compute quota consumed by project workspaces is deducted from the hub workspace quota balance. |
| Storage | Associated resource for storing workspace data. Project workspaces use designated containers starting with a prefix {workspaceGUID}, and have a conditional [Azure Attribute Based Access](/azure/role-based-access-control/conditions-overview) role assignment for the workspace identity for accessing these containers only. |
Expand All @@ -69,7 +69,7 @@ Data that is uploaded in one project workspace, is stored in isolation from data
Once a hub is created, there are multiple ways to create a project workspace using it:

1. [Using ML Studio](how-to-manage-workspace.md?tabs=mlstudio)
1. [Using AI Studio](/azure/ai-studio/how-to/create-projects)
1. [Using AI Foundry](/azure/ai-studio/how-to/create-projects)
2. [Using Azure SDK](how-to-manage-workspace.md?tabs=python)
4. [Using automation templates](how-to-create-workspace-template.md)

Expand All @@ -93,11 +93,11 @@ Features that are supported using hub/project workspaces differ from regular wor
| Feature | Default workspace | Hub workspace | Project workspace | Note |
|--|--|--|--|--|
|Self-serve create project workspaces from Studio| - | X | X | - |
|Create shared connections on hub | |X|X| Only in AI studio |
|Create shared connections on hub | |X|X| Only in AI Foundry portal |
|Consume shared connections from hub | |X|X| - |
|Reuse compute instance across workspaces|-|X|X| |
|Share compute quota across workspaces|-|X|X||
|Build GenAI apps in AI studio|-|X|X||
|Build GenAI apps in AI Foundry portal|-|X|X||
|Single private link endpoint across workspaces|-|X|X||
|Managed virtual network|X|X|X|-|
|BYO virtual network|X|-|-|Use alternative [managed virtual network](how-to-managed-network.md)|
Expand All @@ -115,6 +115,6 @@ To learn more about setting up Azure Machine Learning, see:
+ [Create and manage a workspace](how-to-manage-workspace.md)
+ [Get started with Azure Machine Learning](quickstart-create-resources.md)

To learn more about hub workspace support in AI Studio, see:
To learn more about hub workspace support in AI Foundry portal, see:

+ [How to configure a managed network for hubs](/azure/ai-studio/how-to/configure-managed-network)
6 changes: 3 additions & 3 deletions articles/machine-learning/concept-model-catalog.md
Original file line number Diff line number Diff line change
Expand Up @@ -118,19 +118,19 @@ Models that are available for deployment as serverless APIs with pay-as-you-go b

### Pay for model usage in MaaS

The discovery, subscription, and consumption experience for models deployed via MaaS is in the Azure AI Studio and Azure Machine Learning studio. Users accept license terms for use of the models, and pricing information for consumption is provided during deployment. Models from third party providers are billed through Azure Marketplace, in accordance with the [Commercial Marketplace Terms of Use](/legal/marketplace/marketplace-terms); models from Microsoft are billed using Azure meters as First Party Consumption Services. As described in the [Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage), First Party Consumption Services are purchased using Azure meters but aren't subject to Azure service terms; use of these models is subject to the license terms provided.
The discovery, subscription, and consumption experience for models deployed via MaaS is in the Azure AI Foundry portal and Azure Machine Learning studio. Users accept license terms for use of the models, and pricing information for consumption is provided during deployment. Models from third party providers are billed through Azure Marketplace, in accordance with the [Commercial Marketplace Terms of Use](/legal/marketplace/marketplace-terms); models from Microsoft are billed using Azure meters as First Party Consumption Services. As described in the [Product Terms](https://www.microsoft.com/licensing/terms/welcome/welcomepage), First Party Consumption Services are purchased using Azure meters but aren't subject to Azure service terms; use of these models is subject to the license terms provided.

### Deploy models for inference through MaaS

Deploying a model through MaaS allows users to get access to ready to use inference APIs without the need to configure infrastructure or provision GPUs, saving engineering time and resources. These APIs can be integrated with several LLM tools and usage is billed as described in the previous section.

### Fine-tune models through MaaS with Pay-as-you-go

For models that are available through MaaS and support fine-tuning, users can take advantage of hosted fine-tuning with pay-as-you-go billing to tailor the models using data they provide. For more information, see [fine-tune a Llama 2 model](/azure/ai-studio/how-to/fine-tune-model-llama) in Azure AI Studio.
For models that are available through MaaS and support fine-tuning, users can take advantage of hosted fine-tuning with pay-as-you-go billing to tailor the models using data they provide. For more information, see [fine-tune a Llama 2 model](/azure/ai-studio/how-to/fine-tune-model-llama) in Azure AI Foundry portal.

### RAG with models deployed through MaaS

Azure AI Studio enables users to make use of Vector Indexes and Retrieval Augmented Generation. Models that can be deployed as serverless APIs can be used to generate embeddings and inferencing based on custom data to generate answers specific to their use case. For more information, see [Retrieval augmented generation and indexes](concept-retrieval-augmented-generation.md).
Azure AI Foundry enables users to make use of Vector Indexes and Retrieval Augmented Generation. Models that can be deployed as serverless APIs can be used to generate embeddings and inferencing based on custom data to generate answers specific to their use case. For more information, see [Retrieval augmented generation and indexes](concept-retrieval-augmented-generation.md).

### Regional availability of offers and models

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -438,7 +438,7 @@ Assigned to user does not need compute write (create) permission to enable SSO.

Here are the steps assigned to user needs to take. Please note creator of compute instance is not allowed to enable SSO on that compute instance due to security reasons.

1. Click on compute in left navigation pane in Azure Machine Learning Studio.
1. Click on compute in left navigation pane in Azure Machine Learning studio.
1. Click on the name of compute instance where you need to enable SSO.
1. Edit the Single sign-on details section.

Expand Down
4 changes: 2 additions & 2 deletions articles/machine-learning/how-to-custom-dns.md
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ When using an Azure Machine Learning workspace (including Azure AI hubs) with a
- An Azure Virtual Network that uses [your own DNS server](/azure/virtual-network/virtual-networks-name-resolution-for-vms-and-role-instances#name-resolution-that-uses-your-own-dns-server).

:::moniker range="azureml-api-2"
- An Azure Machine Learning workspace with a private endpoint, including hub workspaces such as those used by Azure AI Studio. For more information, see [Create an Azure Machine Learning workspace](how-to-manage-workspace.md).
- An Azure Machine Learning workspace with a private endpoint, including hub workspaces such as those used by Azure AI Foundry. For more information, see [Create an Azure Machine Learning workspace](how-to-manage-workspace.md).

- If your workspace dependency resources are secured with an __Azure Virtual network__, familiarity with the [Network isolation during training & inference](./how-to-network-security-overview.md) article.
:::moniker-end
Expand Down Expand Up @@ -57,7 +57,7 @@ Another option is to modify the `hosts` file on the client that is connecting to
Access to a given Azure Machine Learning workspace via Private Link is done by communicating with the following Fully Qualified Domains (called the workspace FQDNs) listed below:

> [!IMPORTANT]
> If you are using a hub workspace (including Azure AI Studio hub), then you will have addtional entries for each project workspace created from the hub.
> If you are using a hub workspace (including Azure AI Foundry hub), then you will have additional entries for each project workspace created from the hub.
**Azure Public regions**:
- ```<per-workspace globally-unique identifier>.workspace.<region the workspace was created in>.api.azureml.ms```
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2165,4 +2165,4 @@ For more information on how to track costs, see [Monitor costs for models offere
* [Azure AI Model Inference API](reference-model-inference-api.md)
* [Deploy models as serverless APIs](how-to-deploy-models-serverless.md)
* [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md)
* [Plan and manage costs for Azure AI Studio](concept-plan-manage-cost.md)
* [Plan and manage costs for Azure AI Foundry](/azure/ai-studio/how-to/costs-plan-manage)
Original file line number Diff line number Diff line change
Expand Up @@ -666,4 +666,4 @@ Quota is managed per deployment. Each deployment has a rate limit of 200,000 tok
* [Azure AI Model Inference API](reference-model-inference-api.md)
* [Deploy models as serverless APIs](how-to-deploy-models-serverless.md)
* [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md)
* [Plan and manage costs for Azure AI Studio](concept-plan-manage-cost.md)
* [Plan and manage costs for Azure AI Foundry](/azure/ai-studio/how-to/costs-plan-manage)
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ author: msakande
ms.custom: references_regions, build-2024
ms.collection: ce-skilling-ai-copilot

#This functionality is also available in Azure AI Studio: /azure/ai-studio/how-to/deploy-models-cohere.md
#This functionality is also available in Azure AI Foundry portal: /azure/ai-studio/how-to/deploy-models-cohere.md
---

# How to deploy Cohere Rerank models with Azure Machine Learning studio
Expand Down Expand Up @@ -250,5 +250,5 @@ For more information on how to track costs, see [Monitor costs for models offere

- [Model Catalog and Collections](concept-model-catalog.md)
- [Deploy and score a machine learning model by using an online endpoint](how-to-deploy-online-endpoints.md)
- [Plan and manage costs for Azure AI Studio](concept-plan-manage-cost.md)
- [Plan and manage costs for Azure AI Foundry](/azure/ai-studio/how-to/costs-plan-manage)
- [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md)
6 changes: 3 additions & 3 deletions articles/machine-learning/how-to-deploy-models-jais.md
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ You can find the JAIS model in the model catalog by filtering on the JAIS collec

For a list of regions that are available for each of the models supporting serverless API endpoint deployments, see [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md).

- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Studio](/azure/ai-studio/concepts/rbac-ai-studio).
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Foundry portal. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Foundry portal](/azure/ai-studio/concepts/rbac-ai-studio).


### JAIS 30b Chat
Expand Down Expand Up @@ -270,7 +270,7 @@ Response:

### Cost and quota considerations for models deployed as a service

JAIS 30b Chat is deployed as a service are offered by Core42 through the Azure Marketplace and integrated with Azure AI Studio for use. You can find the Azure Marketplace pricing when deploying the model.
JAIS 30b Chat is deployed as a service are offered by Core42 through the Azure Marketplace and integrated with Azure AI Foundry for use. You can find the Azure Marketplace pricing when deploying the model.

Each time a project subscribes to a given offer from the Azure Marketplace, a new resource is created to track the costs associated with its consumption. The same resource is used to track costs associated with inference; however, multiple meters are available to track each scenario independently.

Expand All @@ -284,6 +284,6 @@ Models deployed as a service with pay-as-you-go are protected by [Azure AI Conte

## Next steps

- [What is Azure AI Studio?](/azure/ai-studio/what-is-ai-studio)
- [What is Azure AI Foundry?](/azure/ai-studio/what-is-ai-studio)
- [Azure AI FAQ article](/azure/ai-studio/faq)
- [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md)
Loading

0 comments on commit ff52827

Please sign in to comment.