Azure Data Factory November 2022 Monthly Update

Azure Data Factory November 2022 Monthly Update

This article is contributed. See the original author and article here.

Welcome to Azure Data Factory’s November monthly update! Here we’ll share the latest updates on What’s New in Azure Data Factory. You can find all our updates here.


 


We’ll also be hosting our monthly livestream next week on December 15th at 9:00am PST/ 12:00pm EST! Join us to see some live demos and to ask us your ADF questions! 


 


Join the livestream here.


 


 


Table of Contents


Continuous Integration and continuous deployment (CI/CD)



Data flow



Developer Productivity



 


Continuous integration and continuous delivery (CI/CD)


 


Disable publish button in ADF studio now available


We have added the ability to disable the publish button from the ADF studio when source control is enabled. Once this is enabled, the Publish button will be grayed out in the ADF Studio. This will help to avoid overwriting the last automated publish deployment.


 


Noelle_Li_0-1670017635045.png


 


To learn more about this update, read Source control – Azure Data Factory | Microsoft Learn.


 


Data flow


 


New improvements to SAP Change Data Capture (CDC)


We added some new updates to the SAP CDC connector in mapping data flows, adding new capabilities:


Incremental only is now available


You can get changes only from SAP system without initial full load. With this feature, you have the flexibility to focus on incremental data process without worrying about big data volume movement on an initial full data load.


 

 

sap-change-data-capture-mapping-data-flow-run-mode.png


 


Performance improvements have been made – source partitions in initial full data load


With a simple radio button, you can now connect your SAP system to multiple sessions in parallel to extract data, which will greatly reduce the time to perform an initial full data load from your SAP system.


 


sap-change-data-capture-mapping-data-flow-optimize-partition.png


 


Pipeline template added to Template Gallery


There is a new pipeline template in the Template Gallery to help speed up your journey on loading big amounts of SAP objects by a parameterized pipeline with SAP CDC enabled in ADF. You can get more information on the template here.


 


To learn more about this, read Transform data from an SAP ODP source with the SAP CDC connector in Azure Data Factory or Azure Synapse Analytics – Azure Data Factory & Azure Synapse | Microsoft Learn.


 


Developer Productivity


 


Pipeline designer enhancements added to ADF Studio preview experience


We have added three new UI updates to the ADF Studio preview experience:


Dynamic Content Flyout


A new dynamic content flyout has been added to make it easier to set dynamic content in your pipeline activities without having to use the expression builder.


 


Noelle_Li_1-1670017635052.png


 


Error Messaging


Error messages have been relocated to the Status column in Debug and Pipeline monitoring. This will make it easier to view errors whenever a pipeline run fails.


 


Noelle_Li_2-1670017635057.png


 


Container view for monitoring


A new monitoring view is available, adding the option to enable the container view. This view provides a more consolidated monitoring view of activities that were run. A big change that you will see is that nested activities are now grouped under the parent activity. To be able to view more activities on one page, you can choose to hide the list of nested activities under the parent activity.  


 


Noelle_Li_3-1670017635064.png


 


To learn more about these enhancements, read Managing Azure Data Factory studio preview experience – Azure Data Factory | Microsoft Learn.


 


We hope that you found this helpful! Let us know in the comments if there’s anything else you’d like to see in our blogs or livestreams. We love hearing your feedback!


 

Generate Azure Policy Compliance Alerts By Sending Custom Data to Log Analytics

Generate Azure Policy Compliance Alerts By Sending Custom Data to Log Analytics

This article is contributed. See the original author and article here.

Purpose


The purpose of this post is to walk through the process of creating an event driven alerting mechanism for Azure Policy compliance.


 


Assumptions


General knowledge of Azure, PowerShell, and resource creation, both in the portal and with command-line interfaces.


 


Challenge


Many organizations use Azure Policy to track, measure, maintain, and enforce regulatory policy compliance. These regulatory compliance initiatives could be standard baseline initiatives that have been assigned or they could be customized regulatory compliance initiatives, created just for that particular organization. Regardless of the regulatory compliance initiative type, organizations have prioritized not just compliance to a regulatory compliance initiative but also when a policy state change occurs. A common question we hear is “How can I be alerted when my policy compliance state changes?”. If an organization would rather use automation instead of a manual method, this article will describe an alerting mechanism that will notify you about what policy changed, when that policy changed, and how you want to be notified about that change.


 


In Azure, there are multiple ways to accomplish the same objective. Azure Policy is no different. These different methods can be broken down into two categories: event driven and time driven. Time driven methods require you to query or retrieve some data from a source on a schedule. This scheduled query would then leverage technology to determine logic and trigger a response. While this time driven method does work, it is complex and inefficient. This particular method introduces delays between the time a policy event occurs and the time in which you get an alert. The other method related to Azure Policy is event driven. An event driven method is a way to trigger a response to a policy event when it happens. This event driven approach will be the focus of this post.


 


Querying the Azure Activity log has been one way that administrators have retrieved Azure Policy events. However, the Azure Activity log does not provide the level of detail that is required for Azure Policy especially in regard to regulatory compliance. Admins would like to know when a regulatory compliance policy is no longer in compliance. Unfortunately, because the Azure Activity log is focused on all Azure Activity, it does not provide specifics about Azure Policy event changes. This event driven solution will walk you through the steps of obtaining a method for rich Azure Policy activity details such as regulatory compliance changes.


 


Solution


The first item needed in the solution is to determine where the type of data source is that we need to capture for Azure Policy regulatory compliance details. The preferred data source for capturing policy specific events is the PolicyInsights event data. PolicyInsights data comes in three event types in Azure: Policy Compliance Status Added, Policy Compliance Status Changed, Policy Compliance Status Deleted. The next question would be: How do I access or capture this PolicyInsights Event data and then create alerts so that I can be notified when certain policy events happen, like compliance change? The following architecture answers that question.


 


diag.png


 


 



  1. Azure Policy: The first step in this process is Azure Policy. Policies are assigned and those policies have compliance states.

  2. Event Grid: When an Azure Policy compliance state changes, the Event Grid will pick it up because it is “listening” for PolicyInsights events.

  3. Event Grid Subscription: An Event Grid Subscription will be configured to send those captured events to an Azure Function.

  4. Azure Function: An Azure Function will be using PowerShell code to harvest the incoming policy event data and use the Log Ingestion API to send it to the Log Analytics Workspace through a Data Collection Endpoint and Data Collection Rule.

  5. Log Ingestion API: The Log Ingestion API will be used to send this data through the Data Collection Endpoint and Data Collection Rule to the Log Analytics Workspace.

  6. Log Analytics Workspace: A Log Analytics Workspace will be configured with a Custom Table created to receive the data coming from the Data Collection Endpoint.

  7. Monitor: Azure Monitor will be used to run queries on the Custom Table to indicate when an alert should be triggered.

  8. Alert: An alert will be configured to be triggered when the Custom Table query indicates something is out of compliance.


Key Vault


In this reference architecture, we will use a Key Vault to store a secret value that gets dynamically retrieved by the PowerShell code in the Azure Function. The purpose of this process is to maintain proper security posture and provide a secure way to store and retrieve this sensitive data. An existing Key Vault may be used, or you may elect to create a new one for this purpose. Rather than walk through the creation of a new Key Vault, we will just be covering the specific configuration items that are needed. Key Vault deployment docs can be found here ( https://learn.microsoft.com/en-us/azure/key-vault/general/quick-create-portal).


 



  1. Secrets: The Key Vault will be used to store the App Registration Secret that you created in the AAD App registration. The value of the secret was one that you should have saved in your reminders notepad. Go to your Key Vault and click “Secrets”.

  2. Now click “Generate/Import” on the top menu bar to create a new secret.

  3. Upload option is Manual. Assign a “Name” and “Secret Value“. We used “PolicyAlert-Secret” as the name and put in the AAD App Registration secret as the value. Leave as enabled. Then click “Create“.KV3a.png


 



  1. Access Configuration: The Key Vault will need to have the Access Configuration set to Azure role-based access control. Click to apply the Azure role-based access control Permission model. This model is required so that you can setup specific access for the Azure Function managed identity in a later step.


    1. KV4a.png





  1. Access Control: You will need to assign the “Key Vault Secrets User” role to the Azure Function managed identity. If you have not yet created that managed identity, you can come back later to do it or do it in the Azure Function section of the portal.KV5a.png

     




Event Grid System Topic


The Event Grid System Topic will capture the PolicyInsights data so it can be sent to a Log Analytics Workspace.



  1. Type Event Grid in the Global Azure search and select it.

  2. Under Azure service events, select System topics.


    1. Create New and use these Basic Settings:

    2. Change the Topic Types to Microsoft PolicyInsights

    3. Ensure that the scope is set to Azure Subscription

    4. Ensure that the appropriate Subscription name has been selected.

    5. Select the appropriate Resource Group

    6. Give the SystemTopic an appropriate name such as egst-PolicyAlert

    7. Tags as needed

    8. Review and create.EG2i.png



Function App


The Function App be used to harvest the PolicyInsights data from the Event Grid and then write it to a Log Analytics Workspace.



  1. Basics:


    1. Go to Azure Global Search and type/select Function App

    2. Click Create

    3. Select Resource Group where Event Grid resides

    4. Give name to function app according to naming convention and that makes sense for example, FN-PolicyAlert or FA-PolicyAlert

    5. Publish: Code

    6. Runtime Stack: PowerShell Core

    7. Version: 7.2 (newest available)

    8. Region: Select the region where you are working and have the EventGrid resource located

    9. Operating System: Windows

    10. Plan type: Consumption (Serverless)

      1. FA1k.png




  2. Hosting: Select a storage account that you already have or accept the storage account that is created automatically.

  3. Networking: Leave network injection to off

  4. Monitoring: Enable Applications Insights; accept the newly created application insight.

  5. Deployment: Accept the defaults.

  6. Tags: As needed

    1. FA7a.png



  7. Review and Create



  1. Identity: Once the Function App is created, you need to configure the Managed Identity.


    1. After updating the slider to enable system assigned managed identity, the following pop up will occur. Click Yes to enable.

    2. As outlined earlier, assign the “Key Vault Secrets User” role to the Azure Function managed identity. Click on Azure role assignments, click on + Add role assignment (preview), select the scope to Key Vault, the Subscription in which you are working, the resource, being the actual Key Vault to which you want to apply the role to the managed identity, and finally what role to which you are assigning the managed identity which is the Key Vault Secrets User

      1. FA8b1.png

      2. DJBartles_8-1667850167212.png




    3. Check permissions on the Key Vault. Go to Key Vault. Select Access Control (IAM). Click Check Access. Select for Find: Managed Identity. Ensure you have the correct subscription, select the managed identity, Function App and select the Function app you created for this solution.

      1. DJBartles_6-1669817092834.png

         





    4. You should see the role of Key Vault Secrets User

      1. DJBartles_7-1669817232347.png

         







Function


The Function inside of the Function App will be running code to format the Event Hub data and write it to Log Analytics.



  1. Create the Function (inside of the Function App).

    1. Go the Function App that you just created.

    2. From the left hand menu, in the Functions sub menu, click on Functions.

    3. Click Create. A flyout menu will pop up on the right hand side of the screen.

    4. Leave the development environment as Develop in portal.

    5. The next section will be Select a Template. In this search bar, type and search for “Azure Event Grid Trigger” and select it. You can accept the default name but you can change the name to something that makes sense to the solution and/or naming convention, in this case we used “PolicyAlertTrigger1“.

      1. F1f.png

         







  2. Once the Function is created, select Integration to view the contents.

    1. F2a.png

       





  3. Click to open the “Trigger” (eventGridEvent in this example).


    1. Please note the “Event Trigger Parameter Name“. This can be customized but it must match in the PowerShell code for the function. In this example, we will use the default value of “eventGridEvent“.

      1. DJBartles_13-1667850167223.png

         






  4. Next, click Create Event Grid Subscription“.

  5. On the “Create Event Subscription” part of the portal, use the following options/values.

    1. Name: EvtSub-PolicyAlert (this is customizable but should following naming conventions)

    2. Event Schema: Event Grid Schema

    3. Topic Types: Search for “policy” and select Microsoft PolicyInsights.

    4. Source Resource: Azure Subscription *(might be autofilled)

    5. System Topic Name: egst-PolicyAlert (or the name you created) *(might be autofilled)

    6. Event Types: Policy Compliance State Created, Policy Compliance State Changed

    7. Endpoint Type: Azure Function * (might be autofilled)

    8. Endpoint: PolicyAlertTrigger1 (Your Azure Function created in the previous step) * (might be autofilled)

      1. F5i.png

         








AAD App Registration


An AAD App registration is needed for the Log Ingestion API, which is used to write the data to Log Analytics.


( https://learn.microsoft.com/en-us/azure/azure-monitor/logs/tutorial-logs-ingestion-portal) – please follow the link for additional information about log ingestion



  1. In the Azure global search, type Azure Active Directory.

  2. Once in your AAD tenant, under the Manage sub menu, click “App Registrations“.

  3. Click +New Registration and use the following settings:


    1. Name: AR-PolicyAlert-Ingestion

    2. Supported account types: Accounts in this organizational directory only

    3. Click RegisterAAR3d.png

       



    4. Once you create the new registration, click to open the registration. Be sure to store your Tenant ID and App IDreminders” in a text document (like notepad) so you can copy/paste it in a later step. The snapshot below just has the values blanked out but shows the IDs that are needed by being highlighted.DJBartles_16-1667850167229.png

    5. When inside of your new registration, click “Certificates and Secrets” and then go to “Client Secrets“.AAR3h.png

       



    6. Create a new client secret. Put the name you want to use in the “Description” and add this to your reminders notepad. Ensure that you have the correct expiration for the secret according to your organization. Click Add.

    7. Once the new secret is created, immediately copy the secret value and add it to the reminders notepad. This value will only be visible for a short time and you cannot get access to it again.



Data Collection Endpoint ( https://learn.microsoft.com/en-us/azure/azure-monitor/logs/tutorial-logs-ingestion-portal)


The Data Collection Endpoint is part of the Log Ingestion for writing the PolicyInsights data to Log Analytics.



  1. In the Azure global search , type Monitor. Select Azure Monitor and open in the Portal.

  2. On the left menu bar, under the Settings submenu, click “Data Collection Endpoints“.

  3. Click + Create to create a new Data Collection Endpoint (DCE)


    1. Name the DCE, in this case, DCE-PolicyAlerts

    2. Ensure that the correct subscription is selected

    3. Ensure that the correct resource group is selected

    4. Ensure that the correct region is selected

    5. Add any appropriate tags.

    6. Review and Create

      1. DJBartles_5-1669819453849.png

         







Custom Log (Table) and Data Collection Rule ( https://learn.microsoft.com/en-us/azure/azure-monitor/logs/tutorial-logs-ingestion-portal)


These steps will show how to create a customer data table in the Log Analytics Workspace and create a Data Collection Rule.



  1. Go to the Azure global search bar and type in Log Analytics Workspaces (LAW)

  2. If you do not have a current Log Analytic workspace, please create one before moving on to the next step.

  3. Open the LAW, from the lefthand side menu, under settings, select Tables.

  4. Select +Create, and then select New custom log (DCR-based)

    1. CL4a.png

       





  5. On the Create page, name the Table in this case, the table is called PolicyAlert

  6. Next, click “Create Data Collection Rule” blue hyperlink. (This Table which we just named, will be where you will be sending your policy data.)

    1. DJBartles_3-1669819291029.png

       


       





  7. Give the new DCR a name, in this case, it is named DCR-PolicyAlert. Click Done.

    1. DJBartles_0-1669819217039.png

       




       





  8. When back at the “Create custom log” page, select the Data Collection Rule, DCR-PolicyAlert, you just created if it did not auto populate.

  9. Select the Data Collection Endpoint (DCE) that you created earlier in Azure Monitor. In this case the DCE is DCE-PolicyAlerts

    1. CL9a.png

       







  1. Click Next to continue

  2. Setting up the data format/schema is the next step, which can be a little confusing. In this step, you will upload/paste in a sample json that represents the data that you will be using. If you would like to use the exact same data that we are using in this article, HERE is a sample file you can use for this upload. This sample data file will basically define the data structure in your LAW Table. This will need to be coordinated with your PowerShell code that is completed in a later step.DJBartles_7-1669819547018.png

     






    1. When you upload the sample file (or any other), you will likely get an error about the “TimeGenerated” data field. This is an expected error. This field is required in Custom Tables so you will next need to “transform” a data/time field in our sample data to suit the Custom Table. If you get this error, click the “Transformation Editor” in the previous image. The following image shows the query you can use. Type in the following KQL. Run the query and then click Apply.DJBartles_0-1669818970747.png

       



    2. You should now be able to click Next to continue.

    3. Click the Create. *Please note you will see the table name appended with _CL. In this case, PolicyAlert_CL is the name of the custom table.



Setup Access to the Data Collection Rule (https://learn.microsoft.com/en-us/azure/azure-monitor/logs/tutorial-logs-ingestion-portal)


This process will setup access to the Data Collection Rule as part of writing the data to Log Analytics via the Log Ingestion API.



  1. Go to Azure global search. Type Monitor and select it.

  2. Within Azure Monitor, go to Settings, and click “Data Collection Rules” on the left menu.

  3. Find the DCR created in the previous step and click on it.

  4. Once on the DCR you created, click “Access Control (IAM)” from the left hand menu.

    1. DCR4a.png

       







  1. Find Grant access to this resource, click “Add Role Assignment“.

  2. Add the role “Monitoring Metric Publisher” to your previously created AAD App Registration. In our example, we named ours “AR-PolicyAlert-Ingestion“. Submit the role assignment when completed.


    1. For Role, search for Monitoring Metrics Publisher. Select it. Click Next.

    2. For Members, select the name of the app registration from earlier in the solution. In this case, AR-PolicyAlert-Ingestion

    3. Click Next

    4. Click Review and Assign.

      1. DCR6e.png

         







Setup Function App Managed Identity to have Access to Key Vault


Here you will setup the managed identity for Key Vault so it can read your secret store in Key Vault.



  1. Setting up access to your Key Vault secrets to the Managed Identity for your Function App can be done from the Key Vault or from the Function App Identity page. In this article we will describe how to do it from the Function App Identity page. This access control will allow your Function App to query the Key Vault for the secret that you stored from the previous step.

  2. Go to the Global Azure search bar and type Function App

  3. Go to your Function App that you created for this solution.


    1. From the Settings menu, click the Identity button on the left menu.


  4. You should see your managed identity there from the previous step. Now click “Azure Role Assignment“.KVMI4a.png

     



  5. Now click “Add Role Assignment“.

  6. As shown in the following image, choose Scope=Key Vault, Resource=(Your Keyvault), Role=”Key Vault Secrets User”.

  7. Assign the role and submit the change.


Setup PowerShell Operating Environment in Function App


This process will prepare your Azure Function powershell environment.



  1. Go to your Function App that you created in a previous step.


    1. Under Functions, click on “App Files” on the left side.


  2. There will be a drop-down menu near the top middle of your screen. Choose the “requirements.psd1” option.DJBartles_0-1668015548475.png

     



  3. Edit the text in the file so that you have an entry for “Az.Accounts” and “Az.KeyVault“, as indicated in the graphic, and then click “Save“. This will tell your function to automatically install the PowerShell modules specified when the app restarts next.



  1. With the Az module line uncommented, you now need to restart the function app.


    1. To do this, click on “Configuration” on the left bar, under Settings.


  2. Navigate to the “General Settings” tab on the top right.

    1. PSOE5a.png

       





  3. Find the setting “PowerShell Core Version” on the right and change it to a different value and click save. Acknowledge the change and wait until it completes. Once the task is completed, reverse that setting back to its original value and click save again. In my example, I went from “PowerShell 7.2” to “PowerShell 7.0” back to “PowerShell 7.2”. This process is just to force Azure to restart the Function App so the Az Modules will load.



  1. Loading the PowerShell modules could take a few minutes so do not immediately expect it be there right after the restart completes.

  2. When the PowerShell is executed in your function app, it should have the PowerShell Az modules loaded by default.


Setup PowerShell in Function


Here you will setup the powershell code within the Azure Function.



  1. In this section we will setup the PowerShell code and operating environment with the Function created in the previous step. For our reference architecture in this article, you can download the PowerShell source code from HERE.

  2. Go to the Function App, FN-PolicyAlert or what you called the Function App for this solution.

  3. From the left-hand menu, under Functions, select Functions.

  4. You should be able to see on the right-hand side, the function that was created earlier in this solution called, “PolictAlertTrigger1“, and click on it to open it.

    1. PSF5a.png




 



  1. On the left menu, under Developer, click “Code + Test“.

  2. This will open the code editor. Grab all of the code from HERE and paste it into the editor. You need to make sure that you have verified that all of the PowerShell variables at the top of the code are filled in with your appropriate values. Many of these values will come from the “remindersnotepad text file mentioned in earlier steps.

  3. Need the DCE logs ingestion URI – go to the DCE in monitor and copy

  4. For the DCR immutable ID, go to the DCR, click on JSON view

    1. PSF8a.png

    2. PSF8b.png




Setup Alerting


Here you will setup the alerting mechanism within Azure Monitor.



  1. Go to the Log Analytics Workspace (LAW) where you have sent the Custom Logs that you created earlier which in the previous steps in this example was called, LAW-. In this example, the custom log table is called PolicyAlert_CL.

  2. In a blank query space, please type or paste in the following KQL query to query the custom log table for the policy compliance event changes


 


 

PolicyAlert_CL
| where event_type =~ "Microsoft.PolicyInsights.PolicyStateCreated" or event_type =~ "Microsoft.PolicyInsights.PolicyStateChanged"
| where compliancestate =~ "NonCompliant"
| extend TimeStamp = timestamp
| extend Event_Type = event_type
| extend Resource_Id = subject
| extend Subscription_Id = subscriptionid
| extend Compliance_State = compliancestate
| extend Policy_Definition = policydefinitionid
| extend Policy_Assignment = policyassignmentid
| extend Compliance_Reason_Code = compliancereasoncode
| project TimeStamp, Resource_Id, Subscription_Id, Policy_Assignment, Policy_Definition, Compliance_State, Compliance_Reason_Code

 


 


 



  1. Run the Query by click the Run button to ensure that it is working properly.


DJBartles_0-1669745200949.png


 



  1. Click on New alert rule to create the alerting for the policy compliance event changes.


DJBartles_1-1669656974839.png



  1. The alert rule that is being created will be based on the current log query. The Scope will already be set to the current resource.

  2. Next, on the Condition tab, the Log query will already be populated with the KQL query that we entered.

  3. On the Condition tab, the Measurement section will define how the records from the log query will be measured. As this query has one or more summarized columns, the Table rows option will be shown. Leave Table rows as the selection. Aggregation granularity refers to the time interval over which the collected values are aggregated. By default, the value is set to 5 minutes.


For this example, leave this set to 5 minutes.



  1. On the Condition tab, the Configure dimensions section allows you to separate alerts for different resources. In this example, we are only measuring one resource so there is no need to configure.


DJBartles_2-1669656974846.png


 



  1. On the Condition tab, the Configure alert logic is where we can configure the Operator and Threshold values against the value returned from what we are measuring. In this case, select the operator as greater than and the threshold value as 0. We would like to be alerted when there is a policy compliance change. With these settings, we will be notified when the threshold value is greater than 0; so every time a change occurs.

  2. Next, we will need to select a frequency of evaluation value; how often do we want the log query to run and be evaluated. Remember, the cost for the alert rule increases the lower the frequency is meaning if I have something run with a frequency of every minute, it will cost more than something that I run every 10 minutes. There will be a pro tip that shows an estimated cost. In this example, select 5 minutes.


DJBartles_3-1669656974847.png



  1. The remaining steps are all about how you want to be notified by using Action Groups. The procedures to create Action Groups is well documented HERE. For our example, this image shows a basic Action Group that just sends an email. In the example, the name of the Action Group is ag-PolicyAlert-Action. Also, the Action Group will contain the email address of the group/individual that will be notified when a policy compliance change occurs. Please keep these items in mind when designing your Action Group.DJBartles_0-1669730517491.png


Wrap-Up


Now that this setup has been complete, you should be able to receive an email notification whenever an alert is generated from an Azure Policy Compliance status change. There are several possible solutions to use and ways to expand functionality in this architecture so stay tuned for future posts that expand on this approach.


References


Azure Key Vault: Quickstart – Create an Azure Key Vault with the Azure portal | Microsoft Learn


Log Ingestion API: Logs Ingestion API in Azure Monitor (preview) – Azure Monitor | Microsoft Learn


Log Ingestion API Tutorial: Tutorial – Send data to Azure Monitor Logs using REST API (Azure portal) – Azure Monitor | Microsoft Learn


Send Custom Events to Azure Function: Quickstart: Send custom events to Azure Function – Event Grid – Azure Event Grid | Microsoft Learn


Azure PolicyInsights Data: Get policy compliance data – Azure Policy | Microsoft Learn


Azure Policy State Change Events: Reacting to Azure Policy state change events – Azure Policy | Microsoft Learn


Azure Function Powershell Dev: PowerShell developer reference for Azure Functions | Microsoft Learn


Azure Action Groups: Manage action groups in the Azure portal – Azure Monitor | Microsoft Learn


 

Improving interoperability between the Visio web and desktop apps

Improving interoperability between the Visio web and desktop apps

This article is contributed. See the original author and article here.

We’ve been working diligently to address user feedback about desktop files not opening in the Visio web editor. Today, we’re excited to announce that we’ve made a number of improvements towards a more seamless user experience between the Visio web and desktop apps.



What’s new



Visio users can now open and edit files containing shapes with the following properties in Visio for the web:



  • Rectangle gradients

  • More fill and line patterns

  • OLE objects

  • Data graphics

  • Shadow text

  • Vertical text

  • Double underlines

  • Shadow effects

  • Reflection effects

  • Glow effects

  • Soft edges

  • Bevel effects

  • 3-D rotations

  • Perspectives and 3-D rotation perspectives

  • Compound line patterns

  • 1-D and 2-D protection


*Note: Some interactions have been optimized to ensure they work in Visio for the web.



An image of a component architecture diagram demonstrating how shapes with gradients appear in Visio for the web.An image of a component architecture diagram demonstrating how shapes with gradients appear in Visio for the web.


Looking ahead



Our goal is to always make it easier for you to think and collaborate on ideas visually—whether improving processes, developing new reporting structures, or brainstorming concepts. With this new update, we hope you will find it easier to work with colleagues, from whichever app you prefer.



We are continuing to make improvements on this front and look forward to introducing support, for example, for diagrams containing shapes with layers or diagrams with more than 1,000 shapes. To learn more about the properties that can still lead to compatibility issues when editing files in Visio for the web, please see Why can’t I open a diagram for editing in Visio for the web?



Watch our on-demand webinar



Watch the event recording here: Editable for all: How to create diagrams that work virtually anywhere and across Visio versions. During this session, we show you how to create Visio files that can be accessed by all stakeholders regardless of which Visio version (or app) you use. We also discuss how to avoid or quickly address any web alerts that prevent in-browser edits, so you can jump right back into your workflow.



We are listening!



We look forward to hearing your feedback and learning more about how we can ensure a more seamless user experience between Visio apps. Please tell us what you think in the comment section below. You can also send feedback via the Visio Feedback Portal or directly in the Visio web app using “Give Feedback to Microsoft” in the bottom right corner.



Did you know? The Microsoft 365 Roadmap is where you can get the latest updates on productivity apps and intelligent cloud services. Check out what features are in development and coming soon on the Microsoft 365 Roadmap homepage or learn more about roadmap item, Improved interoperability between the Visio web and desktop apps.

Microsoft named a Leader in 2022 Gartner® Magic Quadrant™ for Unified Communications as a Service, Worldwide

Microsoft named a Leader in 2022 Gartner® Magic Quadrant™ for Unified Communications as a Service, Worldwide

This article is contributed. See the original author and article here.

We are honored to announce that Gartner® has recognized Microsoft as a Leader in the 2022 Gartner® Unified Communications as a Service (UCaaS) Magic Quadrant™ report.

The post Microsoft named a Leader in 2022 Gartner® Magic Quadrant™ for Unified Communications as a Service, Worldwide appeared first on Microsoft 365 Blog.

Brought to you by Dr. Ware, Microsoft Office 365 Silver Partner, Charleston SC.

Azure Marketplace new offers – November 30, 2022

Azure Marketplace new offers – November 30, 2022

This article is contributed. See the original author and article here.

We continue to expand the Azure Marketplace ecosystem. For this volume, 94 new offers successfully met the onboarding criteria and went live. See details of the new offers below:


 

















































































































































































































































































































































Get it now in our marketplace


Applications-CloudNativeStackVirtualMachineImage.png

Cloud Native Stack Virtual Machine Image: NVIDIA’s Cloud Native Stack VMI is a GPU-accelerated virtual machine image (VMI) that enables advanced functionality for developers, including better GPU performance, utilization, and telemetry.


Applications-Demography.png

Demography: TensorGo’s Demography API detects faces in videos and returns gender, ethnicity, and estimated age in real time by using advanced machine learning models.


Applications-FaceDetection.png

Face Detection: TensorGo’s Face Detection API uses computer vision to detect faces in an image or video and return a bounding box around individual faces.


Applications-FacialLandmarks.png

Facial Landmarks: TensorGo’s Facial Landmarks API localizes a human face in an image or video and tracks the facial points to predict and track behaviors over time.


Applications-FTPServerBasiconWindowsServer2019.png

FTP Server Basic on Windows Server 2019: This preconfigured image from Tidal Media includes a minimized FTP Server on Windows Server 2019. FTP (File Transfer Protocol) lets you transfer computer files between a server and client.


Applications-HeartRateEstimation.png

Heart Rate Estimation: TensorGo’s Heart Rate Estimation API uses non-invasive computer vision techniques to estimate the heart rate of a subject in a video.


Applications-Hyper-VonWindowsServer2022.png

Hyper-V on Windows Server 2022: Cloud Infrastructure Services’ Hyper-V Server lets you deploy nested virtualization on Microsoft Azure to host multiple operating systems and containerized virtual machines.


Applications-JenkinsonDebian10withSupport.png

Jenkins on Debian 10 with Support: This virtual machine from ATH provides Jenkins 2.361 on Debian 10. Jenkins is an open-source automation server that runs servlet containers and version control tools for CI/CD tasks.


Applications-JenkinsonDebian11withSupport.png

Jenkins on Debian 11 with Support: This virtual machine from ATH provides Jenkins 2.361 on Debian 11. Jenkins is an open-source automation server that runs servlet containers and version control tools for CI/CD tasks.


Applications-JenkinsonUbuntu1804LTSwithSupport.png

Jenkins on Ubuntu 18.04 LTS with Support: This virtual machine from ATH provides Jenkins 2.361 on Ubuntu 18.04 LTS. Jenkins is an open-source automation server that runs servlet containers and version control tools for CI/CD tasks.


Applications-JenkinsonUbuntu2004LTSwithSupport.png

Jenkins on Ubuntu 20.04 LTS with Support: This virtual machine from ATH provides Jenkins 2.361 on Ubuntu 20.04 LTS. Jenkins is an open-source automation server that runs servlet containers and version control tools for CI/CD tasks.


Applications-LicensePlateRecognition.png

License Plate Recognition: TensorGo’s License Plate Recognition API uses optical character recognition to identify vehicles in traffic and extract their license plate numbers.


Applications-LowLightandInfraredFaceDetection.png

Low Light and Infrared Face Detection: TensorGo’s Low Light IR Face Detection API overcomes environmental conditions to identify faces and people from video captured in low lighting or by an infrared (IR) camera.


Applications-LTAPPSTimesheet.png

LTAPPS Timesheet: The LTAPPS Timesheet is a time-tracking app for Microsoft Teams and SharePoint, whether desktop or mobile, that lets employees and managers track hours, overtime, and absences.


Applications-PeopleDetection.png

People Detection: TensorGo’s People Detection API lets you detect people moving in videos, regardless of which direction they are facing, the light levels, and their clothing.


Applications-PeopleSegmentation.png

People Segmentation: TensorGo’s People Segmentation API provides real-time detection and segmentation of people in an image or video, allowing you to home in on individuals in crowds.


Applications-PoseEstimation.png

Pose Estimation: TensorGo’s Pose Estimate API can track the pose and orientation of a person or object from an image or video to determine whether the person is standing, sitting, or performing activities.


Applications-RockyLinux8-Generation2VirtualMachine.png

Rocky Linux 8 – Generation 2 Virtual Machine: ProComputers has preconfigured this minimal, ready-to-use generation 2 virtual machine containing Rocky Linux 8, cloud-init, and the Microsoft Azure Linux Agent.


Applications-RockyLinux9Minimal.png

Rocky Linux 9 Minimal: Ntegral has preconfigured this minimal virtual machine containing Rocky Linux 9 for use on Azure. Rocky Linux is used for enterprise cloud environments including Node.js, web, and database applications.


Applications-RubyonRails.png

Ruby on Rails: Data Science Dojo has packaged Ruby on Rails on Ubuntu 20.04 to enable students, developers, and organizations focused on web-based products to easily create Ruby-based data science apps.


Applications-SalesKen-AIAssistance.png

Salesken – AI Assistance: Salesken is a real-time, Azure-based sales assistant that provides AI-driven conversation insights from contextual clues and talking points, enabling you to navigate sales discussions with ease.


Applications-SmartDA.png

SmartDA: SmartDA is an app for Microsoft Teams that provides a versatile and personalized internal assistant using Intumit’s conversational AI to answer user questions about Microsoft 365.


Applications-TrafficAnalyzer.png

Traffic Analyzer: TensorGo’s Traffic Analyzer API detects vehicles, pedestrians, and road signs from images or videos, identifying the location and number of vehicles and people.


Applications-Ubuntu20045LTS.png

Ubuntu 20.04.5 LTS: Ntegral provides this preconfigured image of Ubuntu 20.04.5 LTS for use on Microsoft Azure. Ubuntu is a popular Linux distribution used to serve enterprise workloads such as Node.js, web, and databases.


Applications-VeritasInfoScalefortheCloud.png

Veritas InfoScale for the Cloud (Ireland): Veritas’s InfoScale is a comprehensive availability and storage management solution delivering a common platform across physical, virtual, and multi-cloud infrastructure.


Applications-VeritasInfoScalefortheCloud.png

Veritas InfoScale for the Cloud (Singapore): Veritas’s InfoScale is a comprehensive availability and storage management solution delivering a common platform across physical, virtual, and multi-cloud infrastructure.


Applications-VeritasInfoScalefortheCloud.png

Veritas InfoScale for the Cloud (United States): Veritas’s InfoScale is a comprehensive availability and storage management solution delivering a common platform across physical, virtual, and multi-cloud infrastructure.


Applications-VSMforMicrosoftTeams.png

VSM for Microsoft Teams: Virsae’s VSM Everywhere is a centralized network management and optimization platform to ensure that your Microsoft Teams and contact center environments operate at peak performance.



Go further with workshops, proofs of concept, and implementations


ConsultingServices-AzureARCandAzureStackHCI1-DayWorkshop.png

Azure ARC and Azure Stack HCI: 1-Day Workshop: Available only in German, Bechtle’s practical workshop will present an overview of Microsoft Azure ARC and Azure Stack HCI, as well as use cases to manage your infrastructure uniformly.


ConsultingServices-AzureDevOpsBestPractices1-DayWorkshop.png

Azure DevOps Best Practices: 1-Day Workshop: Learn from PetaBytz about best practices, tips, tricks, and working demos for Microsoft Azure DevOps to get the most out of your Agile software development stack.


ConsultingServices-Business-drivenWebAnalyticsSolution5-WeekImplementation.png

Business-driven Web Analytics Solution: 5-Week Implementation: Polestar will deliver dashboards built on Microsoft Power BI to help you track web statistics across your analytics stores by using Microsoft Azure Data Factory and Azure Services.


ConsultingServices-CloudDeploymentMigrationandModernizationServices.png

Cloud Deployment, Migration, and Modernization Services: ISM’s Cloud Deployment, Migration, and Modernization services are targeted to optimize costs and line-of-business delivery for your enterprise IT landscape by using greenfield implementations on Microsoft Azure.


ConsultingServices-CloudEnd-to-EndManagedServices.png

Cloud End-to-End Managed Services: ISM offers complete management of part or all of your Microsoft Azure environment, whether standalone or hybrid, in alignment with ITIL and the Microsoft Cloud Adoption Framework.


ConsultingServices-CloudMigrationService3-WeekImplementation.png

Cloud Migration Service: 3-Week Implementation: NCS will accelerate your migration and adoption of Microsoft Azure through this low-risk, cost-optimized service that includes strategic planning, technical design, delivery validation, and more.


ConsultingServices-DevSecOpsAutomatedPipelineProcessImplementation.png

DevSecOps Automated Pipeline Process: Implementation: Vismaya India will analyze your traditional automated pipelines and manual development processes, then implement an automation solution built on Microsoft Azure DevOps.


ConsultingServices-HybridCloudSyncServices.png

Hybrid Cloud Sync Services: AlgoSystems’ Hybrid Cloud Sync Services will provide you with a modern and reliable file solution built on Microsoft Azure File Sync and Azure Backup to support your business and users.


ConsultingServices-InfrastructureasCodeonAzureDevOps1-WeekImplementation.png

Infrastructure as Code on Azure DevOps: 1-Week Implementation: Drawing on expert experience, PetaBytz will use industry best practice blueprints to drive your adoption of Microsoft Azure DevOps to optimize delivery of business products and services.


ConsultingServices-InnovationJumpstart6-WeekProofofConcept.png

Innovation Jumpstart: 6-Week Proof of Concept: Quisitive’s Innovation Jumpstart helps organizations build a ready-to-execute app roadmap and rapidly prototype a solution built on Microsoft Azure.


ConsultingServices-NetworkManagementonAzureforMulticloud2-WeekProofofConcept.png

Network Management on Azure for Multi-cloud: 2-Week Proof of Concept: Metanext will implement Aviatrix for Microsoft Azure to improve your management of networks and security in your multi-cloud environment. This offer is available only in French.


ConsultingServices-QuickstartServiceforPrismaCloud6-MonthImplementation.png

Quickstart Service for Prisma Cloud: 6-Month Implementation: Palo Alto Networks’ Quickstart Service for Prisma Cloud CPSM delivers visibility, compliance, and governance for threat detection and behavior analytics on Microsoft Azure.



Contact our partners



Alfresco Consultation: 2-Hour Briefing



Analytics Solution: 4-Week Assessment



Angles for SAP



Anomaly Detection



Application Modernization on Azure: 4-Week Assessment



ARGOS Cloud Security for Microsoft Sentinel



Azure Database Migration: 2-Hour Briefing



Azure Migrate: 1-Hour Briefing



Beak Vulnerability Detection & Response



Canopy Remote Device Management Software



Cisco Cyber Vision



Clobotics REA



Cloud Discovery, Assessment, and Advisory Services



Cloud Readiness: 8-Day Assessment



Data Discovery for Manufacturing and Construction: 3-Week Assessment



Data Governance: 8-Week Assessment



DevOps (CI/CD) for z/OS Mainframes



DevOps Consulting



DrAid AI Cognitive Service



Enable Customer Success for Cloud Adoption Framework: 4-Hour Briefing



FONS Galen



Forcepoint Email Security V8.5.5



Hf.flow Application



Holistic Business Case Solution for Microsoft 365: 4-Week Assessment



Honeywell Forge Performance+ Industrial Asset Performance



IBM Cloud Pak for Data on ARO – BYOL



Info-Tech Mobile Attendance App



Managed GitHub Enterprise Server



mirro.ai Mood Analyzer (SaaS)



Panoptica Cloud Native Application Security For Developers



Percy Intranet



Prime Unified Messaging Platform



Process Runner for Microsoft 365



Proximus Identity Governance: Assessment



Quorum Cyber Managed eXtended Detection & Response (XDR)



ReversingLabs File Enrichment



SAP & Microsoft Analytics Combined: 5-Day Assessment



Seavus Application Modernization: 4-Week Assessment



Spark Unified Media Platform



Spicy Managed Microsoft Sentinel Service



SPiDER TM (Japan)



SPiDER TM 5.5 (South Korea)



Spirent for Azure Public Multi-access Edge Compute



Stipra



SUSE Linux Enterprise Server 15 SP4 – Hardened BYOS (x86_64)



Sustainability Data Insights



SymphonyAI Sensa – SensaAML (AI-based Anti-Money Laundering)



US Exports Bill of Lading



US Imports Bill of Lading



US Imports Bill of Lading Commodity Details



Veridis Carbon Management



WeTrack



Windows 365 Cloud PC



Yanomaly Asset Health Monitoring