Join us at Microsoft Build on May 25-27

This article is contributed. See the original author and article here.

Get ready to learn, connect, and code at Microsoft Build 2021 on May 25-27, 2021! Explore what’s next in tech and the future of hybrid work, discover solutions, and add to your toolbox.


 


Whether you’re new to the industry or are a seasoned developer, the Learning Zone is the place to build your skills and learn something new. Through a variety of learning activities, workshops, sessions, and more, you’ll find something that fits your skill level and personal learning style to expand your knowledge.



  • Cloud Skills Challenge – If you like competition, sign up and complete one or more unique challenges tailored to different themes across our cloud services and be entered for a chance to meet one of our product leaders – Amanda Silver, Donovan Brown, Scott Hanselman, or Mark Russinovich, just to name a few!

  • Learn Live – If hands-on learning is your preferred style, tap into these self-paced modules while you’re guided by subject matter experts.

  • Student Zone – This isn’t just for ‘traditional students.’ If you’re new to the industry or looking to make your next career move in tech, be sure to check out these sessions.


 


In addition, join us in the Ask the Experts: Microsoft Certification sessions in the Connection Zone to get your Microsoft Certification questions answered by our panel experts.



 


You’ll also find a Table Topics session on How the community and certifications can help you achieve more in the Connection Zone – an opportunity to connect with the community and our experts to tap into help with your professional development.


 


To get the latest event details, follow us on LinkedIn and Twitter (#MSBuild, #CloudSkillsChallenge, #ProudToBeCertified). After the event, be sure to continue what you start from one free convenient location – Microsoft Learn. You’ll find deeper technical content, training options, certification details, and developer communities.


 


Now that you know what we have in store for you, register for Microsoft Build today and join us in the Learning Zone on May 25-27 so you can build your learning journey.

End-to-End IoT analytics and machine learning with Azure Data and AI services

End-to-End IoT analytics and machine learning with Azure Data and AI services

This article is contributed. See the original author and article here.

The Internet of Things (IoT) technology stack in Operational Technology (OT) is widely used across various industries, including oil & gas, manufacturing, utilities, and natural resources, for solving operational challenges and delivering mission-critical insights and analytics.


 


More and more organizations are leveraging Microsoft’s Azure cloud platform to perform large scale analytics and machine learning using data from IoT assets, something that had not been simple to do using traditional systems such as Scada and Historians.


 


In this article, we preview an end-to-end Azure Data and AI cloud architecture that enables IoT analytics. This article is based on our 3-part blog series on the Databricks Blog site. You can find more information and code samples starting with


 



Here is the overall architecture discussed in this article and the Databricks blog series:


 


teaserpicture.PNG


For making this article applicable on a common IoT use case, consider the scenario of balancing optimal short-term utilization of an asset, such as a wind turbine, against its long-term maintenance costs.


balance.PNG


In order to develop insights on short-term optimization as well as longer term maintenance costs, various data sources need to be considered and ingested into the cloud for centralized storage and analysis. Here are a few Azure cloud services to consider, depending on whether the data sources can be stream or batch processed.


 



For the wind turbine scenario, streaming data can be the sensor data collected from the turbines, while structured data can be maintenance and failure data collected in a batch process.


 


Once the data sources are ingested into Azure, there are a few options, again depending on stream or batch processing, to process and store the data. In this architecture, the Delta format in Azure Databricks backed by Azure Data Lake Gen 2 is the preferred data format for large-scale IoT data sources: Delta Lake and Delta Engine guide – Azure Databricks – Workspace | Microsoft Docs


 



Once the data is ingested, processed and stored into Delta format, Azure Databricks can be used for big data analytics including data engineering and data science using Spark. As a common pattern, multiple zones within the data lake and aggregations are highlighted below identified through


 



  • Bronze for raw granular IoT data

  • Silver for aggregated data, commonly used for machine learning and data science

  • Gold for enriched data ready for analytics and reporting purposes


Data engineers can use Azure Databricks and create 3 Delta tables corresponding to these three zones. Users can use programming languages namely Python, Scala, R, and SQL in Azure Databricks for accelerated data engineering and data science development.


 


Azure Machine Learning can be used for machine learning, most commonly together with Azure Databricks, in this IoT architecture. For example, Azure Databricks can be used with Spark to engineer features and aggregate data. Then Azure Machine Learning can be used to build models through code, drag-and-drop, or even automated machine learning. In addition, Azure Machine Learning can be used to deploy and operationalize machine learning models.


 


For the wind turbine scenario, the bronze Delta table could be the granular IoT sensor data from the turbines while the silver Delta table is the aggregated (by the hour for example) data. Then Azure Databricks can be used to perform feature engineering and feature selection to build a machine learning and analytics ready dataset. This dataset would then be loaded in Azure Machine Learning to build a predictive maintenance model or a power generation prediction model.


 


Finally, once the predictions and gold enriched data is created in the gold Delta table with Azure Databricks, it can be loaded into Azure Synapse Analytics for BI analytics and reporting scenarios together with Power BI. Azure Data Explorer provides real-time operational analytics so IoT data can be streamed directly from IoT Hub or Event Hub to Data Explorer.


 


In summary, this article covered the end-to-end steps for enabling IoT data analytics and machine learning on the Azure cloud platform, including some best practices, recommended services, and application with wind turbine operations use case. This blog series has the full details and provides code samples as well: Articles by Hubert Duan – The Databricks Blog.


 


 

Join us for an AMA!

Join us for an AMA!

This article is contributed. See the original author and article here.

2021-05-19 17_07_01-PowerPoint Slide Show  -  AMA banner.pptx.png


 


The Microsoft Defender for Office 365 team wants to hear from you! We’re excited to invite you to join us for a Tech Community Ask Microsoft Anything (AMA). Our team will be on hand to answer any of your questions about Microsoft Defender for Office 365, Exchange Online Protection, and email and collaboration security in general, so come prepared!


 


 


The AMA will take place Thursday, May 27, 2021, from 9:00-10:00am Pacific Time. We hope to see you there!


 


Use the link below to add a reminder to your calendar and to join the discussion.


https://aka.ms/ama/DefenderO365  

End of synchronization for WSUS 3.0 SP2

This article is contributed. See the original author and article here.

On October 31, 2021, Windows Server Update Services (WSUS) 3.0 Service Pack 2 (SP2) will no longer synchronize and download updates. 


WSUS is key to the Windows servicing process for many organizations. Whether being used standalone or as a component of other products, it provides a variety of useful features including automating the download and installation of Windows updates.


Extended support for WSUS 3.0 SP2 ended on January 14, 2020, in alignment with the end of support dates for Windows Server 2008 SP2 and Windows Server 2008 R2. It is, however, still possible to synchronize and download updates from Microsoft using WSUS 3.0 SP2.


WSUS relies on several different components for secure communication. The protocol that is used for a given connection depends on the capabilities of the associated components. If any component is out of date, or not properly configured, the communication might use an older, less secure protocol. Microsoft is transitioning all endpoints to the more secure TLS 1.2 cryptographic protocol. WSUS 3.0 SP2 does not support this newer protocol. As a result, any organizations still using WSUS 3.0 SP2 must migrate to a currently supported version of WSUS by October 31, 2021.


For guidance, see Deploy Windows Server Update Services.


 

Experiencing Data Latency issue in Azure Portal for Many Data Types – 05/20 – Resolved

This article is contributed. See the original author and article here.

Final Update: Thursday, 20 May 2021 12:25 UTC

We’ve confirmed that all systems are back to normal with no customer impact as of 05/20, 11:50 UTC. Our logs show the incident started on 05/20, 11:21 UTC and that during the 29 minutes that it took to resolve the issue some of the customers may have experienced data latency issue in China East 2 region.
  • Root Cause: The issue is because of sudden increase of the requests in one of our dependent service.
  • Incident Timeline: 29 Minutes – 05/20, 11:21 UTC through 05/20, 11:50 UTC.
We understand that customers rely on Application Insights as a critical service and apologize for any impact this incident caused.

-Srikanth

Microsoft Identity Platform community call – April 2021

Microsoft Identity Platform community call – April 2021

This article is contributed. See the original author and article here.

EventHubBlockDiagram.png


 


Call Summary:


This month’s in-depth topic:  Get Change notifications delivered via Azure Event HubsBrief introduction to Graph change notification (webhooks) and Change notifications delivered via Azure Event hubs.   Microsoft Program Manager presenters – George Juma and Kalyan Krishna.  This session was delivered and recorded on April 15, 2021.  Live and in chat Q&A throughout call. 


 


 


In-depth topic:


Get Change notifications delivered via Azure Event Hubs


 


A new Microsoft Graph platform capability gives developers using change notification features in Microsoft Graph the option to get change notifications delivered to their applications more quickly and at scale using Azure Event hub.   With Microsoft Graph, changes (in data in Microsoft Graph) are tracked with webhooks, a.k.a., change notifications.   Presently you get change notifications via a API REST end-point exposed on the internet. 


 


Azure Event Hub is a real-time events ingestion and distribution service built for scale.  Ideal for high throughput (no dropped notifications due to timeouts), no public URL (security), no missed notifications if temporally off-line.    Operationally, Microsoft Graph Change Tracking places notifications in Event Hub and your app retrieves messages from Event Hub rather than from publicly exposed end points on the internet.   You need not poll for changes, change notifications are pushed to your app.  You need only subscribe to notifications.     


 


In the end-to-end demo, the presenter steps through app registration, provisioning of an Azure Event Hub, an Azure Storage and a container, add a Shared Access Policy, create an Azure Key Vault (secure connection strings), add subscription connection string and IDs of resources to the app.    


 


Resources:


 



 


Actions:  



 


Stay connected:



 

Microsoft Graph community call – May 2021

Microsoft Graph community call – May 2021

This article is contributed. See the original author and article here.

TitleSlideThumbnail.png


 


This month’s community call features presentations on Application Consent Deep Dive (demystifying permissions and consent when accessing Microsoft Graph) and on Considerations for Creating Online Meetings (Integration into Outlook or into other 3rd party experiences).  Q&A in this call after demos, at end and in chat.  The call was hosted by Brian T. Jackett (Microsoft) | @BrianTJackett. Microsoft Presenters were Philippe Signoret and Fabian Williams.  Recorded on May 4, 2021.


 


Topics:


App Consent Deep Dive – demystifying permissions and consent when accessing Microsoft Graph.  Most Graph permissions allow access to lots of data.  Learn how apps and services access Microsoft Graph resources – application permissions and delegated permissions, requesting permissions, granting permissions, and restricting data access scope.  General concepts – direct access, access on behalf of user, app and user authorizations, permission types, service principals, consents and more.


 


Considerations for creating Online Meetings – Based on specific meeting requirements, create an Online Meeting through the Calendar Events API or through the Cloud Communications API (Teams).  Step through 7 questions to ask yourself to decide when to use what method.  The decision is largely based on the need for a rich integrated Microsoft client (Outlook/Teams) experience vs 3rd party application integration including Microsoft chat integration.       


 


 


Actions:



 


Resources:


From demos:



From Q&A



 


General Resources: 



 


Stay Connected:



 

Microsoft Endpoint Manager and Delivery Optimization

Microsoft Endpoint Manager and Delivery Optimization

This article is contributed. See the original author and article here.

Microsoft Endpoint Manager and Delivery Optimization



 


Hello everyone! I´m Stefan Röll, Customer Engineer at Microsoft Germany for Microsoft Endpoint Manager. In my last Delivery Optimization Blog, I wrote about how you can test Delivery Optimization in your Organization. After many discussions with customers, I have noticed that it is not always clear when Delivery Optimization can be used, so I made a couple of tests and here are my results.


 


TL;DR


Delivery Optimization can be used to optimize setup and update downloads for various products. However, it is important to understand in which scenarios it can be used to optimize traffic. The table below will give you a quick overview. For more details continue reading.


 


What is Delivery Optimization?


Delivery Optimization (DO) is a service that consists of multiple components. It has the goal to minimize the traffic that is downloaded from internet to your internal network by caching and sharing content internally.


The Delivery Optimization service is hosted in the cloud and the Delivery Optimization downloader on Windows 10 uses information provided by the cloud service to find peers in the local network.


However, the Delivery Optimization Downloader can also be used to download Content without seeking for Peers or without contacting the cloud service.


For example, in VPN scenarios the Delivery Optimization Downloader is still downloading the content, but not sharing content with peers.  When a different download method like Background Intelligent Transfer Service (BITS) is used, the download will never be optimized via Delivery Optimization.


 


What is Microsoft Connected Cache?


Microsoft Connected Cache (MCC) is an extra component that you can install in your company network. It provides an additional caching source inside your network for your clients to have even better caching results. Right now, the MCC is only available as an additional component on Configuration Manager (ConfigMgr) Distribution Points but might be available as standalone container in the future.


The challenge with MCC is that clients need to find this caching server. In a ConfigMgr environment it can be automatically configured via the Boundary Group configuration.  If you don’t use ConfigMgr, you can manually configure the MCC to be used by clients via Group Policy or Configuration profiles in Intune.


However, these entries are always static and not ideal for roaming clients. But the solution to this is already here. For Windows 10 2004 and above you can configure an DHCP Option to provide the clients the nearest MCC.


 


Which content can I download over Delivery Optimization?


Now the things can get a little difficult. It is always important to remember which downloader is used for content – DO or BITS.


ConfigMgr uses BITS by default but can be forced to use DO in some scenarios.


Windows Update in Windows 10 uses DO by default – every content that is downloaded via the Windows Update agent can be optimized via DO.  Office Click2Run (2019 + 365) can use DO since version 1912 for the installation. Office Updates work with DO for quite a while now.  Unfortunately, not everything works with MCC yet, so let’s get into the details:


 


 



















































Product



Workload managed via Intune



Workload managed via ConfigMgr



Microsoft Updates



DO + MCC



DO 1



Windows 10 Upgrades



DO + MCC



DO (via Servicing) 1



Office 365 Installs



DO + MCC



DO + MCC (via CDN) 2



Office 365 Updates



DO + MCC



DO not supported  3



Store and Store for Business Installs/Updates



DO + MCC



DO + MCC 4



Intune Win32 Apps



DO + MCC



N/A



Microsoft Edge Updates



DO not supported 5



DO 6



Defender Definition Updates



DO + MCC



DO (via MECM) / DO + MCC (via WU)  7



All tests were performed on Win 10 20H2 19042.928 and ConfigMgr 2103 5.00.9049.1010 in May 2021


 


 


1) Microsoft Updates and Upgrades via ConfigMgr


For Microsoft Updates and Upgrades, DO can only be used if you enable ‘download delta content’ in the Client settings and the content is not available on internal DPs. See my other Blog post for details.


 


2) O365 Installs via ConfigMgr


Office 365 Installs can only be optimized via DO in ConfigMgr if the Content is downloaded from the Content Delivery Network (CDN) and not from internal Distribution Points. See my other Blog post for details.


 


3) O365 Updates via ConfigMgr


When you deploy O365 Updates over ConfigMgr, DO will never be used. The current implementation of the Delta Downloader does not support O365 content.


 


4) Store and Store for Business Installs and Updates via ConfigMgr


Microsoft Store Apps deployed via ConfigMgr use DO + MMC when deployed in Online mode. Offline Apps (where the content is stored on your local DPs) cannot use DO + MCC.


 


Picture1.png


 


5) Microsoft Edge Updates via Intune


If you don´t mange Microsoft Edge Updates via ConfigMgr, the internal Updater from Microsoft Edge is used. Unfortunately, this Updater doesn’t support DO.


 


6) Microsoft Edge Updates via ConfigMgr


If you use ConfigMgr to Update Microsoft Edge you need to disable automatic Updates inside Edge. Updates for Microsoft Edge are deployed as regular Microsoft Updates via ConfigMgr and therefore DO can only be used if you enable ‘download delta content’ in the Client settings and when the content is not available on internal DPs.


 


7) Defender Definition Updates via ConfigMgr


If you manage Defender Definition Updates via ConfigMgr, it depends on which Update Source you configured if DO can be use. If you configure ‘Microsoft Update’ as source, DO+MCC can be used. But keep in mind that you might need to reduce the File Size cached by DO via Group Policy or Configuration profiles in Intune. If you configure ‘Configuration Manager’ as source, DO can only be used if you enable ‘download delta content’ in the Client settings and the content is not available on internal DPs.


 


 


Summary


When you manage your Clients over Intune, DO and MCC can be used for almost all scenarios without complex configuration. If you manage your Clients with ConfigMgr, you need to know in which situations DO + MCC can be used.


I hope my blog helps you to understand the behavior a bit better.


 


Happy Caching!


 


Stefan Röll


Customer Engineer – Microsoft Germany


 


 


Resources: 


Stay current while minimizing network traffic: The power of Delivery Optimization


https://myignite.techcommunity.microsoft.com/sessions/81680?source=sessions


Delivery Optimization reference


https://docs.microsoft.com/en-us/windows/deployment/update/waas-delivery-optimization-reference


Delivery Optimization and Office 365 ProPlus


https://docs.microsoft.com/en-us/deployoffice/delivery-optimization


Optimize Windows 10 update delivery with Configuration Manager


https://docs.microsoft.com/en-us/configmgr/sum/deploy-use/optimize-windows-10-update-delivery


Microsoft Connected Cache in Configuration Manager


https://docs.microsoft.com/en-us/configmgr/core/plan-design/hierarchy/microsoft-connected-cache


Modern Content Distribution: Microsoft Endpoint Manager and Connected Cache


https://techcommunity.microsoft.com/t5/core-infrastructure-and-security/modern-content-distribution-microsoft-endpoint-manager-and/ba-p/1148669


 


Disclaimer:


This posting is provided “AS IS” with no warranties and confers no rights


 


Azure Neural Text-to-Speech extended to support lip sync with viseme

Azure Neural Text-to-Speech extended to support lip sync with viseme

This article is contributed. See the original author and article here.

Neural Text-to-Speech (Neural TTS), part of Speech in Azure Cognitive Services, enables you to convert text to lifelike speech for more natural user interactions. One emerging solution area is to create an immersive virtual experience with an avatar that automatically animates its mouth movements to synchronize with the synthetic speech. Today, we introduce the new feature that allows developers to synchronize the mouth and face poses with TTS – the viseme events.


 


What is viseme


 


A viseme is the visual description of a phoneme in a spoken language. It defines the position of the face and the mouth when speaking a word. With the lip sync feature, developers can get the viseme sequence and its duration from generated speech for facial expression synchronization. Viseme can be used to control the movement of 2D and 3D avatar models, perfectly matching mouth movements to synthetic speech.


 


Traditional avatar mouth movement requires manual frame-by-frame production, which requires long production cycles and high human labor costs.


 


Viseme can generate the corresponding facial parameters according to the input text. It greatly expands the number of scenarios by making the avatar easier to use and control. Below are some example scenarios that can be augmented with the lip sync feature. 



  • Customer service agent: Create an animated virtual voice assistant for intelligent kiosks, building the multi-mode integrative services for your customers;

  • Newscast: Build immersive news broadcasts and make content consumption much easier with natural face and mouth movements;

  • Entertainment: Build more interactive gaming avatars and cartoon characters that can speak with dynamic content;

  • Education: Generate more intuitive language teaching videos that help language learners to understand the mouth behavior of each word and phoneme;

  • Accessibility: Help the hearing-impaired to pick up sounds visually and “lip-read” any speech content.


 


How viseme works with Azure neural TTS


 


The viseme turns the input text or SSML (Speech Synthesis Markup Language) into Viseme ID and Audio offset which are used to represent the key poses in observed speech, such as the position of the lips, jaw and tongue when producing a particular phoneme. With the help of a 2D or 3D rendering engine, you can use the viseme output to control the animation of your avatar.


 


The overall workflow of viseme is depicted in the flowchart below.


Yueying_Liu_0-1621061769441.jpeg


The underlying technology for the Speech viseme feature consists of three components: Text Analyzer, TTS Acoustic Predictor, and TTS Viseme Generator.


 


To generate the viseme output for a given text, the text or SSML is first input into the Text Analyzer, which analyzes the text and provides output in the form of phoneme sequence. A phoneme is a basic unit of sound that distinguishes one word from another in a particular language. A sequence of phonemes defines the pronunciations of the words provided in the text.


 


Next, the phoneme sequence goes into the TTS Acoustic Predictor and the start time of each phoneme is predicted.


 


Then, the TTS Viseme generator maps the phoneme sequence to the viseme sequence and marks the start time of each viseme in the output audio. Each viseme is represented by a serial number, and the start time of each viseme is represented by an audio offset. Often several phonemes correspond to a single viseme, as several phonemes look the same on the face when pronounced, such as ‘s’, ‘z’.


 


Here is an example of the viseme output.


(Viseme), Viseme ID: 1, Audio offset: 200ms.


(Viseme), Viseme ID: 5, , Audio offset: 850ms.          


……


(Viseme), Viseme ID: 13, Audio offset: 2350ms.


 


This feature is built into the Speech SDK. With just a few lines of code, you can easily enable facial and mouth animation using the viseme events together with your TTS output.


 


How to use the viseme


 


To enable viseme, you need to subscribe to the VisemeReceived event in Speech SDK (The TTS REST API doesn’t support viseme). The following snippet illustrates how to subscribe to the viseme event in C#. Viseme only supports English (United States) neural voices at the moment but will be extended to support more languages later.


 

using (var synthesizer = new SpeechSynthesizer(speechConfig, audioConfig))
{
    // Subscribes to viseme received event
    synthesizer.VisemeReceived += (s, e) =>
    {
        Console.WriteLine($"Viseme event received. Audio offset: " +
            $"{e.AudioOffset / 10000}ms, viseme id: {e.VisemeId}.");
    };

    var result = await synthesizer.SpeakSsmlAsync(ssml));
}

 


After obtaining the viseme output, you can use these outputs to drive character animation.  You can build your own characters and automatically animate the characters.


 


For 2D characters, you can design a character that suits your scenario and use Scalable Vector Graphics (SVG) for each viseme ID to get a time-based face position.  With temporal tags provided by viseme event, these well-designed SVGs will be processed with smoothing modifications, and provide robust animation to the users. For example, below illustration shows a red lip character designed for language learning. Try the red lip animation experience in Bing Translator, and learn more about how visemes are used to demonstrate the correct pronunciations for words.


 


Yueying_Liu_1-1621061769468.png


 


For 3D characters, think of the characters as string puppets. The puppet master pulls the strings from one state to another and the laws of physics will do the rest and drive the puppet to move fluidly. The Viseme output acts as a puppet master to provide an action timeline. The animation engine defines the physical laws of action. By interpolating frames with easing algorithms, the engine can further generate high-quality animations.


 


Yueying_Liu_2-1621061769510.png


 (Note: the character image in this example is from Mixamo.)


 


Learn more about how to use the viseme feature to enable text-to-speech animation with the tutorial video below.


 


Get started 


 


With the viseme feature, Azure neural TTS expands its support for more scenarios and enables developers to create an immersive virtual experience with automatic lip sync to synthetic speech. 


 


Let us know how you are using or plan to use Neural TTS voices in this form. If you prefer, you can also contact us at mstts [at] microsoft.com. We look forward to hearing about your experience and developing more compelling services together with you for the developers around the world.


 


See our documentation for Viseme


Add voice to your app in 15 minutes


Build a voice-enabled bot


Deploy Azure TTS voices on prem with-Speech Containers


Build your custom voice


Improve synthesis with the Audio Content Creation tool


Visit our Speech page to explore more speech scenarios


 

How to use Logic App to get/send outlook item (i.e. *.msg)

How to use Logic App to get/send outlook item (i.e. *.msg)

This article is contributed. See the original author and article here.

If you tried to use logic app to pick up all attachments with “When an email is received” Office 365 trigger, we might notice that some of the attachments are not able to be retrieved from this trigger. By checking the raw output from the trigger, you might find even “hasAttachments” is “true”, the “attachments” is an empty array.


Yanbo_Deng_0-1621481133825.png


 


 


This is caused by Office 365 connector does not support “item attachment”. From our official document, we should use HTTP with Azure AD connector to work around such limitation.
Reference:
Office 365 connector limitation: https://docs.microsoft.com/en-us/connectors/office365/#known-issues-and-limitations


 


To workaround it, you could follow the steps below:



  1. First of all, using a “When a new email arrives V3” trigger to monitor if an email arrived, and get the message id of emails received. And then use HTTP to call Graph API and list attachments from the message id retrieved from the trigger.
    Reference:
    Graph API list attachments: https://docs.microsoft.com/en-us/graph/api/message-list-attachments?view=graph-rest-1.0&tabs=http
    Yanbo_Deng_1-1621481133837.png

     





  2. Then for each attachments listed from the email, we get the attachment raw contents from the attachment id. Note that we added “$value” in the request to get attachment raw contents.
    Reference:
    https://docs.microsoft.com/en-us/graph/api/attachment-get?view=graph-rest-1.0&tabs=http#example-6-get-the-raw-contents-of-a-file-attachment-on-a-message
    Yanbo_Deng_2-1621481133845.png

     





  3. The last step is to send this attachment in an email. We need to add an extension “.eml” to the attachment name, so our outlook could recognize it as an valid attachment.
    Yanbo_Deng_3-1621481133852.png

     




Sample:


The original email received:


Yanbo_Deng_6-1621481188589.png


 


The new email sent by logic app:


Yanbo_Deng_5-1621481133861.png