Stream Microsoft Defender for IoT alerts to a 3rd party SIEM

Stream Microsoft Defender for IoT alerts to a 3rd party SIEM

This article is contributed. See the original author and article here.

Overview


As more businesses convert OT systems to digital IT infrastructures, security operations center (SOC) teams and chief information security officers (CISOs) are increasingly responsible for handling threats from OT networks.


 


Defender for IoT’s built-in integration with Sentinel helps bridge the gap between IT & OT securitychallenge. Sentinel enables SOC teams to reduce the time taken to manage and resolve OT incidents efficiently by providing out-of-the-box capabilities to analyze OT security alerts, investigate multistage IT/OT attacks, utilize Azure Log Analytics for threat hunting, utilize threat intelligence, and automate incident response using SOAR playbooks.


 


Customer engagements have taught us that sometimes customers prefer to maintain their existing SIEM, alongside Microsoft Sentinel, or as a standalone SIEM.


In this blog, we’ll introduce a solution that sends Microsoft Defender for IoT alerts to an Event Hub that can be consumed by a 3rd party SIEMs. You can use this solution with Splunk, QRadar, or any other SIEM that supports Event Hub ingestion.


 


Preparation and use


In this blog, we’ll use Splunk as our example.


Screen Shot 2022-07-25 at 9.02.11.png


 


The following describe the necessary preparation steps:



  1. Connect your alerts from Defender for IoT to Microsoft Sentinel

  2. Register an application in Azure AD

  3. Create an Azure Event Hub Namespace

  4. Prepare Azure Sentinel to forward Incidents to Event Hub

  5. Configure Splunk to consume Azure Sentinel Incidents from Azure Event Hub



1. Connect your alerts from Defender for IoT to Microsoft Sentinel


The first step is to enable the Defender for IoT data connector so that all Defender for IoT alerts are streamed into Microsoft Sentinel (a free process).


 


In Microsoft Sentinel, under Configuration, select Data Connectors and then locate Microsoft Defender for IoT data connector. Open the connector page, select the subscription whose alerts you want to stream into Microsoft Sentinel, and then select Connect.


 


For more information, see Connect your data from Defender for IoT to Microsoft Sentinel


2. Register an application in Azure AD


You’ll need Azure AD to be defined as a service principal for Splunk Add-on for Microsoft Cloud Services.



  1. To register an app in Azure AD, open the Azure Portal and navigate to Azure Active Directory > App Registrations > New Registration. Fill the Name and click Register.

    Screen Shot 2022-07-25 at 9.16.35.png



  2. Click Certificates & secrets to create a secret for the Service Principle. Click New client secret and note its value.
    Screen Shot 2022-07-25 at 9.27.04.png


  3. To grant the required permissions to read data from the app, click API permissions > Add a permission and select Microsoft Graph > Application permissions > SecurityEvents.ReadWrite.All.
    Screen Shot 2022-07-25 at 9.28.43.png

    Ensure that the granted permission is approved by admin.



  4.  For the next step of setting up Splunk Add-on for Microsoft Cloud Services, note the following settings:


    • The Azure AD Display Name

    • The Azure AD Application ID

    • The Azure AD Application Secret

    • The Tenant ID





3. Create an Azure Event Hub Namespace



  1. In the Azure Portal, navigate to Event Hubs > New to create a new Azure Event Hub Namespace. Define a Name, select the Pricing Tier and Throughput Units and click Review + Create.
    Screen Shot 2022-07-25 at 9.29.48.png


  2. Once the Azure Event Hub Namespace is created click Go to resource and click + Event Hubs to create an Azure Event Hub within the Azure Event Hub Namespace.


  3. Define a Name for the Azure Event Hub, configure the Partition CountMessage Retention and click Review + Create.
    Screen Shot 2022-07-25 at 9.33.29.png


  4. Navigate to Access control (IAM) and Click + Add > Add role assignment to add the Azure AD Service Principle created before and delegate as Azure Event Hubs Data Receiver and click Save.
    Screen Shot 2022-07-25 at 9.30.15.png


  5. For the configuration of Splunk Add-on for Microsoft Cloud Services app, make a note of following settings:

    • The Azure Event Hub Namespace Host Name

    • The Azure Event Hub Name





4. Prepare Azure Sentinel to forward Incidents to Event Hub


To forward Microsoft Sentinel incidents or alerts to Azure Event Hub, you’ll need to define your Microsoft Sentinel workspace with a data export rule.



  1. In the Azure Portal, navigate to Log Analytics > select the workspace name related to Microsoft Sentinel > Data Export > New export rule.
    Screen Shot 2022-07-25 at 9.30.24.png


  2. Name the rule, configure the Source as SecurityIncident and the Destination as Event Type utilizing the Event Hub Namespace and Event Hub Name configured previously. Click on Create.
    Screen Shot 2022-07-25 at 9.30.43.png


5. Configure Splunk to consume Microsoft Sentinel Incidents from Azure Event Hub


For Microsoft Defender for IoT alerts to be ingested into Azure Event Hub, install the Splunk Add-on for Microsoft Cloud Services app.



  1. For the installation, open the Splunk portal and navigate to Apps > Find More Apps. For the dashboard find the Splunk Add-on for Microsoft Cloud Services app and Install.
    Screen Shot 2022-07-25 at 9.30.53.png


  2. To add the Azure AD Service Principal, open the Splunk app and navigate to Azure App Account > Add. Use the details you’d noted earlier:

    Define a Name for the Azure App Account


    Add the Client ID, Client Secret, Tenant ID


    Choose Azure Public Cloud as Account Class Type


    Click Update to save and close the configuration.
    Screen Shot 2022-07-25 at 9.31.10.png




  3. Now navigate to Inputs within the Splunk Add-on for Microsoft Cloud Services app and select Azure Event Hub in Create New Input selection. 

    Define a Name for the Azure Event Hub as Input, select the Azure App Account created before, define the Event Hub Namespace (FQDN), Event Hub Name, let the other settings as default and click Update to save and close the configuration.
    Screen Shot 2022-07-25 at 9.31.24.png




Once the ingestion is processed, you can query the data by using sourcetype=”mscs:azure:eventhub” in search field.
Screen Shot 2022-07-25 at 9.31.35.png

Disclaimer: The use of EventHub and Log Analytics export rule may incur an additional charge. For more information, see Event Hubs pricing and Log Data Export pricing 


 

Speech Recognition for Singlish

Speech Recognition for Singlish

This article is contributed. See the original author and article here.

Mithun Prasad, PhD, Senior Data Scientist at Microsoft and Manprit Singh, CSA at Microsoft


 


Speech is an essential form of communication that generates a lot of data. As more systems provide a modal interface with speech, it becomes critical to be able to analyze human to computer interactions. Interesting market trends point that voice is the future of UI. This claim is further bolstered now with people looking to embrace contact less surfaces with the recent pandemic.


 


Interactions between agents and customers in a contact center remains dark data that is often untapped. We believe the ability to transcribe speech in the local dialects/slang should be in the midst of a call center advanced analytics road map such as the one proposed in this McKinsey recommendation. To enable this, we want to bring the best from the current speech transcription landscape, and present it in a coherent platform which businesses can leverage to get a head start on local speech to text adaptation use cases. 


 


There is tremendous interest in Singapore to understand Singlish.


 


Singlish is a local form of English in Singapore that blends words borrowed from the cultural mix of communities.


miprasad_0-1658626696072.png


An example of what Singlish looks like


 


A speech recognition system that could interpret and process the unique vocabulary used by Singaporeans (including Singlish and dialects) in daily conversations is very valuable. This automatic speech transcribing system could be deployed at various government agencies and companies to assist frontline officers in acquiring relevant and actionable information while they focus on interacting with customers or service users to address their queries and concerns.


 


Efforts are on to understand calls made to transcribe emergency calls at Singapore’s Civil Defence Force (SCDF) while AI Singapore has launched Speech Lab to channel efforts in this direction. Now, with the release of the IMDA National Speech Corpus, local AI developers now have the ability to customize AI solutions with locally accented speech data. 


 


IMDA National Speech Corpus


The Infocomm Media Development Authority of Singapore has released a large dataset, which is:


 


• A 3 part speech corpus each with 1000 hours of recordings of phonetically-balanced scripts from ~1000 local English speakers.


• Audio recordings with words describing people, daily life, food, location, brands, commonly found in Singapore. These are recorded in quiet rooms using a combination of microphones and mobile phones to add acoustic variety.


• Text files which have transcripts. Of note are certain terms in Singlish such as ‘ar’, ‘lor’, etc.


 


This is a bounty for the open AI community in accelerating efforts towards speech adaptation. With such efforts, the trajectory for the local AI community and businesses are poised for major breakthroughs in Singlish in the coming years.


 


We have leveraged the IMDA national speech corpus as a starting ground to see how adding customized audio snippets from locally accented speakers drives up accuracy of transcription. An overview of the uptick is in the below chart. Without any customization, the holdout set performed with an accuracy of 73%. As more data snippets were added, we can validate that with the right datasets, we can drive accuracy up using human annotated speech snippets.


 


miprasad_2-1658626909593.png


 


On the left is the uplift in terms of accuracy. The right correspondingly shows the Word Error Rate dropping on addition of more audio snippets


 


Keeping human in the loop


 


The speech recognition models learn from humans, based on “human-in-the-loop learning”. Human-in-the-Loop Machine Learning is when humans and Machine Learning processes interact to solve one or more of the following:



  • Making Machine Learning more accurate

  • Getting Machine Learning to the desired accuracy faster

  • Making humans more accurate

  • Making humans more efficient


 


An illustration of what a human in the loop looks like is as follows. 


miprasad_3-1658626952260.png


 


In a nutshell, human in the loop learning is giving AI the right calibration at appropriate junctures. An AI model starts learning for a task, which eventually can plateau over time. Timely interventions by a human in this loop can give the model the right nudge. “Transfer learning will be the next driver of ML success.”- Andrew Ng, in his Neural Information Processing Systems (NIPS) 2016 tutorial 


 


Not everybody has access to volumes of call center logs, and conversation recordings collected from a majority of local speakers which are key sources of data to train localized speech transcription AI. In the absence of significant amounts of local accented data with ground truth annotations, and our belief behind transfer learning to be a powerful driver in accelerating AI development, we leverage existing models and maximize their ability to understand towards local accents. 


 


miprasad_4-1658627000430.png


 


 


The framework allows extensive room for human in the loop learning and can connect with AI models from both cloud providers and open source projects. A detailed treatment of the components in the framework include:



  1. The speech to text model can be any kind of Automatic Speech Recognition (ASR) engine or Custom Speech API, which can run on cloud or on premise. The platform is designed to be agnostic to the ASR technology being used. 

  2. Search for ground truth snippets. In a lot of cases when the result is available, a quick search of the training records can point to the number of records trained, etc. 

  3. Breakdown on Word Error Rates (WER): The industry standard to measure Automatic Speech Recognition (ASR) systems is based on the Word Error Rate, defined as the below


miprasad_5-1658627047292.png


 


where S refers to the number of words substituted, D refers to the number of words deleted, and I refer to the number of words inserted by the ASR engine.


 


A simple example illustrating this is as below, where there is 1 deletion, 1 insertion, and 1 substitution in a total of 5 words in the human labelled transcript.


 


miprasad_6-1658627713891.png


 


Word Error Rate comparison between ground truth and transcript (Source: https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/how-to-custom-speech-evaluate-data)


 


So, the WER of this result will be 3/5, which is 0.6. Most ASR engines will return the overall WER numbers, and some might return the split between the insertions, deletions and substitutions. 


 


However, in our work (platform), we can provide a detailed split between the insertions, substitutions and deletions. 



  1. The platform built has ready interfaces that allow human annotators to plug audio files with relevant labeled transcriptions, to augment data

  2. It ships with dashboards which show detailed substitutions, such as how often was the term ‘kaypoh’ transcribed as ‘people’. 


The crux of the platform is the ability to control the existing transcription accuracy, by getting a detailed overview of how often the engine is having trouble transcribing certain vocabulary, and allowing human to give the right nudges to the model. 


 


References and useful links



  1. https://yourstory.com/2019/03/why-voice-is-the-future-of-user-interfaces-1z2ue7nq80?utm_pageloadtype=scroll

  2. https://www.mckinsey.com/business-functions/operations/our-insights/how-advanced-analytics-can-help-contact-centers-put-the-customer-first

  3. https://www.straitstimes.com/singapore/automated-system-transcribing-995-calls-may-also-recognise-singlish-shanmugam

  4. https://www.aisingapore.org/2018/07/ai-singapore-harnesses-advanced-speech-technology-to-help-organisations-improve-frontline-operations/

  5. https://livebook.manning.com/book/human-in-the-loop-machine-learning/chapter-1/v-6/17

  6. https://www.youtube.com/watch?v=F1ka6a13S9I

  7. https://ruder.io/transfer-learning/

  8. https://www.imda.gov.sg/programme-listing/digital-services-lab/national-speech-corpus


 


*** This work was performed in collaboration with Avanade Data & AI and Microsoft.


 

MTC Weekly Roundup – July 22

MTC Weekly Roundup – July 22

This article is contributed. See the original author and article here.

Hey there, MTC’ers! It’s been a busy week, so let’s jump right on in and look at what’s been happening in the Community this past week.


 


MTC Moments of the Week


 


This week, Community Events made a triumphant return with a double hitter!


 


Earlier this month, @Alex Simons published a blog post announcing the general availability of Microsoft Entra Permissions Management, and this past Tuesday, July 19, we had our first Entra AMA featuring @Nick Wryter, @Laura Viarengo, and @Mrudula Gaidhani.


 


Then, on Thursday, we had Tech Community Live: Endpoint Manager edition, which featured four AMA live streams all about the latest Endpoint Manager capabilities, including Windows Autopilot, Endpoint Analytics, and more! Thank you to everyone who attended :)


 


On the blogs this week, @Rafal Sosnowski published a post announcing the sunset of Windows Information Protection (WIP) and sharing resources on its successor, Microsoft Purview Data Loss Prevention (DLP), which you can try for free by enabling the free trial from the Microsoft Purview compliance portal.


 


Cecilia_Bergstedt_0-1658531799277.jpeg


 


I also want to shout out @Sergei Baklan for helping @Jammin2082 with their Morse code translator in Excel. What a cool way to use Excel!


 


 


Unanswered Questions – Can you help them out?


 


Every week, users come to the MTC seeking guidance or technical support for their Microsoft solutions, and we want to help highlight a few of these each week in the hopes of getting these questions answered by our amazing community!


 


This week, @Florian Hein shared a scenario they’ve run into involving links to Sharepoint pages not opening from within Teams. Have you experienced this before?


 


Cecilia_Bergstedt_1-1658531799281.png


 


 


Meanwhile, new contributor @eliekarkafy is looking for guidance in building documentation for an Azure Governance Framework. If you have recommendations or a template to share, hop in and help a fellow MTC’er!


 


Next Week – Mark your calendars!


Lesson Learned #230: Microsoft Reactor -Azure SQL Developer and DBA Best Practices (Spanish Version)

This article is contributed. See the original author and article here.

We had the great opportunity to deliver a session within program Microsoft Reactor with our colleague Pablo Javier Fernandez – Cloud Solution Architect – Data & AI LATAM SQL Advanced Cloud Expert about Azure SQL Developer and DBA Best Practices.


 


In this video you see how is an interaction between a DBA and developer to find:


 



  • How to identify the application is taking the resources.

  • How to review the database metrics. 

  • How to implement a maintenance plan using runbooks


 


Microsoft Reactor connects you with the developers and startups that share your goals. You could Learn new skills, meet new peers, and find career mentorship. Virtual events are running around the clock so join us anytime, anywhere!


 


You could find additional information below:


 


Lesson Learned #221:Hands-On-Labs: Activity Monitor in my Azure SQL Managed Instance – Microsoft Tech Community


Lesson Learned #220:Hands-On-Labs: Activity Monitor in my Elastic Database Pool – Microsoft Tech Community


Lesson Learned #219:Hands-On-Labs: What do I need to do in case of high CPU wait time – Microsoft Tech Community


Lesson Learned #218:Hands-On-Labs: What do I need to do in case of high LOG_RATE_GOVERNOR wait time – Microsoft Tech Community


Lesson Learned #207: Hands-On-Labs: 40613-Database ‘xyz’ on server ‘xyz2′ is not currently available – Microsoft Tech Community


Global Azure 2022 – No encuentro donde esta el problema de la query (Spanish Version Delivered) – Microsoft Tech Community


Lesson Learned #196: Latency and execution time in Azure SQL Database and Azure SQL Managed Instance – Microsoft Tech Community


Blog – Automating DB maintenance for all SQL Databases in a single server using Azure Data Factory pipeline (microsoft.com)


(and many others…)


 


Watch this video (Spanish version)


 


 


 


Enjoy!

Cisco Releases Security Updates for Multiple Products

This article is contributed. See the original author and article here.

Cisco has released security updates to address vulnerabilities in multiple products. Some of these vulnerabilities could allow a remote attacker to execute take control of an affected system. For updates addressing lower severity vulnerabilities, see the Cisco Security Advisories page

CISA encourages users and administrators to review the Cisco advisories and apply the necessary updates.

Atlassian Releases Security Advisory for Questions for Confluence App, CVE-2022-26138

This article is contributed. See the original author and article here.

Atlassian has released a security advisory to address a vulnerability (CVE-2022-26138) affecting Questions for Confluence App. An attacker could exploit this vulnerability to obtain sensitive information. Atlassian reports that the vulnerability is likely to be exploited in the wild.

CISA encourages users and administrators to review Atlassian’s security advisory, Questions For Confluence Security Advisory 2022-07-20, and apply the necessary updates immediately.