This article is contributed. See the original author and article here.

Azure Event Hubs enables you to stream millions of events per second from any source using Kafka, AMQP or HTTPS protocols. Using Event Hubs capture feature, you can load real-time streaming data to data lakes, warehouses, and other storage services, so that they can be processed or analyzed by analytics services.


Today we are excited to announce the preview of Apache Parquet capturing support in Azure Event Hubs.


Why Apache Parquet for big data analytics?

Apache Parquet is column-oriented storage format that is designed for efficient data storage and retrieval. It’s open source and is not tied to any processing framework, data model or programming language. Parquet is ideal for storing any kind of big data and is built to support efficient compression and encoding schemes.  


Capture streaming data in Parquet format using Event Hubs

Using Azure Event Hubs, no code editor for event processing, you can automatically capture streaming data in an Azure Data Lake Storage Gen2 account in Parquet format. The no code editor allows you to easily develop an Azure Stream Analytics job without writing a single line of code.



Once data is captured, any analytics service of your choice can process or analyze Parquet data.


Get Started Today



Brought to you by Dr. Ware, Microsoft Office 365 Silver Partner, Charleston SC.

%d bloggers like this: