In this post I’d like to review some information about using ORC, Parquet and Avro files in Azure Data Lake, in particular when we’re extracting data with Azure Data Factory and loading it to files in Data Lake.
I’d like to tell you a story about a customer who chose to migrate from Oracle to Azure Data Warehouse and tell you their reasons for doing so, as well as the benefits they’re seeing after making the move.
Today’s post is in response to a question I was recently asked. It’s about using Azure Data Lake Store with Azure Data Factory, in particular about the Copy Activity within Data Factory to read data from Azure Data Lake.
In today’s post I’ll look at some considerations for choosing to use Azure Blob Storage or Azure Data Lake Store when processing data to be loaded into a data warehouse. My basis here is a reference architecture that Microsoft published, see diagram below.
We talk to a lot of customers about their data strategies, specifically their data cloud strategies. One great tool we have is Azure Data Lake. I’d like to introduce that tool and tell you about some benefits you will gain.