Azure Data Engineer (Az-200, 201)

Aspired to build your career as an Azure Data Engineer? Follow this positive vocation way and become an Azure Data engineer.

Azure Data Engineer Associate is the greatest superpower in right now. Gathering the proper and important Data can help organizations drive better choices. Moreover, the legitimate utilization of data could uphold upgrades in client care. This has been one of the reasons for an abrupt ascent in requests for information researchers. Nonetheless, nobody calls attention to the meaning of an Data engineer in the midst of the commotion for information researchers.

DP 200 and DP 201  You should take note of that an Azure Data engineer is additionally quite possibly the most pined for work parts in the Microsoft Azure scene. The accompanying conversation would target showing a guide for the Azure Data engineer profession way, alongside the Azure confirmation needed for the Data engineer job. The conversation would zero in on pivotal perspectives, for example, the way for turning into an Data Engineer. The conversation would likewise put forth an attempt to introduce insights concerning motivations to pick Data engineer occupations.

Azure Data Engineer Course Content:

Azure Data Engineer + Data bricks Developer 

  • Python Basics 
  • Data Bricks
  • Spark 
  • PySpark 
  • Delta Lake
  • Azure Data Factory
  • Azure ADF & Data bricks Projects Projects

Python Basics

1. Python Data Types
  a. Numbers, Strings….
2. Data Structures
a. Lists, and Tuples ,Dictionaries and Sets
3. Conditionals and Loop Control Statements
a. If, for, while, pass, break, continue
4. Regular Expressions
5. Functions
6. Files and Input/Output
7. Errors and ExceptionsYouTube Channel: Pyspark telugu
Azure Databricks Concepts.
1) Azure Databricks Introduction
A. Databricks Architecture
B. Databricks Components overview
C. Benefits for data engineers and data scientists
2) Azure Databricks concepts
A. Workspace – Creation and managing workspace.
B. Notebook – creating notebooks, calling and managing different notebooks.
C. Library – installing libraries, managing libraries
D. Experiment – ML and dependency libraries usage.
3) Data Management
A. Databricks File System. – DBFS commands copy and manage files using DBFS.
B. Database – Creating database, tables and managing databases and tables.
C. Table – Creating Tables, dropping tables, loading data ..
D. Metastore – managing metadata and delta tables creation, managing delta tables.
4) Computation Management
A. Cluster — Creating Clusters , managing clusters
B. Pool – creating pools and using pools for Auto scaling.
C. Databricks RunTime – understanding and using Databricks runtimes based on requirement.
D. Jobs – creating jobs from notebooks and assigning types of clusters for jobs.
E. Workload – monitoring jobs and managing loads.
F. Execution Context – understanding context.
5) Security
A. User – Creating users
B. Group – creating groups.
C. Managing Access – managing access to users and groupsYouTube Channel: Pyspark telugu
6) Databricks Advanced topics.
A. Databricks Workflows
B. Calling one notebook into another notebook.
C. Creating global variables (widgets) and using into Azure ADF pipeline.
D. How to implement parallelism in notebooks execution.
E. Mounting azure blob storage and data lake storage accounts.
F. Integrating source code (notebooks) with GitHub
G. Calling DataBricksnotebooks into Azure Data factory.
H. Databricks Clusters logs monitoring flow.
SPARK Concepts
1) Introduction to Spark -Getting started
A. What is Spark and what is its purpose?
B. Components of the Spark unified stack
C. Resilient Distributed Dataset (RDD)
D. Downloading and installing Spark standalone
E. Scala and Python overview
F. Launching and using Spark’s Scala and Python shell ©
2) Resilient Distributed Dataset and DataFrames
A. Understand how to create parallelized collections and external datasets
B. Work with Resilient Distributed Dataset (RDD) operations
C. Utilize shared variables and key-value pairs
3) Spark application programming
A. Understand the purpose and usage of the Spark Context
B. Initialize Spark with the various programming languages
C. Describe and run some Spark examples
D. Pass functions to Spark
E. Create and run a Spark standalone application
F. Submit applications to the cluster
4) Introduction to Spark libraries
A. Understand and use the various Spark librariesYouTube Channel: Pyspark telugu
5) Spark configuration, monitoring and tuning
A. Understand components of the Spark cluster
B. Configure Spark to modify the Spark properties, environmental variables, or logging
C. Monitor Spark using the web UIs, metrics, and external instrumentation ,Understand
performance tuning considerations
PySpark Content
 Introduction To Pyspark
1) What is SparkSession
2) How to create spark session
3) What is SparkContext
4) How to create SparkContext
5) What is SQLContext
 How to Use Jupyter Notebooks & Databricks notebooks for Python Development.
 Install and configure PySpark in Local System for development.
 Introduction to Big Data and Apache Spark
 Apache Spark Framework & Execution Process.
 Introduction To RDDs
1) Different Ways to Create RDD’s in Pyspark.
2) RDD Transformations
3) RDD Actions
4) RDD Cache & Persist
 Introduction to DataFrame.
1) Different Ways to Create Data Frame’sin Pyspark.
2) Dataframe Transformations
3) Dataframe Actions
4) Dataframe Cache & Persist
 Different types of Big Data File systems.
1) Difference between Row store format and column store format.
2) Avro File
3) Parquet file
4) ORC File
 Reading and Writing Different Types of Files using Dataframe.
1) Csv files
2) Json filesYouTube Channel: Pyspark telugu
3) Xml files
4) Excel files
5) Complex Json files
6) Avro files
7) Parquet files
8) Orc files

Need for Spark SQL
 What is Spark SQL
1) SQL Table Creation
2) SQL Join Types
3) SQL Nested Queries
4) SQL DML Operations
5) SQL Merge Scripts
6) SQL SCD Type 2 implementation
User-Defined Functions
 Performance Tuning
 Spark-Hive Integration
 Pyspark Project with execution.
1) End to End Pyspark Projects implementation
2) Executing Pyspark Project in Databricks
3) Executing PySpark project in Azure ADF.

1) Delta Lake usage in Databricks.
A. Delta Lake Architecture
B. Delta Lake Storage Understanding
C. Delta lake table creation
D. Delta Lake DML Operations usage.
E. Delta Lake Snapshots
Azure Data Engineer YouTube Channel: Pyspark telugu
1) Overview of the Microsoft Azure Platform
A. Introduction to Azure
B. Basics of Cloud computing
C. Azure Infrastructure
D. Walkthrough of Azure Portal
E. Overview of Azure Services
2) Azure Data Architecture
A. Traditional RDBMS workloads.
B. Data Warehousing Approach
C. Big data architectures.
D. Transferring data to and from Azure
3) Azure Storage options
A. Blob Storage
B. ADLS Gen1 & Gen2
D. Hadoop
F. Disk
4) Blob Storage
A. Azure Blob Resources
B. Types of Blobs in Azure
C. Azure storage account data objects
D. Azure storage account types and Options
E. Replications in distribution
F. Secure access to an application’s data
G. Azure Import/Export service
H. Storage Explorer
I. Practical section on Blob Storage
5) Azure Data Factory
A. Azure Data Factory Architecture
B. Creating ADF Resource and Use in azure cloud
C. Pipeline Creation and Usage Options
D. Copy Data Tool in ADF Portal, Use
E. Linked Service Creation in ADF
F. Dataset Creation, Connection Reuse
G. Staging Dataset with Azure Storage
H. ADF Pipeline Deployments
I. Pipeline Orchestration using Triggers
J. ADF Transformations and other tools integration.
K. Processing different type’s files using ADF.
L. Integration Runtime
M. Monitoring ADF Jobs
N. Manage IR’s and Linked Services.
6) Azure SQL Database Service
A. Introduction to Azure SQL Database
B. Relational Data Services in the Cloud
C. Azure SQL Database Service TiersYouTube Channel: Pyspark telugu
D. Database Throughput Units (DTU)
E. Scalable performance and pools
F. Creating and Managing SQL Databases
G. Azure SQL Database Tools
H. Migrating data to Azure SQL Database
7) Azure Data Lake Gen1 & Gen2
A. Explore the Azure Data Lake enterprise-class security features.
B. Understand storage account keys.
C. Understand shared access signatures.
D. Understand transport-level encryption with HTTPS.
E. Understand Advanced Threat Protection.
F. Control network access.
G. Differences between Gen1 & Gen2
8) Azure HD-Insight cluster
A. Creating HD-Insight Cluster
B. Understanding HD-Insight Architecture
C. Using Spark in HD-insight
D. Using Hadoop in HD-insight
E. Understanding Amabari view
F. Pricing structure and calculations
G. Monitoring and manage.