Learn HDFS commands, Hadoop, Spark SQL, SQL Queries, ETL & Data Analysis| Spark Hadoop Cluster VM | Fully Solved Qs
What you will learn
Students will get hands-on experience working in a Spark Hadoop environment thatβs free and downloadable as part of this course.
Students will have opportunities solve Data Engineering and Data Analysis Problems using Spark on a Hadoop cluster in the sandbox environment that comes as part
Issuing HDFS commands.
Converting a set of data values in a given format stored in HDFS into new data values or a new data format and writing them into HDFS.
Loading data from HDFS for use in Spark applications & writing the results back into HDFS using Spark.
Reading and writing files in a variety of file formats.
Performing standard extract, transform, load (ETL) processes on data using the Spark API.
Using metastore tables as an input source or an output sink for Spark applications.
Applying the understanding of the fundamentals of querying datasets in Spark.
Filtering data using Spark.
Writing queries that calculate aggregate statistics.
Joining disparate datasets using Spark.
Producing ranked or sorted data.
Description
Apache Spark is currently one of the most popular systems for processing big data.
Apache Hadoop continues to be used by many organizations that look to store data locally on premises. Hadoop allows these organisations to efficiently store big datasets ranging in size from gigabytes to petabytes.
As the number of vacancies for data science, big data analysis and data engineering roles continue to grow, so too will the demand for individuals that possess knowledge of Spark and Hadoop technologies to fill these vacancies.
This course has been designed specifically for data scientists, big data analysts and data engineers looking to leverage the power of Hadoop and Apache Spark to make sense of big data.
This course will help those individuals that are looking to interactively analyse big data or to begin writing production applications to prepare data for further analysis using Spark SQL in a Hadoop environment.
The course is also well suited for university students and recent graduates that are keen to gain exposure to Spark & Hadoop or anyone who simply wants to apply their SQL skills in a big data environment using Spark-SQL.
This course has been designed to be concise and to provide students with a necessary and sufficient amount of theory, enough for them to be able to use Hadoop & Spark without getting bogged down in too much theory about older low-level APIs such as RDDs.
On solving the questions contained in this course students will begin to develop those skills & the confidence needed to handle real world scenarios that come their way in a production environment.
(a) There are just under 30 problems in this course. These cover hdfs commands, basic data engineering tasks and data analysis.
(b) Fully worked out solutions to all the problems.
(c) Also included is the Verulam Blue virtual machine which is an environment that has a spark Hadoop cluster already installed so that you can practice working on the problems.
- The VM contains a Spark Hadoop environment which allows students to read and write data to & from the Hadoop file system as well as to store metastore tables on the Hive metastore.
- All the datasets students will need for the problems are already loaded onto HDFS, so there is no need for students to do any extra work.
- The VM also has Apache Zeppelin installed. This is a notebook specific to Spark and is similar to Pythonβs Jupyter notebook.
This course will allow students to get hands-on experience working in a Spark Hadoop environment as they practice:
- Converting a set of data values in a given format stored in HDFS into new data values or a new data format and writing them into HDFS.
- Loading data from HDFS for use in Spark applications & writing the results back into HDFS using Spark.
- Reading and writing files in a variety of file formats.
- Performing standard extract, transform, load (ETL) processes on data using the Spark API.
- Using metastore tables as an input source or an output sink for Spark applications.
- Applying the understanding of the fundamentals of querying datasets in Spark.
- Filtering data using Spark.
- Writing queries that calculate aggregate statistics.
- Joining disparate datasets using Spark.
- Producing ranked or sorted data.
Content