Training Cloudera

Training goals dlearning

This four-day Analyzing with Data Warehouse course will teach you to apply traditional data analytics and business intelligence skills to big data. This course presents the tools data professionals need to access, manipulate, transform, and analyze complex data sets using SQL and familiar scripting languages.

What you'll learn

Through instructor-led discussion and interactive, hands-on exercises, participants will navigate the ecosystem, learning how to:

  • Use Apache Hive and Apache Impala to access data through queries
  • Identify distinctions between Hive and Impala, such as differences in syntax, data formats, and supported features
  • Write and execute queries that use functions, aggregate functions, and subqueries
  • Use joins and unions to combine datasets
  • Create, modify, and delete tables, views, and databases
  • Load data into tables and store query results
  • Select file formats and develop partitioning schemes for better performance
  • Use analytic and windowing functions to gain insight into their data
  • Store and query complex or nested data structures
  • Process and analyze semi-structured and unstructured data
  • Optimize and extend the capabilities of Hive and Impala
  • Determine whether Hive, Impala, an RDBMS, or a mix of these is the best choice for a given task
  • Utilize the benefits of CDP Public Cloud Data Warehouse

What to expect

This course is designed for data analysts, business intelligence specialists, developers, system architects, and database administrators. Some knowledge of SQL is assumed, as is basic Linux command-line familiarity.

Conspect Show list

  • Foundations for Big Data Analytics
    • Big Data Analytics Overview
    • Data Storage: HDFS
    • Distributed Data Processing: YARN, MapReduce, and Spark
    • Data Processing and Analysis: Hive and Impala
    • Database Integration: Sqoop
    • Other Data Tools
    • Exercise Scenario Explanation
  • Introduction to Apache Hive and Impala
    • What Is Hive?
    • What Is Impala?
    • Why Use Hive and Impala?
    • Schema and Data Storage
    • Comparing Hive and Impala to Traditional Databases
    • Use Cases
  • Querying with Apache Hive and Impala
    • Databases and Tables
    • Basic Hive and Impala Query Language Syntax
    • Data Types
    • Using Hue to Execute Queries
    • Using Beeline (Hive's Shell)
    • Using the Impala Shell
  • Common Operators and Built-In Functions
    • Operators
    • Scalar Functions
    • Aggregate Functions
  • Data Management
    • Data Storage
    • Creating Databases and Tables
    • Loading Data
    • Altering Databases and Tables
    • Simplifying Queries with Views
    • Storing Query Results
  • Data Storage and Performance
    • Partitioning Tables
    • Loading Data into Partitioned Tables
    • When to Use Partitioning
    • Choosing a File Format
    • Using Avro and Parquet File Formats
  • Working with Multiple Datasets
    • UNION and Joins
    • Handling NULL Values in Joins
    • Advanced Joins
  • Analytic Functions and Windowing
    • Using Analytic Functions
    • Other Analytic Functions
    • Sliding Windows
  • Complex Data
    • Complex Data with Hive
    • Complex Data with Impala
  • Analyzing Text
    • Using Regular Expressions with Hive and Impala
    • Processing Text Data with SerDes in Hive
    • Sentiment Analysis and n-grams in Hive
  • Apache Hive Optimization
    • Understanding Query Performance
    • Cost-Based Optimization and Statistics
    • Bucketing
    • ORC File Optimizations
  • Apache Impala Optimization
    • How Impala Executes Queries
    • Improving Impala Performance
  • Extending Hive and Impala
    • User-Defined Functions
    • Parameterized Queries
  • Choosing the Best Tool for the Job
    • Comparing Hive, Impala, and
    • Relational Databases
    • Which to Choose?
  • CDP Public Cloud Data Warehouse
    • Data Warehouse Overview
    • Auto-Scaling
    • Managing Virtual Warehouses
    • Querying Data Using CLI and Third-Party Integration
  • Appendix: Apache Kudu
    • What Is Kudu?
    • Kudu Tables
    • Using Impala with Kudu
Download conspect training as PDF

Additional information

Prerequisites

This course is designed for data analysts, business intelligence specialists, developers, system architects, and database administrators. Some knowledge of SQL is assumed, as is basic Linux command-line familiarity.

Difficulty level
Duration 4 days
Certificate

The participants will obtain certificates signed by Cloudera (course completion).

Upon completion of the course, attendees are encouraged to continue their study and register for the CDP Data Analyst exam https://www.cloudera.com/about/training/certification/cdp-dataanalyst-exam-cdp-4001.html and/or CDP Data Engineer exam https://www.cloudera.com/about/training/certification/cdp-data-engineer-exam-guide-cdp-3002.html

Certification is a great differentiator. It helps establish you as a leader in the field, providing employers and customers with tangible evidence of your skills and expertise.

Trainer

Certified Cloudera Instructor

Other training Cloudera | Cloudera Data Analyst

Contact form

Please fill form below to obtain more info about this training.







* Fields marked with (*) are required !!!

Information on data processing by Compendium - Centrum Edukacyjne Spółka z o.o.

3000 EUR

FORM OF TRAINING ?

 

TRAINING MATERIALS ?

 

SELECT TRAINING DATE

    • General information
    • Guaranteed dates
    • Last minute (-10%)
    • Language of the training
    • English
Book a training appointment
close

Traditional training

Sessions organised at Compendium CE are usually held in our locations in Kraków and Warsaw, but also in venues designated by the client. The group participating in training meets at a specific place and specific time with a coach and actively participates in laboratory sessions.

Dlearning training

You may participate from at any place in the world. It is sufficient to have a computer (or, actually a tablet, or smartphone) connected to the Internet. Compendium CE provides each Distance Learning training participant with adequate software enabling connection to the Data Center. For more information, please visit dlearning.eu site

close

Paper materials

Traditional materials: The price includes standard materials issued in the form of paper books, printed or other, depending on the arrangements with the manufacturer.

Electronic materials

Electronic materials: These are electronic training materials that are available to you based on your specific application: Skillpipe, eVantage, etc., or as PDF documents.

Ctab materials

Ctab materials: the price includes ctab tablet and electronic training materials or traditional training materials and supplies provided electronically according to manufacturer's specifications (in PDF or EPUB form). The materials provided are adapted for display on ctab tablets. For more information, check out the ctab website.

Upcoming Cloudera training

Training schedule Cloudera