Sie sind auf Seite 1von 14

Post Graduate Certification In Data Science & Artificial Intelligence

Get Certified By TCS iON ProCert - a TCS brand


The Post Graduate Certification In Data Science & Artificial Intelligence (PGC-DSAI) is a 6 month weekend
classroom program for working professionals & freshers alike. This program is one of the most comprehensive
ones available in India and covers every aspect of data science. It will equip you with all the technologies,
conceptual knowledge and skillsets that you need to crack any data science & AI interview, transition into a career
in this field and prosper in it. This is a course designed to get you a job in data science! TCS iON ProCert has tied up
with Edvancer to be our certification partner. You will be certified by TCS iON ProCert post a final exam held by
them and by Edvancer too post completing all the required projects in the course.

Program Highlights

180 HOURS OF INDUSTRY BOOT-CAMP STYLE TRAINING TOP CLASS FACULTY FROM
EXPERT LED CLASSROOM WITH 70% PRACTICAL MICROSOFT, GOOGLE,
SESSIONS CONTENT AMAZON, DELOITTE ETC.

24X7 LIFETIME ACCESS TO CREATE A FULL- FLEDGED CERTIFICATE FROM TCS ION
ONLINE LEARNING CONTENT AI PRODUCT AS YOUR PROCERT & EDVANCER WITH
& VIDEOS CAPSTONE PROJECT JOB ASSISTANCE

Why Should You Take This Course

Become a full-stack data Create a job ready project Most comprehensive


scientist with core skills in portfolio to establish your curriculum covering everything
Machine Learning and skills & credibility and attract from predictive analytics,
Artificial Intelligence recruiters machine learning & AI

Get a huge hike in your We will work closely with Learn data science from
salary on becoming a data you to help build your data India's top data science
scientist science portfolio and start a training institute as ranked
data science career by industry & students
Technologies Covered

Our Students Are Placed In

Pre-requisites
1. Education: B.E/B.Tech/M.Tech/MCA/MS/MBA or B.Sc/M.Sc in Statistics/ Maths/ Physics/ Economics or
any quantitative field
2. Experience: 0-10 years of work experience.
3. You should be comfortable in learning maths and programming. Not mandatory to have a prior
programming or maths background though.

List of Industry Projects You Will Work On


1. Healthcare: Predict patient re-admission risk for a hospital chain based on their medical history data.
2. Pharma: Predict the sales volume of counterfeit medicines in order to guide law enforcement agencies
in cracking down on top counterfeiters.
3. BFSI: Identify fraudulent transactions for a credit card providing company to create an early warning
system to prevent frauds in real time.
4. BPO: Create a machine learning system which will automatically pick out customer complaints most
likely to be unresolved and escalate them.
5. Real Estate: Predict property prices and help the prop-tech startup’s agents close deals quicker.
6. Manufacturing: Predict inventory requirement of spare-parts for a large engineering co.
7. HR: Predict employee attrition for an IT company
8. Retail: Identify locations where a retail giant’s new stores should be opened and will do well.
9. E-Commerce: Analyze web server logs of an ecommerce portal to understand more about the products
being browsed and sold.
10. Music Streaming: Create an AI system which will automatically assign genres to any music file
11. Marketing: Identify prospects for a luxury products company
12. Spam Filter: Build a robust AI based spam filter for online forums like Stack Overflow at scale.
13. Question Tag System: Create an automated system which will tag questions for Stack Overflow
14. Car Safety System: Build an AI safety system for an auto major to identify distracted drivers in their cars
and sound an alert.
15. Image recognition system: Recognize and caption images automatically through an AI model

Scholarship Process

1 2
16. 3

Enter your application details Your application will be Enrol for the course and get
on our site & upload your
17.
reviewed by our experts and if the scholarship of Rs. 30,000.
CV/profile. shortlisted you will be eligible
for the scholarship

Fees
Final Course fee: Rs. 75,000 + 18% GST after a 20% discount and scholarship of Rs. 30,000.
Pay the fee in 6 interest-free instalments* post a 10% down-payment. Effectively you will be paying just Rs.
14,750/- per month for 6 months post down-payment, if you clear the scholarship application process &
get the discount.
(*instalment offer subject to approval from our financing partner based on Aadhar card and 4 months bank statements)

Payments can be made online through credit cards, debit cards or net-banking.

About Edvancer
Edvancer is India's leading data science training institute where we provide a range of courses on data
science to all levels of learners. We have trained over 5,000 students and delivered over 10,00,000+ hours
of learning. Our alumni work with some of India's top companies in data science and even globally. Our
corporate clients include PwC, E&Y, L&T, HP, JP Morgan, Cognizant, Accenture, TCS, Microsoft etc.
Full Curriculum
Module 1: Predictive Analytics in R
What is this module about?: Predictive Analytics is the scientific process of deriving insights from raw
data to support decision making and is the core of data science. Through this module you will learn how to
use analytical techniques and the R language to solve business problems. This is a comprehensive module
which will take you from the basics of statistical techniques, R language right up to building predictive
models.

Tools to be learnt: R
Class Duration: 64 hours

Topic What does it mean?


Introduction to business analytics
• What is analytics & why is it so important? In this section we shall provide you an overview
into the world of analytics. You will learn about the
• Applications of analytics
various applications of analytics, how companies
• Different kinds of analytics are using analytics to prosper and study the
• Various analytics tools analytics project methodology through a real-world
• Analytics project methodology case study
• Real world case study
R Training
Fundamentals of R
• Installation of R & R Studio
• Getting started with R
• Basic & advanced data types in R
• Variable operators in R This part is all about learning how to manage and
• Working with R data frames manipulate data and datasets, the very first step of
• Reading and writing data files to R analytics. We shall teach you how to use R to work
• R functions and loops with data using a case study.
• Special utility functions
• Merging and sorting data
• Case study on data management using R
• Practice assignment
Data visualization in R
• Need for data visualization Data visualization is extremely important to
• Components of data visualization understand what the data is saying and gain
insights in just one glance. Visualization of data is a
• Utility and limitations
strong point of the R software and you will learn
• Introduction to grammar of graphics the same in this module.
• Using the ggplot2 package in R to create visualizations
Data preparation and cleaning using R
• Needs & methods of data preparation Real world data is rarely going to be given to you
• Handling missing values perfect on a platter. It will always be dirty with
• Outlier treatment missing data points, incorrect data, variables
needing to be changed or created in order to
• Transforming variables
analyze etc. A typical analytics project will have
• Derived variables 60% of its time spent on preparing data for analysis.
• Binning data This is a crucial process as properly cleaned data
• Modifying data with Base R will result in more accurate and stable analysis. We
• Data processing with dplyr package shall teach you all the techniques required to be
• Using SQL in R successful in this aspect.
• Practice assignment
Setting the base of business analytics
Understanding the data using univariate statistics in R This is where you shall learn how to start
• Summarizing data, measures of central tendency understanding the story your data is narrating by
• Measures of variability, distributions summarizing the data, checking its variability and
• Using R to summarize data shape by visualizing it. We shall take you through
• Case study on univariate statistics using R various ways of doing this using R and also solve a
• Practice assignment case study
Hypothesis testing and ANOVA in R to guide decision
making With 95% confidence we can say that there is an
• Introducing statistical inference 85% chance, people visiting this site twice will
• Estimators and confidence intervals enroll for the course ☺. Here, you learn how to
• Central Limit theorem create a hypothesis, test and validate it through
• Parametric and non-parametric statistical tests data within a statistical framework and present it
• Analysis of variance (ANOVA) with clear and formal numbers to support decision
making.
• Conducting statistical tests
• Practice assignment
Predictive modelling in R

1. Correlation and Linear regression


• Correlation A statistical model is the core of predictive analytics
and regression is one of the most powerful tools for
• Simple linear regression
making predictions by finding patterns in data. You
• Multiple linear regression shall learn the basic of regression modelling hands-
• Model diagnostics and validation on through real world cases
• Case study
Logistic regression is the work-horse of the
2. Logistic regression predictive analytics world. It is used to make
• Moving from linear to logistic predictions in cases where the outcomes are dual in
• Model assumptions and Odds ratio nature i.e. an X or Y scenario where we need to
• Model assessment and gains table predict if X will be the case or will Y, given some
• ROC curve and KS statistic data. This is a must-know technique and we shall
• Case Study make you comfortable with it through real world
problems.
3. Techniques of customer segmentation
• Need for segmentation Learn why and how to statistically divide a broad
• Criterion of segmentation customer market into various segments of
• Types of distances customers who are similar to each other so as to be
• Hierarchical clustering able to better target and meet their needs in a cost
• K-means clustering effective manner. This is one of the most essential
• Deciding number of clusters techniques in marketing analytics.
• Case study
4. Time series forecasting techniques The ability to forecast into the future is very
• Need for forecasting important for any business and it is necessary to
• What are time series? have as accurate a forecasting as possible for
• Smoothing techniques financial and strategic planning. In this module
• Time series models learn the techniques of time series analysis without
• ARIMA being misled by seasonal and cyclical impacts.

5. Decision trees & Random Forests


• What are decision trees Decision trees are predictive models which map
• Entropy and Gini impurity index observations about an item to conclusions about
• Decision tree algorithms the item's target value. Learn the technique of
• CART decision trees, one of the most popular predictive
• Random Forest analytics techniques
• Case Study
6. Boosting Machines Want to win a data science contest on Kaggle or
• Concept of weak learners data hackathons or be known as a top data
• Introduction to boosting algorithms scientist? Then learning boosting algorithms is a
• Adaptive Boosting must as they provide a very powerful way of
• Extreme Gradient Boosting (XGBoost) analysing data and solving hard to crack problems
• Case study
7. Cross Validation & Parameter Tuning
Learn how to make your model more accurate and
• Model performance measure with cross validation perform the best on real -world data
• Parameter tuning with grid & randomised grid search

Module 2: Machine Learning in Python


What is this module about?: Through this Machine Learning module, you will learn how to process,
clean, visualize and automate decision making through data science by using Python, one of the most
popular machine learning tools. You will learn cutting edge machine learning techniques in Python.

Tools to be learnt: Python (Libraries like pandas, numpy, scipy, scikit-learn, bokeh, beautifulsoup)
Class Duration: 56 hours

Topic What does it mean?


In this section we shall provide you an overview
Introduction to Machine Learning in Python into the world of machine learning. You will learn
• What is machine learning & why is it so important? about the various applications of machine learning,
how companies from all sort of domains are solving
• Applications of machine learning across industries
their day to day to long term business problems.
• Machine Learning methodology We’ll learn about required skill sets of a machine
• Machine Learning Toolbox learning expert which make them capable of filling
• Tool of choice- Python: what & why? up this vital role. Once the stage is set and we
• Course Components understand where we are heading we discuss why
Python is the tool of choice in data science.
Python Training
Introduction to Python
• Installation of Python framework and packages: Anaconda
and pip Python is one of the most popular & powerful
languages for data science used by most top
• Writing/Running python programs using Spyder, Command
companies like Facebook, Amazon, Google, Yahoo
Prompt
etc. It is free and open source. This module is all
• Working with Jupyter Notebooks about learning how to start working with Python.
• Creating Python variables: Numeric, string and logical We shall teach you how to use the Python language
operations to work with data.
• Basic Data containers: Lists, Dictionaries, Tuples & sets
• Practice assignment
Iterative Operations & Functions in Python This is where we move beyond simple data
• Writing for loops in Python containers and learn about amazing possibilities
• List & Dictionary Comprehension and functionalities hidden in various associated
operators. We get introduced to wonderful world
• While loops and conditional blocks
of loops, list and dictionary comprehensions. In
• List/Dictionary comprehensions with loops addition to already existing functions and classes
• Writing your own functions in Python we learn to write our own custom functions and
• Writing your own classes and functions as class objects classes. This module sets the stage for handling
• Practice assignment data and ML algorithm implementation in python.
Data Summary; Numerical and Visual in Python
• Need for data summary Data summary is extremely important to
• Summarising numeric data in pandas understand what the data is saying and gain
• Summarising categorical data insights in just one glance. Visualization of data is a
• Group wise summary of mixed data strong point of the Python software using the latest
• Need for visual summary ggplot package using much celebrated grammar of
• Introduction to ggplot & Seaborn graphics. We also introduce you another powerful
package seaborn in additional material section.
• Visual summary of different data combinations
• Practice Exercise
Data Handling in Python using NumPy & Pandas Python is a very versatile language and in this
• Introduction to NumPy arrays, functions &properties module we expand on its capabilities related to
• Introduction to pandas data handling. Focusing on packages numpy and
pandas we learn how to manipulate data which will
• Dataframe functions and properties
be eventually useful in converting raw data suitable
• Reading and writing external data for machine learning algorithms.
• Manipulating Data Columns
Machine Learning in Python
In this module we understand how we can
transform our business problems to data problems
Basics of Machine Learning
so that we can use machine learning algorithms to
• Business Problems to Data Problems solve them. We will further get into discovering
• Broad Categories of Business Problems what all categories of business problems and
• Supervised and Unsupervised Machine Learning Algorithm subsequently which machine learning algorithms
• Drivers of ML algorithms are there.
• Cost Functions
• Brief introduction to Gradient Descent We’ll learn what is the ultimate goal of any
• Importance of Model Validation machine learning algorithm and go through a brief
description of the mother of many modern
• Methods of Model Validation
optimisation methods- Gradient Descent. We’ll
• Introduction to Cross Validation and Average Error wrap up this module with discussion on importance
and methods of validation of our results.
Generalised Linear Models in Python
• Linear Regression We start with implementing machine learning
• Limitation of simple linear models and need of algorithms in this module. We also get exposed to
regularisation some important concepts related to regression and
• Ridge and Lasso Regression (L1 & L2 Penalties) classification which we will be using in the later
• Introduction to Classification with Logistic Regression modules as well. Also this is where we get
• Methods of threshold determination and performance introduced to scikit-learn, the legendary python
measures for classification score models library famous for its machine learning prowess.
• Case Studies

Tree Models using Python In this module we will learn a very popular class of
• Introduction to decision trees machine learning models, rule based tree
• Tuning tree size with cross validation structures also known as Decision Trees. We'll
• Introduction to bagging algorithm examine their biased nature and learn how to use
bagging methodologies to arrive at a new
• Random Forests
technique known as Random Forest to analyse
• Grid search and randomized grid search data. We’ll further extend the idea of randomness
• ExtraTrees (Extremely Randomised Trees) to decrease bias in ExtraTrees algorithm.
• Partial Dependence Plots In addition, we learn about powerful tools used
• Case Studies with all kind of machine learning algorithms,
• Home exercises gridSearchCV and RandomizedSearchCV.
Boosting Algorithms using Python
• Concept of weak learners Want to win a data science contest on Kaggle or
• Introduction to boosting algorithms data hackathons or be known as a top data
• Adaptive Boosting scientist? Then learning boosting algorithms is a
• Extreme Gradient Boosting (XGBoost) must as they provide a very powerful way of
• Case study analysing data and solving hard to crack problems.
• Home exercise
Support Vector Machines (SVM) and KNN in Python We step in a powerful world of “observation based
• Introduction to idea of observation based learning algorithms” which can capture patterns in the data
• Distances and Similarities which otherwise go undetected. We start this
• K Nearest Neighbours (KNN) for classification discussion with KNN which is fairly simple. After
• Introduction to SVM for classification that we move to SVM which is very powerful at
• Regression with KNN and SVM capturing non-linear patterns in the data.
• Case study
• Home exercises
Unsupervised learning in Python Many machine learning algorithms become difficult
• Need for dimensionality reduction to work with when dealing with many variables in
• Introduction to Principal Component Analysis (PCA) the data. In comes to rescue PCA which solves
problems arising from data which has highly
• Difference between PCAs and Latent Factors
correlated variables. The same idea can be
• Introduction to Factor Analysis extended to find out hidden factors in our data
• Patterns in the data in absence of a target with Factor Analysis which is used extensively in
• Segmentation with Hierarchical Clustering and K-means surveys and marketing analytics.
• Measure of goodness of clusters
• Limitations of K-means We also learn about two very important
• Introduction to density based clustering (DBSCAN) segmentation algorithms; K-means and DBSCAN
and understand their differences and strengths.
Neural Networks
• Introduction to Neural Networks
Artificial Neural Networks are the building blocks of
• Single layer neural network
artificial intelligence. Learn the techniques which
• Multiple layer Neural network replicate how the human brain works and create
• Back propagation Algorithm machines which can solve problems like humans.
• Moment up and decaying learning rate in context of
gradient descent
• Neural Networks implementation in Python
Text Mining in Python
Unstructured text data accounts for more and more
• Quick Recap of string data functions interaction records as most of our daily life moves
• Gathering text data using web scraping with urllib online. In this module we start our discussion by
• Processing raw web data with BeautifulSoup looking at ways to collect all that data. In addition
• Interacting with Google search using urllib with custom to scraping simple web data; we’ll also learn to use
user agent data APIs with example of Twitter API, right from
• Collecting twitter data with Twitter API the point of creating a developer account on
• Introduction to Naive Bayes twitter. Further we discuss one of the very
• Feature Engineering for text Data powerful algorithm when it comes to text data;
Naive Bayes. Then we see how we can mine the
• Feature creation with TFIDF for text data
text data.
• Case Studies
Ensemble Methods in Machine Learning Individual machine learning models extract pattern
• Making use of multiple ML models taken together from the data in different ways , which at times
• Simple Majority vote and weighted majority vote results in them extracting different patterns from
the data. Rather than sticking to just one algorithm
• Blending
and not making use of other’s results is what we
• Stacking move past in this module. We learn to make use of
• Case Study multiple ML models taken together to make our
predictive modelling solutions even more powerful.
Bokeh For making quick prototypes of your solutions
• Introduction to Bokeh charts and plotting which can be scaled later as interactive
visualisation in the form of standalone or hosted
web pages, we introduce you to Bokeh, an evolving
library in python which has all the tools that you’ll
need to do the same.
Version Control using Git and Interactive Data We finish this module with a discussion on two very
Products important aspects of a data scientist’s work. First is
• Need and Importance of Version Control version control which enables you to work on large
• Setting up git and github accounts on local machine projects with multiple team members scattered
• Creating and uploading GitHub Repos across the globe. We learn about git and most
• Push and pull requests with GitHub App widely used public platform version control that is
• Merging and forking projects GitHub.
• Examples of static and interactive data products

Module 3: Deep Learning (Artificial Intelligence) Using Tensorflow and Keras


What is this module about?: Through this module, you will learn the various techniques used in the
world of artificial intelligence like deep learning techniques, reinforcement learning, NLP and computer
vision using latest Python libraries Tensorflow and Keras. This module will put you on the cutting edge of
technology and make you future proof.

Tools to be learnt: Tensorflow and Keras


Class Duration: 44 hours
Topic What does it mean?
Introduction to AI and Deep Learning
• What is AI?
Get introduced to the world of Artificial Intelligence
• How will AI change the world?
which is poised to change the entire world/
• What is Deep Learning? Understand what is deep learning and how it is used
• Uses of Deep Learning? in AI
• Examples of Deep Learning & AI

Getting Started with Tensorflow TensorFlow™ is an open source software library in


• Setting up tensor flow, gpu instance on gcp Python for high performance numerical
• Understanding computation graph and basics of tensorflow computation. Originally developed by researchers
• Implementing simple perceptron in tensor flow and engineers from the Google Brain team within
• Implementing multi layer neural network in tensor flow Google’s AI organization, it comes with strong
• Visualising training with tensor board support for machine learning and deep learning.
Deep Feed Forward & Convolutional Neural Networks
• Implementing deep neural net for image classification A feedforward neural network is an artificial neural
• Understanding convolutions, strides, padding, filters etc network wherein connections between the nodes
• Implementing CNN with tensor flow do not form a cycle and information only flows in
one direction. a convolutional neural network
• Regularizing with dropout
(CNN, or ConvNet) is a class of deep, feed-forward
• Learning rate decay and its effects artificial neural networks, most commonly applied
• Batch normalisation and its effects to analyzing visual imagery. Learn these techniques
for classifying images.
Introduction to Keras Keras is a high-level neural networks API, written in
• Basics of Keras Python and capable of running on top of
• Composing various models in Keras TensorFlow. It was developed with a focus on
enabling fast experimentation and allows for easy
• Parameter tuning in Keras with previous examples
and fast prototyping.
Recurrent Neural Networks, Long-Short Term A recurrent neural network (RNN) is a class of
Memory and Gated Recurrent Unit artificial neural network where connections
• Intro to RNN architecture between nodes form a directed graph along a
sequence. Long Short Term Memory networks –
• Modelling sequences
usually just called “LSTMs” – are a special kind of
• Limitations of RNNs RNN, capable of learning long-term dependencies.
• Introduction to LSTM and use cases with implementation Gated recurrent units (GRUs) are a gating
(text data) mechanism in recurrent neural networks. These
• Introduction to GRU and implementation (text data) techniques are very popular for Natural Language
Processing.
Autoencoders, Generative Adverserial Networks, An autoencoder is a type of artificial neural network
Hopfield networks used to learn efficient data codings in an
• Autoencoders and dimensionality reduction unsupervised manner. The aim of an autoencoder is
to learn a representation (encoding) for a set of
• GANs and their implementation
data, typically for the purpose of dimensionality
• Hopfield networks reduction. Generative adversarial networks (GANs)
• Variational auto encoders are, implemented by a system of two neural
• Word2Vec networks contesting with each other. The purpose
• Glove of a Hopfield net is to store 1 or more patterns and
to recall the full patterns based on partial input.
These are techniques used in computer vision
problems.

Module 4: Data Visualization in Tableau (Videos Only)


What is this module about?: This video based Tableau tutorial teaches you how to visualize data using
Tableau Desktop 10. Tableau is one of the world’s most popular data visualization and BI tools and through
this module you will learn how to create data dashboards and stories to present your analysis to top
management. You get to work on 17 case studies and 3 projects in this course.

Tools to be learnt: Tableau


Class Duration: 6 hours of pre-recorded videos

Topic What does it mean?


Introduction to Business Intelligence & Visualization
In this module, you will understand what exactly is
• What is Business Intelligence?
data visualization, how does it relate to business
• What is data visualization?
intelligence & analytics and the need and uses of
• Need for Visualization visualization
• Uses of visualization
Introduction to Tableau
• What is Tableau
• Tableau vs. Excel We will introduce you to the Tableau software
• Installing Tableau Desktop which is one of the most popular tools globally for
• Overview of Tableau Desktop visualizing data. Here you will learn how to
• Various Applications of Tableau download and install Tableau on your system, why
• Components of Tableau Desktop Tableau is used instead of Excel and the various
• Benefits of Tableau and Opportunities components of the Tableau software along with
uses and benefits.
• Tableau Products & Certifications
• Tableau Architecture
• Saving and publishing your work in Tableau
Dive into Tableau
• Explore Tableau Interface
Here you will get to know Tableau more in details
• Understand various Tableau terminologies
and explore how to start the visual analysis of data.
• Create Different Views to Analyze Data
• Case Study
Connecting Data Sources In this module you will learn how to connect
• Connection Options various sources of data to Tableau, join various
• Data Types datasets to one another and write custom SQL
• Data Roles queries within Tableau.
• Joins & Over Joins
• Unions
• Custom SQL Query
• Data Blending
• Editing Connections
• Case Study 1
• Case Study 2
Tableau Generated Fields
• Use of Measure Names and Measure Values
• Compare Multiple Measures Understand how to use various fields in Tableau to
• Fetch Number of Records In Database build various views of your data
• Latitude & Longitude Fields
• Case Study
Data Manipulation in Tableau
• Creating Groups
• Creating Combined Fields
• Sorting
• Filtering Data This module will teach you how to manipulate the
• Sets data like sorting, combining fields, binning etc.
• Binning Data through multiple case studies
• Hierarchies
• Case Study 1
• Case Study 2
• Case Study 3
Working with Dates in Tableau
• Changing Date Levels
• Different Date Parts Learn how to visualize data which has dates or time
• Custom Dates in it
• Create Fiscal Dates
• Case Study
Data Customization with Calculations
• Calculated Fields
• Arithmetic Calculations
• Date Calculation
• String Calculation
Undertake calculations on your data in Tableau to
• Logical Calculation come up with new views of the data.
• Type Conversion Calculation
• Table Calculation
• Level of Detail Calculations
• String Calculation
• Case Study 1
Adding Dynamism to a View with Parameters
• Introduction to Parameters
• Create a Parameter
• Explore Parameters Parameters are dynamic values that can replace
• Use Parameters in Calculations constant values in calculations, filters, and
reference lines. Use parameters to make your
• Parameters in Reference Lines
visualization more interactive.
• Parameters in Filters
• Make Estimates using Parameters
• Case Study
Geographical Analysis & Maps
• Where is a Geographical or Map View Useful? Learn how to create map based views of your
• Creating a Map view location related data.
• Custom Geocoding
• WMS Maps
• Modify Locations

Creating Visualizations
• Bar in Bar Chart
• Scatter Plots
• Histogram
• Heat Maps
• Highlighting in Tables
• Motion Charts
• Pie Chart In this module you will learn how to create a
• Bullet Chart number of charts each with its own specific utility.
• Box & Whisker Plot Learn through multiple case studies
• KPI Chart
• Market Basket Analysis Chart
• Pareto Chart
• Waterfall Chart
• Best Practice for Selecting Chart Type
• Case Study 1
• Case Study 2
Adding Statistics to Data
• Reference Lines
• Reference Bands
• Distribution Bands Understand how to add statistics to the charts and
• Trend Lines tables
• Forecasting
• Clustering
• Summary Card
• Case Study 1
Formatting & Annotation
• Add Titles, Captions & Annotations Format your visualizations and make them more
• Formatting Options - Fonts, Shading, Borders etc. informative
• Formatting Axes, Mark Labels and Legends
Dashboards & Stories
• What are Dashboards?
• Why and How are Dashboards Useful?
• Creating an Interactive Dashboard In this module you will learn how to create the end
• Adding Actions to a Dashboard output of visualization in Tableau which is creating
• Best Practices for Dashboard Design entire dashboards & data based storyboards to
• What is a Story? present to clients and management.
• Creating a Story
• Adding a Background Image to a Story
• Case Study

Module 5: Data Analysis in SQL (Videos only)


What is this module about?: This Data Analyst using SQL video tutorial teaches you how to use the ever
popular SQL language to analyse data stored in databases. SQL is a requirement in almost all analytics roles
and this module will make you eligible to work as a data analyst. In this SQL tutorial you will learn how to
communicate with databases, extract data from them, manipulate the data, analyse it and create reports
using the data.

Tools to be learnt: MS SQL


Class Duration: 6 hours of pre-recorded videos
Topic What does it mean?
Introduction To SQL
• What is SQL? Structured Query Language (SQL) is a standard
• Why SQL? language for storing, manipulating and retrieving
• What are relational databases? data in databases. It is a heavily used language and
• SQL command group a must know for every data scientist. Here we will
• MS SQL Server installation introduce you to SQL using MS SQL.
• Exercises
SQL Data Types & Operators
• SQL Data Types
• Filtering Data Learn about various types of data and how to filter
• Arithmetic Operators and conduct basic operations on data in databases
• Comparison operators using SQL.
• Logical Operators
• Exercises
Useful Operations in SQL
• Distinct Operation
• Top N Operation
• Sorting results
Learn more advanced operations on data.
• Combine results using Union
• Null comparison
• Alias
• Exercises
Aggregating Data in SQL
• Aggregate functions
Aggregate data using various conditions and
• Group By clause
clauses in SQL to gain the answers you are looking
• Having clause for.
• Over clause
• Exercises
Writing Sub-Queries in SQL
A subquery is a SQL query within a query.
• What are sub-queries?
Subqueries are nested queries that provide data to
• Sub-query rules the enclosing query. In this module you will learn
• Writing sub-queries how to write various sub-queries.
• Exercises
Common function in SQL
• Ranking functions
• Date & time functions
Learn some of the common functions available in
• Logical functions
SQL to transform the data into more meaningful
• String functions data.
• Conversion functions
• Mathematical functions
• Exercises
Analytic Functions in SQL
• What are analytic functions?
Here you will learn various analytics function in SQL
• Various analytic functions to undertake data analysis in SQL.
• SQL syntax for analytic functions
• Exercises
Writing DML Statements
• What are DML Statements?
DML is abbreviation of Data Manipulation Language
• Insert statement
in SQL. It is used to retrieve, store, modify, delete,
• Update statement insert and update data in databases.
• Delete statement
• Exercises
Writing DDL Statements
• What are DDL Statements? DDL refers to "Data Definition Language", a subset
• Create statement of SQL statements that change the structure of the
database schema in some way, typically by
• Alter statement
creating, deleting, or modifying schema objects
• Drop statement such as databases, tables, and views.
• Exercises
Using Constraints in SQL
• What are constraints?
• Not Null Constraint
Constraints provide a standard mechanism to
• Unique constraint
maintain the accuracy and integrity of the data
• Primary key constraint inside table. There are several different types of
• Foreign key constraint constraints in SQL which you will learn here.
• Check constraint
• Default Constraint
• Exercises
SQL Joins
• What are joins?
• Cartesian Join A SQL Join statement is used to combine data or
• Inner Join rows from two or more tables. Learn the various
• Left & Right Join joins in SQL in this module.
• Full Join
• Self Join
Views in SQL A view is a virtual table that consists of columns
from one or more tables. Though it is similar to a
• What are views?
table, it is not stored in the database. It is a query
• Create View stored as an object. Hence, a view is an object that
• Drop view derives its data from one or more tables. Learn how
• Update view to create these views in this module.

Das könnte Ihnen auch gefallen