Professional Documents
Culture Documents
WHAT IS ANALYTICS?
Analytics is the process of discovering, interpreting, and communicating
significant patterns in data. . Quite simply, analytics helps us see insights and
meaningful data that we might not otherwise detect. Business analytics focuses
on using insights derived from data to make more informed decisions that will
help businesses increase sales, reduce costs, and make business improvements.
ANALYTICS IN FINANCE
A top consumer bank in Asia enjoyed a significant market share but lagged its
competitors in products per customer. It used advanced analytics to explore
several big data sets: customer demographics and key characteristics, products
held, credit-card statements, transaction and point-of-sale data, online and mobile
transfers and payments, and credit-bureau data. The bank discovered
unsuspected similarities that allowed it to define 15,000 microsegments in its
customer base. It then built a next-product-to-buy model that increased the
likelihood to buy three times over. This is one of the many examples where
analytics has changed the way of functioning of the Finance Industry.
Analytics helped modernize the financial processes and information standards
and put the core finance data in one place, which helped the companies from
many legacy inefficiencies. It also helped build machine learning and artificial
intelligence models for businesses using practical drivers, which helped to make
better strategic decisions and a fair resource allocation. Portfolio management is
one such activity where ML models manage the funds and generate returns
outperforming the Index actively. Advanced statistical or visualization methods
like cognitive frameworks, interactivity and storytelling helped companies avoid
unpleasant surprises and incorporate risk mitigating strategies in their day-to-day
functioning. The past decade has also seen the growth of blockchain-based
financial applications. It enabled more open, inclusive, and secure business
networks, shared operating models, reduced costs, and new products and
services. JP Morgan is an early adopter of this technology. They created JPM
Coin to facilitate real-time value movement and solve hurdles of cross border
transactions.
Financial specialists often have to work on semi-structured or structured data.
Various data science tools such as Apache Spark, SAS, Scikit-learn etc., are
used to process and handle the unstructured data sets.
These data help gain customer insights and understand/predict their behaviour.
Meaningful insights from data can be generated using various tools such as text
analytics, data mining, Natural Language Processing (NLP) and many more. A
thorough analysis is conducted on the data of customers using machine learning
algorithms to analyze the changes and trends in the financial market and values.
Based on these customer insights, personalized delivery of services is made
possible and personalized relationship with customers is ensured by increased
engagement. Simple application is suggestion of schemes or funds to the
customer based on the behavioral patterns.
These were some of many uses of big data & analytics in finance.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
ANALYTICS IN
OPERATIONS
Supply chain analytics use data and quantitative methods to improve decision
making across the Supply chain. It first expands the data set for analysis and then
applies statistical tools to draw meaning insights. Several use-cases are
mentioned below:
Sourcing
Manufacturers deal with 2 broad issues using analytics. The supply risk which is
modeled around Predictive Risk Management by incorporating the necessary
internal and external factors. The cost of supply is handled by building ‘clean cost
sheets’ to declutter the vast pricing information to gain enhanced bargaining
power.
Manufacturing
Machine breakdowns are costly and recurrent in the manufacturing industry.
Predictive maintenance helps operators to monitor the health of the machine
thereby reducing the risk of breakdowns. Another unavoidable cost is the cost of
rework/reject, where the main problem is that it’s very time-consuming to identify
the cause of the defect and which leads to a line of defective items in the interim.
Collecting data across the assembly line and running automatic diagnostics
(checks dimensions of the item at multiple 3D points) on the defective item helps
in identifying both the defect and its root cause quickly.
Warehousing
Long shipping times lead to high inventories. This resulted in the increase of both
size and number of warehouses. Reducing the shipping times will reduce the
warehousing costs manifold. Analytics (using data from sensors) allows business
to track shipments and better manage warehouses
Transportation
Data Analytics helps track fleet performance and suggest maintenance schedules
by gathering data from on-fleet sensors to enable cost savings and better fleet
management.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
ANALYTICS IN
OPERATIONS
Point of Sale
Shelf space management and Product placement are the two key drivers in a
retail store. Managing the shelf space & bundling relevant products will lead to a
significant rise in the store’s revenue. Chaotic sales data can be used for this
purpose to decide which products to place where and which products are to be
put together. Detecting stockouts is another major concern. Manual inspection to
check inventory levels is often costly. Data analytics tools monitor the sales
pattern and send an alert to the inventory department when a high selling item
has suddenly stopped appearing at the POS. The inventory management checks
whether that item is out of stock in the shelf. It also identifies high sale products
depending on the seasonal/erratic unexpected demand spike and recommends
the personnel to place them in strategic locations. This leads to better
supply/demand management thereby increase in revenues.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
ANALYTICS IN PRODUCT
MANAGEMENT
Product analytics is the process of analyzing how users engage with a product or
service. It enables product teams to track, visualize, and analyze user
engagement and behaviour data. Teams use this data to improve and optimize a
product or service. Product Analytics helps in understanding engagement and
retention while Marketing Analytics helps in understanding traffic and acquisition.
In order to get a quantitative understanding of what users are doing with the
product, the first step is instrumenting it with product analytics. The idea is to fire
an event for every action that a user can take in the product, so you get an
aggregated view of how many users use a feature, and how often they're using it.
For example, if a PM wants to track the number of times a user clicks a specific
button, he/she might fire an event called "big-red-button. click." From there a PM
can see which features need work, which are your most important, and use that
information to prioritize changes.
User actions are commonly called Events. Events include clicks, slides, gestures
(for mobiles), play commands (for audio and video), downloads, page loads, and
text field fills. The event includes the type of element, the name of the element,
and the action the user took. Generic examples of events include Create Account,
Add to List, Submit Feedback, Share Dashboard, Select Option, Play Tutorial,
Change View, and Complete Onboarding.
The way one understands the specific attributes of the tracked interactions is the
work of Event Properties. PMs also want to know the context that distinguishes
activity from impact when analyzed longitudinally. Event properties can include
details like time, duration, count, device, software version, geography, user
demographic, account firmographic, element characteristics (like colour, size,
shape), Boolean (like login: yes/no), and custom attributes (like
basic/pro/enterprise).
There are 3 underlying units of measurements that help in quantifying the product
analytics data: Goals, Key Performance Indicators (KPIs), and Metrics. Goals are
a company’s highest-level priority, such as driving revenue; KPIs measure
progress toward goals; and metrics measure progress toward KPIs.
Product management metrics are usually divided into categories based on what
part of the user lifecycle they’re intended to measure:
• Engagement: How frequently, do users engage with the product? This could
be the number of key actions taken or minutes of video watched.
• Retention: What proportion of users come back to use the product? This is
usually measured across 7-, 30-, or 90-day periods.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
ANALYTICS IN PRODUCT
MANAGEMENT
• Activation: How quickly do users reach a point that helps them realize value?
This could be making a first purchase or listening to three podcasts.
• Acquisition: How broad of a user base does the product have over a recent
time period? This can be the number of paid accounts in the past year, or the
number of users who made a purchase in the past three months.
• Monetization: How many, or how often, are transactions made? How well is
usage of the product translating to revenue for the business? Transaction
metrics can cover in-app purchases to revenue and ad conversions.
• Business-specific: Metrics specific to the business model. For example, an e-
commerce company might want to track average order value.
ANALYTICS IN
CONSULTING
The word Analytics is now ubiquitous. Whether it is IT, BFSI, Manufacturing,
Supply Chain or the Automotive industry, data analytics has spread its wings in
every industry. Two things have led to this unprecedented rise in usage of data
analytics: The amount and variety of data generated in the world today and the
Technological advancements in Big Data and Artificial Intelligence in capturing
and analyzing this data.
By leveraging this technological disruption, consulting firms can help clients
identify and capture most value and meaningful insights from their data and turn
them into competitive advantages. Analytics can help companies eliminate
strategies which will not work out or identify a company's strengths and
weaknesses to build a strategy around it. The big three management
consultancies have started placing a lot of emphasis on usage of Data Analytics
to create "Change that Matters”. Firms can also advise clients on leveraging
analytics to enhance their businesses. The bottom-line here is that in the era of
digital disruption, analytics is essential for survival.
However, consultants should also learn new skill sets to stay ahead of this
disruption, tools such as Excel, Tableau or Microsoft Power BI allow instant
analysis of big data and help in generation of reports and dashboards. These
tools help in discovering hidden business insights and consultants can use these
reports to make preliminary analysis of client's problems and then delve down
deeper into key pain points and strategize solutions around them.
Analytics can give consulting firms an edge by accomplishing the following:
Reduce bias in decisions and provide absolute clarity: By using data analytics,
consultants can add a new data-oriented view to their strategy, this helps in
evaluating the odds of success of their strategy before allocating resources to that
strategy. Data driven decision making provides absolute clarity, helps in making
better decisions, gives less errors and more positive results.
Extract new growth opportunities: Analytics can also enhance strategic planning
by unearthing growth opportunities that would otherwise be hard to spot, be they
attractive industry segments and acquisition targets, ideas for new products or
services, or even new applications for existing offerings. With NLP (Natural
Language Processing) gaining traction, parsing and extracting patterns from
unstructured and structured textual data has become affective, thus helping firms
extract hidden insights and implement broader range of growth opportunities. Text
Analytics solutions, Chatbot solutions, Sentiment and Intent Analysis are some
examples where firms can use NLP to standout.
Identify early-stage trends: Machine Learning can analyze, in real time, publicly
available data spanning millions of web pages, social media content, news
sources, patent filings, and more. Discovering patterns from these data sources,
can help businesses identify emerging trends by, for example, let us assume Ola
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
ANALYTICS IN
CONSULTING
Electric is trying to decide which electric-vehicle battery technology to invest in.
This is not a minor decision—the capital expenditure alone would run into billions
of dollars and will also lock the manufacturer into a specific technology for many
years. Senior management in this company would benefit from knowing how
associated trends are evolving and when a specific technology is likely to have a
clear advantage. They could gain these insights through near real-time tracking of
patent and academic-publication momentum, announcements, and investments
across different technologies. They could also track government regulatory
changes such as zero-emission- vehicle mandates that stimulate demand for
electric vehicles or local ownership rules in countries where lithium supply is
concentrated. By using analytics to track emerging trends, consultancy firms can
help their clients make smart decisions before their competitors do.
BASIC OF STATISTICS
SAMPLING
Sampling is the process of choosing small number of observations from a large
dataset. The biggest challenge is to choose a subset which accurately represents
your original dataset with limited datapoints
One of the most common sampling techniques used in everyday life, where all
elements supposedly have equal chance of occurrence. Hence, every element is
equally likely making the sample an ideal representation of population
Simple Random Sampling - All elements of the sample are randomly selected,
preserving the quality of data.
Example – Choosing a number from 1 -10
BASIC OF STATISTICS
Cluster Sampling - We divide the sample into random clusters and then pick
elements from them. The difference between cluster and stratified random is that
in stratified random, the user divides population based on similar characteristics
while in clustering sample we just randomly divide the population and pick
samples from each cluster.
Example – An NGO creating a sample of girls from 5 neighboring districts to
provide education
BASIC OF STATISTICS
Quota Sampling is where the members are selected according to some specific
characteristics chosen by the researcher. These specific characteristics serve as
a quota for selection of members of the sample. Hence, the members are
selected based on these specific characteristics.
HYPOTHESIS TESTING
It is an elemental form of statistical testing to compare the characteristics of 2 or
more samples. Hypothesis testing begins with a Null Hypothesis. A null
hypothesis is the assumption that 2 features in 2 samples are not statistically
different. We associate a confidence level with this line of argument –Alpha.
Alpha is the probability of incorrectly rejecting the null hypothesis or rejecting a
null hypothesis even though it is true. Using some statistical test (which we will
soon discuss) we calculate a p-value which is the probability of existence of
extreme cases(in violation of null hypothesis)in the data as indicated by null
hypothesis if it was true. Thus, if the p-value is greater than alpha, then we reject
null hypothesis and conclude that the feature in our sample is statistically
different. This means that occurrence of cases in the data is much larger than our
tolerable levels, which means it is not rare or extreme –but rather than the trend
of the data.
TESTING METHODS
• Z Test: The most basic and common type of testing done. This is used to
compare a similar feature in samples from two different populations. Eg:
Comparing the effect of a new drug on the BMR of youths and old people.
• T Test: This is the most common test used in clinical trials and tests. This is
used to compare statistical difference in samples from the same population
with the effect of an external agent. Thus, one sample is control – no effect of
agent & the other is experiment –with effect of agent. Thus, the test compares
before and aftereffects of agents. E.g.: Effect of chamomile tea on youth BMR
• ANOVA – Analysis of Variance: ANOVA helps to compare more than 2
samples of the same population (which is the limitation of t Test).
• Chi Square Test: This is the statistical test for qualitative data. This is the
used compare the dependence of 2 features on each other. E.g.: The
concurrence of male gender and taller height in children of age 5
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASIC OF STATISTICS
CORRELATION
BASIC TYPES
TYPES OF PROBLEMS
UNSUPERVISED LEARNING
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
DECISION TREES
Decision Trees (DTs) are a non-parametric supervised learning method used
for classification and regression. The goal is to create a model that predicts the
value of a target variable by learning simple decision rules inferred from the data
features. A tree can be seen as a piecewise constant approximation.
COMMON TERMINOLOGIES
• Root Node: Root node is from where the decision tree starts. It represents the
entire dataset, which further gets divided into two or more homogeneous sets.
• Leaf Node: Leaf nodes are the final output node, and the tree cannot be
segregated further after getting a leaf node.
• Splitting: Splitting is the process of dividing the decision node/root node into
sub-nodes according to the given conditions.
• Branch/Sub Tree: A tree formed by splitting the tree.
• Pruning: Pruning is the process of removing the unwanted branches from the
tree.
• Parent/Child node: The root node of the tree is called the parent node, and
other nodes are called the child nodes.
• Attributes : These were the specific features of a data set. For a dataset of
employees, we can consider salary, working hours, rating etc., as the attributes
Step–1 : Begin the tree with the root node, say S, which contains the complete
dataset
Step–2 : Find the best attribute in the dataset using Attribute Selection Measure
(ASM). Two popular techniques for ASM are Information gain and Gini Index.
Step–3: Divide S into subsets containing possible values for the best attributes.
Step–4: Generate the decision tree node, which contains the best attribute.
Step–5: Recursively make new decision trees using the subsets of the dataset
created in step–3. Continue this process until a stage is reached where you cannot
further classify the nodes and called the final node as a leaf node.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
DECISION TREES
INFORMATION GAIN & GINI INDEX
EXAMPLE
Problem Statement : We are trying to classify flowers into different classes like
Versicolor, Virginica, Setosa based on the attributes like petal length, petal width,
sepal width, sepal length etc.,
ASM Used : Gini Index
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
Variables Factors
Records
Imagine three
variables, X1, X2 and
X3 that are correlated
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
Selection of the factors is done to reduce the variables that would be used
for further processing. The tradeoff is to minimize the information lost by
reducing the initial variables.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
TYPES OF CLUSTERING
HIERARCHICAL CLUSTERING
• A clustering procedure characterized by the development of a hierarchy
or tree like structure
• Search for the 2 points that are most similar and group the next most
similar points – going in an iterative manner
• Consists of 2 distinct methods: Agglomerative where we assimilate from
single point clusters to whole dataset and Divisive where we break the
whole dataset to reach single point clusters
NON–HIERARCHICAL CLUSTERING
• Number of clusters are pre-defined, and clusters are built around it
• Algorithm find ‘k’ points farthest from one another and ten starts
grouping other points basis the distance from these ‘k’ seeds (K-Means
Clustering)
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
=LET(x,1,x+1) → Output = 2
=LET(x,1,y,1,x+y) → Output = 2
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
Lookup_value (Required)
The value to search for
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
LAMBDA FUNCTION
LAMBDA function allows the user to create custom, reusable functions and
call them at will.
=LAMBDA(x,y,x+y)
=mylambda(1,3) → Output = 4
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
TEXTJOIN FUNCTION
The TEXTJOIN function combines the text from multiple ranges and/or
strings, and includes a delimiter you specify between each text value that
will be combined. If the delimiter is an empty text string, this function will
effectively concatenate the ranges.
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
EXCEL SHORTCUTS
FUNCTION WINDOWS MAC
Display the Paste Special dialog box Ctrl + Alt + V ⌘+⌃+V
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
EXCEL SHORTCUTS
FUNCTION WINDOWS MAC
Move active cell clockwise in selection Ctrl + . ⌃+.
Extend the selection to the last cell right Ctrl + Shift + → ⌃+⇧+→
Extend the selection to the last cell left Ctrl + Shift + ← ⌃+⇧+←
Extend the selection to the last cell down Ctrl + Shift + ↓ ⌃+⇧+↓
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
EXCEL SHORTCUTS
FUNCTION WINDOWS MAC
Start a new line in the same cell Alt + Enter ⌃ + ⌥ + Return
Display Format Cells with Font tab selected Ctrl + Shift + F ⌃+⇧+F
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
EXCEL SHORTCUTS
FUNCTION WINDOWS MAC
Apply general format Ctrl + Shift + ~ ⌃+⇧+~
Calculate worksheets F9 Fn + F9
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
EXCEL SHORTCUTS
FUNCTION WINDOWS MAC
Calculate active worksheet Shift + F9 Fn + ⇧ + F9
Source: Link
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
DATA WAREHOUSING
BASICS OF SQL
OLTP & OLAP
Online transaction processing (OLTP) captures, stores, and processes data
from transactions in real time. It uses 3NF Normalisation to reduce data
redundancy and speed of the system. It's write speeds are very fast and it
does not contain historical data in most cases.
Online analytical processing (OLAP) uses complex queries to analyse
aggregated historical data from OLTP systems.
STAR SCHEMA
The most basic unit of a star schema consists of a set of dimension tables and a
fact table.
The dimension tables contain core attributes of the data e.g. date, product, place
etc. They contain a PK that uniquely identifies a row in that dimension
(sometimes surrogate keys are used to add this unique identifier).
The fact tables contains the business measure, KPIs and values. They contain
FK references for the PK values of dimension tables. The dimensionality of the
fact table is the number of FK references it contains. The granularity of the data is
determined by how deep we can drill down from the broadest level of the data.
It can also be understood as the last level at which data can be viewed
successfully.
Generally, dimension tables contain a relatively small number of rows. Fact
tables, on the other hand, can contain a very large number of rows and continue
to grow over time.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
STAR SCHEMA
BASICS OF SQL
SNOWFLAKE SCHEMA
It is an extension of a star schema model. The only variation is that her the
dimension side is normalized whereas that is not the case in STAR schema.
So essentially a dimension from STAR schema is broken down into normalized
entities to give a SNOWFLAKE schema.
They reduce the size of the dimensions by reducing redundancies but can adds
to additional query time because multiple joins have to be made to get to a
filtering criteria.
In a Type 1 SCD the new data overwrites the existing data. Thus the existing
data is lost as it is not stored anywhere else. This is the default type of
dimension you create. You do not need to specify any additional
information to create a Type 1 SCD.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
SLOWLY CHANGING DIRECTIONS
DATA NORMALISATION
For data to be of the 1NF you need to ensure that the data is atomic. A data item
is atomic if only one item is in each cell of a table.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
To be in the second normal form the table has be to 1NF and the non-PK data
should have full functional dependency on the PK in the table
For a table to be in the 3NF form, it needs to be in the 2NF form and should not
have any transitive dependencies.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
THIRD NORMAL FORM (3NF)
Also known as 3.5 NF, it is the higher version 3NF. The table should be in 3NF
form and then should be in such a state that no non-PK attribute determines any
PK attribute.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
TYPES OF STATEMENTS
• Data Definition Language (DDL) statements defines data structures. Use
these statements to create, alter, or drop data structures in a database.
These include - ALTER, CREATE, DROP, RENAME, UPDATE STATISTICS
• Data Manipulation Language (DML) statements affect the information
stored in the database. Use these statements to insert, update, and
change the rows in the database. These include - DELETE, INSERT,
SELECT, UPDATE, MERGE, TRUNCATE TABLE
• Data Control Language(DCL) statements defines the control over the
data in a database. These include - GRANT and REVOKE
• Transaction Control Language (TCL) statements are used to manage the
transactions in a database. These are used to manage the changes
made by DML statements. It also allows the statements to be grouped
together into logical transactions. These include - COMMIT and
ROLLBACK
BASICS OF SQL
SUBQUERIES
BASICS OF SQL
DDL & DML STATEMENTS
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
DDL & DML STATEMENTS
VIEWS
A view is a virtual table whose contents are defined by a query. Like a table,
a view consists of a set of named columns and rows of data. Unless
indexed, a view does not exist as a stored set of data values in a database.
The rows and columns of data come from tables referenced in the query
defining the view and are produced dynamically when the view is
referenced.
A view acts as a filter on the underlying tables referenced in the view. The
query that defines the view can be from one or more tables or from other
views in the current or other databases.
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
VIEWS
JOINS
BASIC JOIN
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
LEFT JOIN (OR LEFT OUTER JOIN)
BASICS OF SQL
EXCEPT, UNION AND UNION ALL
EXCEPT
BASICS OF SQL
OTHER QUERIES
CASE
DATEADD
SUBSTRING
BASICS OF ANALYTICS IN
ANALYTICAL TOOLS EXCEL & SQL
ANALYTICS PRACTICE
BASICS OF SQL
COALESCE
The PARTITION BY clause divides the result set into partitions (another
term for groups of rows). The ROW_NUMBER() function is applied to each
partition separately and reinitialized the row number for each partition. The
PARTITION BY clause is optional. If you skip it, the ROW_NUMBER()
function will treat the whole result set as a single partition.
The ORDER BY clause defines the logical order of the rows within each
partition of the result set. It essentially determines the order of the flow.
BASICS OF SQL
CONTENT &
DESIGN TEAM
Vinayak Nadir
Ashwin Palnitkar
REACH US
AT
Srinivasan G