Professional Documents
Culture Documents
220
Nowadays, the impact of data science and especially machine learning in the
business world is increasing. One of the most obvious examples of this effect
is experienced in the field of price optimization. Pricing strategies are a
critical factor that directly affects the revenue of any business, and
optimizing these strategies is one of the keys to surviving in a competitive
market. The integration of machine learning technologies in this field both
strengthens data-driven decision-making processes and enables businesses
to adapt to market dynamics more quickly and effectively.
This article will discuss the opportunities and challenges that the integration
of machine learning into pricing strategies offers for businesses from an
academic perspective and applied examples and will provide information
about the latest trends and best practices in this field. In this study;
Mathematics, economics, econometrics, and machine learning information
were used together. In this respect; It is an enjoyable and fully
interdisciplinary study.
This curve states that as the price of the product increases in the first period,
the seller's income will also increase, but when the price continues to
increase after a certain optimum level, customers will leave this seller and
therefore the seller's profit will decrease (Clarkson and Black, 2016).
Laffer Curve: The relationship between tax rates and the tax revenue that
the state can collect (This view was first expressed by Ibn Khaldun (1332–
1406))
If we want to write this in the econometric form that we will use in Machine
Learning analysis;
We will use mathematics to find the optimum price level in Equation (3). We
know from mathematics that at the vertices of the curves the first-order
derivative is equal to zero (i.e. the slope of the tangent at this point is zero):
This P value obtained is the optimum price level that maximizes revenue.
Equation (3) can also be expanded with other explanatory variables that are
deemed useful to include in the model. These can be briefly indicated by X;
When the partial derivative of Equation (7) is taken concerning P and set
equal to zero, the resulting P ( the P value in Equation (6) ) will be the
optimum price level. That is, at this price level, the seller's profit will reach
its maximum level. If the seller wants to increase prices even further, he will
start to lose customers and his profits will tend to decrease.
In this context, when Equation (7) was expanded with e-commerce data of a
product of the relevant company, Equation (8) was obtained:
Here:
R: Revenue is the amount of income or profit that the seller will earn.
BSRSub: The best sales rank in the subcategory of the product (Best Seller
Rank)
The profit (R) used in Equation (8) can be calculated with the help of
Equation (9):
Here;
C = The purchase price of the product from the USA (from Amazon.com) +
The handling fee that the intermediate warehouse will receive for changing
the label of the product and sending it to the final buyer in Canada + The
average 15% service fee that Amazon will receive (Amazon Fee )..……….(11)
An important point to note here is; This is because Amazon Fee is calculated
on the final sales price, including the seller's profit.
In this study, Profit (R) was calculated using Equation (10) and Equation (11).
Note: When the data were used as P and R, since an inverted U-shaped
relationship could not be detected between these two variables, the P
component of the analysis was transformed into Q*P, or simply QP. Thus, the
turnover (total money entering the cash register) is included on the
horizontal axis, not the price. In this case, equation (8) was rearranged to
obtain Equation (12):
After Equation (12) was estimated, the partial derivative was taken over the
QP variable.
The QP obtained here is; It gives the optimum turnover, and when this value
is divided by the number of products, the optimum price per unit product
will be reached.
Since this study was conducted on a product basis and in time series form,
Equation (15) was obtained when Equation (12) was written in the form of a
time series econometric model:
Equation (15) was used in the Machine Learning phase.
In the first stage of Machine Learning, to transform the data into similar
sizes and ensure that the model gives equal importance to these variables
(otherwise, machine learning algorithms may tend to give more importance
to features containing large numbers) and thus prevent the results from
being biased. The scaling process has been applied. In this context; the
standard scaler method was preferred. Standard Scaler; It is a method
applied to prevent the ML model from giving more importance to series
containing large values by rearranging the values of the series so that the
mean is 0 and the standard deviation is 1. As a result of this process; All data
used in the analysis become of similar numerical size.
Note: An important point to remember here is; If the series are subjected to
scaling from the beginning, when interpreting the findings (coefficients)
obtained as a result of the analysis, the data must be converted to their
original state by performing a reverse scaling process. Another important
point is that if predictions are to be made with these models (that is if
predictions are to be made for observations), new data must also be
subjected to the same scaling process (or even if we go into more detail, they
must be subjected to scaling that is trained with train data).
First, the prediction was made with the Multiple Linear Regression
model, R2 = 0.47 for the train data and R2 = 0.38 for the test data. Since
the relatively large difference between these two values suggests the risk
of overfitting, a 10-fold cross-validation was performed and it was
observed that R2 decreased to 0.41. In this case, since the difference
between the R2 values of the train and test data decreased, it was decided
that there was no overfitting problem in the model. However, since the
success (R2) of the model was low, it was decided that other Machine
Learning algorithms should also be used.
In the third stage; Three of the ML algorithms in the second stage ( XGB,
SVM, and ADA ) were rearranged to produce coefficients, and predictions
were made with these methods. To make these predictions,
booster='gblinear' in XGBoost, kernel='linear' in SVR, and
base_estimator=LinearRegression() hyperparameters in AdaBoost were
used. Hyperparameter optimization (fine-tuning) operations for these 3
methods were carried out with the Grid Search CV method. As a result of
these analyses, it was seen that the algorithm that gave the best results
was AdaBoost.
For this reason, AdaBoost estimation of the Final Model was made using
all data (without separating it as train and test), this model was trained
(fit) with all data, and the obtained result was reported. ( AdaBoost is a
tree-based algorithm that makes predictions on more than one tree. To
obtain the general result of the model (B1, B2 parameters); the
coefficients obtained in each tree were averaged ).
In the study, df.describe().T code was used to find the standard deviation and
average of the QP value, that is, descriptive statistics were used. By
substituting the standard deviation and mean values of QP taken from
descriptive statistics into Equation (16), the real QP value was found to
Open in app
be
Sign up Sign in
6669.16.
Search Write
IN CONCLUSION
Although price, which is the invisible hand that regulates the markets
according to Adam Smith (Date: 1790, Scotland), is at the center of economic
activities, a simple and clear explanation of how to do price optimization
could not be found in the research. In particular, a significant deficiency has
been identified in how price optimization can be done using Machine
Learning algorithms. With this study, an attempt was made to close the gap
in question to some extent. Researchers who follow this study step by step
can easily develop Machine Learning algorithms that calculate the optimum
price level for their own data sets and obtain the optimum price with a more
realistic (more scientific) method. Hoping it will be useful…
Thanks
OneAMZ Data Scientist Elif Canduz, for her contributions in compiling the data of
this study, Techpro Education Machine Learning instructor Mustafa Erdogan,
who provided the opportunity to discuss/test the accuracy of the established
Machine Learning models, Techpro Education assistant instructor Mert Urper, for
his contributions in the analysis and deployment of the model in Stramlit. I would
like to thank Seher Gumus for her devoted efforts in depositing the work in
Streamlit.
References
Clarkson, A. and Black, A. (2016). Price Optimisation: A Case Study of New
Pricing Techniques.
https://www.actuaries.org.uk/system/files/field/document/G2%20ClarksonBl
ack.pdf
Ruparelia, S., Pezzuli, S. and Looft, M. (2022). Price Optimisation Issues &
Challenges. The Actuarial Profession.
https://www.actuaries.org.uk/system/files/documents/pdf/a02ruparelia.pdf
Econometrics
Written by Ismetgocer Follow
7 min read · Oct 27, 2023 7 min read · Apr 14, 2023
55 94 1
Dr. Fatih Hattatoglu in Academy Team Ismetgocer in Academy Team
14 2
87 71
Lists
526 9 1K 7
Jakob Runge in Causality in Data Science Tahera Firdose
169 2 126