Professional Documents
Culture Documents
June 9, 2022
Abstract
Modern digital advertising platforms allow ads to be targeted in a variety of ways, and
generally aim to match the ad being shown with either the user or the content being shown.
In this study, we examine the effect of matching in emotional content of ads and the video
on which the ad is shown on consumers’ engagement with the ad. On the one hand, ads
that are emotionally matched with the content video could lead to greater engagement with
the ad because of the desire for consumers to extend the emotion they are experiencing. On
the other hand, emotional mismatch between the ad and content video can lead to greater
perceptual contrast, thereby drawing more attention to the ad. Additionally, consumers
viewing a video with negative emotions might have a preference for positive emotions in
the advertising to get themselves to a happier state. Thus, whether emotional matching
is more effective in driving ad engagement, and thereby potentially consumers’ evaluation
of, affect towards, and purchase of the advertised good is an empirical question. We study
this question through a field experiment run in collaboration with VDO.AI, a video and ad
serving platform. In this experiment, we manipulate the video/ad combination consumers
see, with variation across the experimental conditions in the emotional content of ads and
videos. We find that in our setting incongruence, where emotional content of videos and ads
are mismatched, leads to greater ad engagement. We find suggestive evidence for attention
being the mechanism through which incongruence leads to better outcomes. Our findings
contribute to the literature on advertising, and provide an important targeting variable for
firms in the AdTech space.
In this paper, we empirically study the effect of matching or congruence between ads and the videos
in which they are placed in terms of their emotional content. Digital video and ad serving platforms
allow for targeted advertising, using a variety of targeting variables. These include consumer
characteristics, the content of the video on which they are placed, and contextual variables such
as time of day. An important distinguishing factor of video advertising, as opposed to forms of
advertising relying on text or even static images, is the potential for rich emotional content in these
ads. Thus, an important potential variable for targeting of ads could be the emotional content of
the content video in which advertising is placed and how well matched it is with the emotional
content of the ad. For this, an understanding of the impact of emotional matching of videos and
ads is important. In this study, we aim to gain such an understanding through a field experiment
where we exogenously vary the degree to which there is such emotional matching, and examine
the impact of such matching on consumers’ engagement with the ad.
There has been a growing literature on the study of matching in persuasive messaging, includ-
ing advertising. In particular, the role of targeting or tailoring or what has been referred to as
’personalized matching’ has received a lot of attention in the marketing and psychology literatures
(see, for instance Teeny et al. 2021 for a recent review). The focus in this literature has typically
been on the match between the advertising message and the recipient of the message. There has
been on the other hand more limited attention paid to ’non-personalized matching’, particularly
in the advertising context. Non-personalized matching refers to the match between the situation
or context of the message and the message itself. Studies in the past have examined matching
between the type of product being written about - hedonic vs. utilitarian - and the type of message
- emotional vs. cognitive - (Rocklage and Fazio, 2020), or the characteristics of the message source
and the message itself (Karmarkar and Tormala, 2010). Additionally, past studies have examined
the effect of matching valence (Coulter, 1998; Kamins et al., 1991), arousal (Puccinelli et al., 2015),
mood (Kamins et al., 1991) and theme (Dahlén et al., 2008) on the effectiveness of ads. A recent
study (Fong, 2021) uses the emotional content of music in videos and ads to indirectly examine
the impact of matching. There is however little direct evidence of the effect of matching between
the emotional content of a video ad and that of the content video in which the ad is placed. We
attempt to fill this gap by directly examining matching in emotional content of ads and the videos
they are placed in. Further, in contrast to previous studies that have examined attitudes of con-
sumers towards the subject of the ad, we focus on actions that consumers can take to engage or
disengage with the ad, including skipping the video altogether. This is of great relevance in the
online video ad context, where consumers often have the option to skip an ad.
We design and implement a large-scale field experiment in collaboration with the video/ad
2 Background
Our empirical setting is in the context of video advertising on online platforms. The most ubiq-
uitous of these is YouTube, where viewers watch content videos and ads are inserted into the
Notes: The figure depicts a sample ”In-stream” and ”Post-roll” video ad placement deployed in the experiment.
3 Experimental Design
Our experiment aims to understand how emotional matching or congruence between an ad and
the content video in which it is placed affects consumers’ engagement with the ad. We designed
a field experiment in collaboration with the video and ad serving platform VDO.AI, where we
exogenously manipulated the content video and the ad that the consumer could view. This design
allowed us to both examine the main question of the role of matching/congruence and explore the
underlying mechanisms by which it affects behavior.
The experiment was conducted across a large number of publishers on the internet, with a focus
on consumers in India or those viewing India-related content in other parts of the world. This
focus on Indian consumers was because emotions in videos could relate to the cultural context of
the viewer. The focus on Indian consumers allowed us to reduce the disparities in cultures that
could cause ambiguity of the emotions felt by users for a given video. For identifying the videos
and ads to be included in the experiment, we used Indian research assistants to score videos on
the emotions. Thus, the identification of videos, as well as the consumers viewing it were Indian
or India-related.
Users who arrived at a website included in the experiment were randomly assigned to one of
four conditions in a two by two design - two conditions for the content video - happy or sad,
crossed with two conditions for the ad video - again happy or sad. Thus, they were shown either a
happy content video from a pre-selected set of videos, or a sad video from a pre-selected set. After
viewing the video, they were shown ads from one of two ad conditions - happy or sad, once again
from a pre-selected set of happy or sad videos.
Figure 2 shows a schematic of the experimental design.
First, the curation team from our partner VDO.AI selected a number of videos and ads that they
classified as happy or sad. This was a manual process, where the team had access to a larger set of
videos from which they shortlisted a set of videos as ranging in emotions from happy to sad. This
selection process focused on finding videos and ads that were clear in their emotional content, and
typically had only one dominant emotion running through the entire course of the video.
Then, we recruited a set of 7 research assistants to view all these videos and rate them on their
emotional content on a 7-point Likert scale ranging from happy (3) to sad (-3). These research
assistants were also recruited in India, so as to be culturally aligned with the consumers who would
be viewing the ads in the field experiment. We then selected a set of videos that were at least 2
or above on in terms of their average ratings across the 7 research assistants, or below -2 (thus,
selecting the videos that were more extreme in their emotional content on either side of the scale -
happy or sad). This reduced our set of videos to a set of 24 content videos and 33 ad videos. Table
1 lists the number of unique ads and content videos in the experiment by experimental condition.
Happy Sad
Content Video 11 13
Ad 20 13
Table 2 reports 3 particular coefficients of inter-rater agreement for more than 2 raters specifi-
cally adapted for cardinal data (the measures for Ad and Video Emotions are cardinal in nature).
As we can see, Krippendorff’s Alpha is 0.839 for the videos and 0.792 for the first quartile and
0.694 for the videos and 0.619 for the second quartile which is an acceptable score (Krippendorff,
2018). Similarly, Kendall’s W values are at a reasonable level.
4 Empirical Analysis
4.1 Data
We first describe the data, conduct randomization checks and discuss our overall empirical strategy.
The data were collected at the level of an impression. The impression recorded the unique user-ID
along with the IP address the user was accessing the website from. A total of about 25 million
impressions were included in the experiment. They were then randomly assigned to be shown
one of the 24 content videos, or one of the 33 ads. Since the content videos and ads were not
equal across conditions, and it was easiest for our partner platform to implement a design where
one of these 24 content videos or 33 ads were shown at random, our final sample was not equally
distributed across the four different conditions. Table 3 below shows the distribution of the ad
impressions across each experimental condition.
Notes: The Table above shows the count of impressions across the four experiment sub-conditions, i.e. Happy ad
& Happy video, Happy ad & Sad video, Sad ad & Sad video, Sad ad & Sad video.
We next report a set of randomization checks in Section B of the Appendix. In this section,
we show that randomization is induced properly. Specifically, we verify that the users across
experimental conditions have no systematic differences in the probability of being shown the videos
they are eligible to be shown, that their probability of being shown a happy or sad video does not
systematically vary across the days of week, hours of day, device type they are using, etc.
For each impression, we study four main outcomes based on whether the user viewed: (1) the
first quartile of the duration of the advertisement, (2) the midpoint of the duration of the adver-
tisement, (3) the third quartile of the duration of the advertisement or (4) the full ad. All four
outcome measures [(1) - (4)] are behavioral measures that are important indicators of the degree of
engagement of the consumer with the ad, and are important to the video platform 4 . Table 5 lists
the raw data for the outcome measures across Happy and Sad Videos, Table 4 lists the outcome
measures across Happy and Sad Videos and finally, Table 6 lists the outcome measures across the
four experiment sub-conditions, i.e. Happy ad & Happy video, Happy ad & Sad video, Sad ad &
Sad video, Sad ad & Sad video.
Notes: The Table above lists the means and standard deviations of the outcome measures across Happy and Sad
Ads.
4
This is related to the question of what constitutes relevant “actions.” In general, major advertising platforms
like Facebook and Google list the amount of ad viewed as an important performance metric.
10
Notes: The Table above lists the means and standard deviations of the outcome measures across Happy and Sad
videos.
Notes: The Table above shows the performance of outcome variables across the four experiment sub-conditions, i.e.
Happy ad & Happy video, Happy ad & Sad video, Sad ad & Sad video, Sad ad & Sad video.
We first run a series of regressions to explore the relationship between the matching/congruence
between content video and ad, and the ad watching behavior. Let WatchBehaviori denote a binary
0/1 indicator of the ad watch outcome for exposure i - these denote the four different outcomes
we examine - watching the first quartile, midpoint, third quartile and the complete ad. Thus, if
the user corresponding to impression i watched the ad until the first quartile, the corresponding
W atchB ehaviori is 1, otherwise it is 0. Let IncongruenceConditioni be a dummy variable indicating
whether the experimental conditions correspond to incongruence between the content video and ad
(where a happy ad is played after a sad video or vice-versa). To examine the effect of incongruence
on these outcomes, we run the following regressions, one for each of the four outcome variables.
WatchBehaviori = α + β1 IncongruenceConditioni + ϵi
Here, β captures the effect of the incongruence between the content video and the ad on user’s
11
Notes: The dependent variable is completion till first quartile, completion till second quartile, completion till third
quartile, completion till full. The “OLS” columns show the estimates of a linear probability model.
The main results of our experiment are that incongruence between content videos and ads
leads to greater viewing of the ads across all four outcome variables. The baseline viewership of
the first quartile of the ad is about 81.8%, and this goes up by .39% for the incongruent conditions.
Thus, when the ad and content videos are mismatched, this viewership goes up to about 82.2%.
Similarly, the viewership of the ad until its midpoint goes up from 72.6% to about 73.1%. For the
third quartile, these numbers are approximately 65.9% and 66.5%, and for the viewership of the
complete video, these are approximately 60.6% and 61.2%.
Of note here is that these are relatively high baseline levels of viewership of the ads compared
to the numbers typically seen for a platform like YouTube. This is because of two reasons. First,
the platform shows these ads on other content-filled web pages - the consumer navigates to these
web pages for this other content and not for the videos themselves. Thus, they could be ’viewing’
the ad, but actually consuming some other content simultaneously. Also, the ads play by default
in the muted state, allowing for such multitasking on the part of the consumer. This is the issue
of ”viewability” that was discussed in the section 2. Thus, only a proportion of those recorded as
viewing the ad actually view it. Second, since our experiment only involves post-roll ads, where
the ad is shown conditional on viewing the content video to completion, the set of consumers spend
a lot of time on the website by design. Since exiting the video is the same as exiting the webpage,,
a larger proportion of them ’view’ the ad to completion than if the consumers were not selected
in this way. While we don’t have a way of directly estimating the proportion of these users that
actively view the ad, the estimates of our project partner, VDO.AI and of the industry itself is
that approximately 5 to 20% of viewers of an ad actively view it.
12
where X = f (user − agent, country, country × T ime, city, day, hour − of − day)
Our main findings are that even with a very detailed set of controls, with fixed effects for every
combination of country and time, browser/device type and city of login, the effects of incongruence
are significant and positive.
13
Time(Hourly) ✓ ✓ ✓
Day ✓ ✓ ✓
User-Agent(Browser, Device) ✓ ✓
Country ✓ ✓
Country x Time ✓ ✓
City ✓
Notes: The dependent variable is completion till first quartile, completion till second quartile, completion till third
quartile, completion till full. The “OLS” columns show the estimates of a linear probability model.
We also report the relative effect sizes in Figure 3. We convert the absolute effect into a
percentage change relative to baseline, computing the robust standard errors of this percentage
change using the delta method. The figure reports the percentage change in viewership across the
four different outcome measures, and the confidence intervals of this percentage change.
14
Notes: The Figure above reports the relative effect sizes. The absolute effect is converted into a percentage change
relative to baseline, computing the robust standard errors of this percentage change using the delta method. The
figure reports the percentage change in viewership across the four different outcome measures, and the confidence
intervals of this percentage change.
Further, in Table 9, rather than classifying the ads and emotions as happy and sad, we use the
emotion scores directly as a continuous measure. The covariate in the regression is the product
of the emotion scores for the content video and the ad. When the content video and ad are
matched, these scores are either both positive (in the case of both being happy) or both negative
(in the case of both being sad), resulting in a positive product. When they are mismatched or
incongruent, this product is negative. The coefficient for this variable would thus indicate the
effect of incongruence. If this coefficient is positive, it would imply that consumers’ viewership
of the ad at the four different time points (first quartile, midpoint, third quartile and complete
video) is greater when the content video and ad are congruent. If it is negative, it would imply
that incongruence leads to greater viewership. The estimate shows the same pattern as earlier –
15
Notes: The dependent variable is completion till first quartile, completion till second quartile, completion till third
quartile, completion till full. The “OLS” columns show the estimates of a linear
We conducted Monte Carlo simulations to determine the statistical power needed to detect the
ad-video congruence effect in our sample. We define the effect size to be the percentage difference
in occurrence of an event (say reached first quartile) between the incongruence and congruence
cohorts. We consider a range of effect sizes(0.1%, 0.25% ,0.5%, 1%) and see if we will be able to
detect the delta in ad consumption behavior due to the ad-video congruence. Expressed simply,
we test if for a given effect size, the experiment and the sample has the ability to reject the null
hypothesis of equality between the congruent and incongruent population. To assess this, we took
3 different sample sizes (100,000 ; 500,000 ; 1,000,000) and for each sample, We simulate power
estimates by bootstrapping. The step-by-step procedure for the simulation exercise is presented in
the next few sentences. As a first step, we bootstrap samples from the experiment for a given effect
size, sample size and condition. Then, we compute estimates for the main regression specification.
Next, we collate the p-values from the statistical tests. The power is finally, the proportion of
these p-values ≤ 0.05.
Figures 4 and 5 plot the statistical power against a range of effect sizes. We do block sampling at
the user-level with replacement by pooling together the data from both the experimental conditions
i.e. congruence and Incongruence for a given effect size x. Specifically, we simulate the change in
the decision of x% of the population. The effect is the difference between % occurrence of an event
under the conditions of incongruence and congruence. In this simulation, we produce this difference
by changing the dummy for the event to 1 in x% of the incongruent samples. Then we test the null:
% occurrence of the event is the same in congruence and incongruence using a regression framework
16
17
We now explore the mechanism behind the main results and present corresponding empirical
evidence to corroborate our theoretical underpinning.
So far, we have documented that incongruence in emotions between the ads and the content video
increases the consumer’s probability of watching the ad up to various points of the video.
The first mechanism we examine for our effects is the attention mechanism (Han and Marois,
2014; Kahneman, 1973), which refers to the fact that consumer attention is higher for more novel
content in a given space. This has been specifically applied to the space of emotions as well
(Belanche et al., 2017; Cornelis et al., 2012) and is consistent with the research in advertising and
eye-tracking where attention is directly measured (Pieters and Warlop, 1999). The implication of
this broad idea in our specific context is that attention to an ad that follows a content video should
be increasing in the degree to which the two differ in terms of their emotional content. This in
turn should increase viewership of the ad as well. In our experiment, we have content videos and
ads with a range of emotion scores, and they were all randomly assigned to consumers. This allows
us to examine empirically how consumers behave differently when the ad and content videos were
emotionally very far apart, versus when they were closer together.
Specifically, the above discussion suggests that the exact levels of emotion scores of the incongru-
ence condition matter (Noseworthy et al., 2014). We expect the ad in the incongruent ad-video pair
to be less novel and attention grabbing if the ad’s emotion score is close to the emotion score of the
video. Thus, as the distance between the emotion scores of the content video and ad increases (and
thereby novelty of the ad increases), the increased attention this generates should lead to an in-
crease in the effect of incongruence. Furthermore, the optimal stimulation theory (Joachimsthaler
and Lastovicka, 1984) would suggest that there is a point beyond which the contrast between the
content video and ad could lead to ad avoidance as the consumer’s experience is less pleasurable
(Westbrook and Oliver, 1991).
Building on these theoretical underpinnings, we present our propositions:
• Proposition 1a: The treatment effect is higher when we compare the incongruent group to
a congruent group where the ad shown in the incongruent ad-video pair has an emotion score
closest to the emotion score of the video content. That is, as we increase the emotion score
difference between ads and videos in the incongruent group, the treatment effect increases.
18
Now, we look for empirical support for the above two propositions. As a first step, we partition
the incongruent group in our data into sub-groups on the basis of the emotion scores of the
respective ads and videos in the incongruent ad-video pair. We split the data into bins with a
bin-spacing of 1.25 (Binning is done on the product of the scores i.e. emotion score of the ad ×
emotion score of the placement video.
As a next step, we individually estimate the causal effect of treatment against each of these
incongruent ad-video pairs. We present the outcomes in Figure 6.
We wish to clarify that the covariate in the regression is not the product. It is still incongruence
(hence the positive effect). What we do is split the data according to the product of the ad and
video, classify it as congruent or incongruent based on the value of this product and then regress
the outcome measures against the incongruence dummy. This would demonstrate that the positive
effect (coefficient) of incongruence changes according to the level of incongruity (the product of
ad and video emotion). The products in the graph on the x axis listed as 4,5,6,7,8,9 are actually
negative.
As expected, we see an initial increasing trend in the causal effect of treatment as the difference
between the emotion scores of the content videos and the ad increases. Therefore, we infer that
when the difference between the advertisement-video emotions is low, the advertisement will draw
less attention from the user. Interestingly, as we further increase the difference between the emotion
of the ad and the video, we observe a decreasing pattern – the treatment effect decreases as the
spacing between the emotions of the ads and videos in the incongruent ad-video pair further
increases.
19
Notes: The Figure shows the plots of the coefficients of the ad-video incongruence against the corresponding prod-
uct of emotion score of the ad × emotion score of the placement video. The top-left panel plots this for the first
quadrant viewership, the top-right panel does this for the midpoint, the bottom-left for the third quadrant and
bottom-right for the completion of the video
Figures 7 and 8 plot the same graphs with a bin size of 1.1 and 1.2 respectively. We see a
consistent U-shaped pattern across all the bin sizes.
20
Notes: The Figure shows the plots of the coefficients of the ad-video incongruence against the corresponding prod-
uct of emotion score of the ad × emotion score of the placement video. The top-left panel plots this for the first
quadrant viewership, the top-right panel does this for the midpoint, the bottom-left for the third quadrant and
bottom-right for the completion of the video
21
Notes: The Figure shows the plots of the coefficients of the ad-video incongruence against the corresponding prod-
uct of emotion score of the ad × emotion score of the placement video. The top-left panel plots this for the first
quadrant viewership, the top-right panel does this for the midpoint, the bottom-left for the third quadrant and
bottom-right for the completion of the video
To further explore the proposition that the users’ behavior under the incongruence treatment effect
is driven by stimulus-novelty and attention, we focus on geography-level features that capture
the user’s exposure to the emotion of the video and the ads: (1) video watching volume – the
inherent tendency of users to consume extreme (sad) emotional videos in that geography, and (2)
ad watching volume - which denotes the inherent tendency of the users to consume extreme (sad)
emotional ads in that geography. These variables capture user’s saturation level in that geography
to an extreme video or ad.
The theoretical mechanism proposed earlier would suggest that it is harder to shift users’
attention with ads having congruent emotional content if they are saturated with more similar
emotion videos or ads. Conversely, when consumers are more saturated with viewing of emotional
content of one kind, incongruent ads should have greater engagement due to their ability to get
the attention of the viewer. Hence, we expect to see higher treatment effects (more viewership in
22
• Proposition 2a: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively higher for users with high consumption of extreme emotion (happy)
video content.
• Proposition 2b: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively higher for users with high consumption of extreme emotion (sad)
video content.
• Proposition 2c: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively lower for users with high consumption of less extreme emotion
(happy) ads.
• Proposition 2d : The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively lower for users with high consumption of extreme emotion (happy)
video content.
• Proposition 2e: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively higher for users with overall low completion rates of video content.
• Proposition 2f : The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively lower for users with overall high completion rates of video content.
We estimate the treatment effects separately for the above propositions. We present the results
in Figure 9. Video completion rates are calculated for each city, followed by a median split
across these completion rates to stratify data into HVC(High Video Completion) and LVC(Low
23
Notes: Heterogeneity in the incongruence effects across video and advertisement preferences aggregated across
geographies (cities). Confidence intervals are built using robust standard errors. HVC refers to the High Video
Completion. LVC refers to the Low Video Completion. HAP refers to Happy ad preference. SAP refers to sad ad
preference. HVP stands for the Happy video preference. SVP stands for the sad ad preference.
We further cut our data across user’s heterogeneous preferences across happy and sad place-
ment video preference (Figure 10) and predict:
• Proposition 3a: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively higher for users who consumer more sad video and especially,
24
• Proposition 3b: The incongruence effect (i.e. preference for incongruent emotion ad-
content pair) is relatively higher for users with higher preference for Happy video and when
happy ad plays after the sad placement video.
Prediction 3(a) stems from our theoretical mechanism which suggests that users in the sad
emotional state are more likely to regulate their emotions and therefore, try to move away from
the sad state and hence, prefer happy ad content. Extrapolating this to our scenario, users who
prefer sad content are more likely to pay attention to happy ad content as they have higher need to
regulate their emotions (by virtue of consuming more sad content) are therefore, happy ad content
is likely to pick her attention. Therefore, this need for regulating their emotions prompts users who
prefer sad content in general to be inconsistent in their choice and they prefer, happy ad content
after consuming sad video content.
On the other hand, our theoretical mechanism supports Prediction 3b as preference for less
extreme emotions like Happiness don’t generate much need for emotion regulation in the users.
Since, the need for regulating their emotions is not high in such users, they stay consistent with their
preferences and prefer happy ads. We see the coefficient for HVP-SH is gretaer than the coefficient
for HVP-HS as the sad placement video does create a small need for the user to regulate their
emotion - Though we don’t expect the two coefficients to be statistically significantly different
from each other. Therefore, with lower need for emotional regulation, users with preference for
happy content stay show a consistent preference for happy ad content.
25
Notes: Heterogeneity in the incongruence effects across video and advertisement preferences aggregated across
geographies (cities). Confidence intervals are built using robust standard errors.
We now explore the differences in the impact of emotion incongruence across contextual factors
like day of the week and device used. Based on the attention as the theoretical mechanism, we
expect the preference for incongruent ad-video pair to be more, in case, the prior attention is
already high. That’s when the increment in attention from the incongruent ad-video pair will add
up to the already high attention level in hand. We formulate the formal propositions below:
• Proposition 4: The treatment effect (greater engagement with the ad when the ad-video
pair has incongruent emotional content) is higher when the contextual factors provide users
with extra attention & novelty apart from the attention & novelty provided by the treatment
itself (serving an ad with a different emotion than the content video). Therefore, a viewer
26
We test this prediction by using a causal forest (Wager and Athey, 2018) model for assessing the
impact of different heterogeneity effects on the main test statistic. The test statistic is the difference
in the fraction of viewers who reach 25% of the ad between the congruence and incongruence
cohorts. We look for heterogeneous treatment effects across the device being used for watching
the content and the day of the week. User consuming content over desktop during the first half
of the week is expected to have less time in hand and hence, likely to be more actively consuming
the content and pay additional attention to the video ad. The causal forest model partitions the
sample into clusters using these heterogeneity effects, maximizing the difference between CATE
(Conditional Average Treatment Effects) value of these clusters. The results are reported in Figures
11 & 12.
The maximal difference obtained between CATE values using this model were
∆(CAT E) = 0.0028
Notes: We generate dummy variables for each of the heterogeneity effects. Thus the following dummy variables
were made:
27
To assess the relative impact of the different heterogeneity effects, we make use of SHAP value
plot for these effects.
Notes: or every dummy variable, the red data points denote the variable set to 1 and blue denote the variable
having value 0. The dummies that have red data points on the positive side tend to increase viewership in the con-
gruence cohort, reducing the CATE. Dummy variables with red points in the negative side reduce viewership in the
congruence cohort, increasing the CATE. Thus we can infer that a viewer watching on a desktop in the first half of
the week shows maximum magnitude of CATE.
Yi is the binary indicator that the user i watched at least 25% of the advertisement; xi is a vector
of covariates (a categorical variable with values desktop and days of the week); wi is the binary
indicator representing whether the ad-video emotions were congruent(0) or incongruent(1). From
the potential outcomes framework, (Yi (1), Yi (0)), represent respectively, the behavioral outcome
of the user i under the treatment and control conditions. We estimate the conditional average
treatment effect function τ (x) = E[Yi (1) - Yi (0) | xi = x]. The CATE function was estimated
using the CausalModel function from the python package dowhy. We then obtain the SHAP values
for the covariates and plot them. The scatter plots for each of the covariaes shows the relative
magnitude and direction of bias that it introduces in the ATE. The CausalModel function also
gives us the maximum value of difference between CATE obtained by partitioning the covariates.
28
In this section, we estimate the characteristics of a user segment that is more likely to be impacted
by incongruent ad-video pairs. We first estimate the conditional average treatment effect (CATE)
on streaming of congruent ad-video pairs. We define a user as relatively “strongly” impacted by
incongruity, if their estimated CATE on congruent ad-video pair is greater in magnitude than the
median estimated CATE in the data, and relatively “weakly” affected otherwise. Next, the ATE
for each sub-group is calculated. Lastly, the averages of the co-variates in compared across the
treatment and control groups (e.g., (Athey and Wager, 2019; Athey et al., 2021; Chernozhukov
et al., 2018)). The CATE is estimated non-parametrically by fitting a causal forest (Athey and
Wager, 2019). In Table 10, we report the individual CATE values for each of the covariates,
confirming the theoretical hypothesis.
Notes: Strong subgroup has users whose estimated CATE on congruent ad-video pair is greater in
magnitude than the median estimated CATE in the data . Robust standard errors in parenthesis.
7 Robustness
The main test statistic here is the difference in the fraction of viewers who reached 25 % of the ad
between the congruence and incongruence cohorts. Though we have an adequate sample size by
virtue of which the standard results based on the Central Limit Theorem work, we seek to verify the
statistical significance of the results with greater confidence. Therefore, we simulate exact p-value
that does not assume normal approximation. We simulate the sampling distribution of the test
statistic assuming the null hypothesis is true. The null hypothesis is that the emotion incongruence
has no effect. We lay out the procedure in the next few sentences. Firstly, we do block sampling at
29
Notes: The Figure plots the empirical CDF of these differences, representing the distribution of
the test statistic under the null.
P-values may prove to be unreliable while testing our null hypothesis. This is where Gelman and
Carlin (2014) concept of Type-III and Type-M errors comes into picture. Type - S error is the
probability that the estimate has the incorrect sign, even if it is statistically significantly different
30
Cohen’s D defines the effect size of a sample by comparing two data-sets. Cohen’s D is defined as:
x1 − x2
d= ∗
ss
(n1 − 1)s21 + (n2 − 1)s22
s∗ =
n1 + n2 − 2
s
n1 + n2 d2
SE ≈ +
n1 n2 2(n1 + n2 )
where d is the estimated effect size, x1 ; x2 are the means of the samples, n1 ; n2 are the sizes of the
two samples, s1 ; s2 are the standard deviations of the two samples and SE represents the standard
error in the effect size estimate.
In our analysis, we first split the data on the basis of cohorts. We remove some outliers with
fewer observations. We only select those cohorts which have larger number of observations and
the number of observations in each cohort is almost the same. The number of observations in each
cohort is made same via bootstrapping. After this, we take every possible pair of cohorts and note
their cohen’s d and standard error. Finally, we take the average of the Cohen’s d and the standard
error. Using the obtained effect size estimate and the standard error, we calculate the Type-S and
Type-M errors.
31
• Number of Simulations that we perform while calculating the error rates = 10,000.
We perform around 10,000 replications of the measure of probability that the incongruence
effect is overstated. In each replication, where the incongruence effect is estimated to be statis-
tically significantly different from zero, the absolute value of the estimated incongruence effect is
divided by A = 0.398 (effect size) to obtain a ratio. The distribution of these ratios across 10,000
replications is plotted in Figure 14 below. The mean ratio is 1.000014 (a low value according to
Gelman and Carlin (2014)
7.2 Are the Effects Driven by Advertisements themselves and not by the incongruity
between the emotions of the ads and content video?
Table 11 reports whether there is support for an alternative hypothesis for the difference in out-
comes between incongruent and congruent groups by positing stronger emotions to the ad itself
and not by virtue of an ad’s emotion being aligned with the emotion of the placement video.
32
Dependent variable:
Notes: The Table above reports outcome measures regressed against Ad emotion. Ad Emotions are categorised into
happy and sad based on their emotion scores. Happy ads act as a reference for the above regression, and coefficients
for sad ads are reported with respect to happy ads. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
7.3 Are incongruent effects simply an artefact of users trying to regulate their emo-
tions?
Prior research suggests that consumption of extreme valence emotions make users look for alle-
viating, repairing, or managing an emotion in the short term, also known as emotion regulation
(Kemp and Kopp, 2011). Therefore, to regulate their emotions, users who consume extreme emo-
tions through placement videos for longer period of time are more likely to opt for an ad with an
opposite valence emotion. This is not what we see in the data. In fact, we see that the incongruent
ad-video pairs perform better for shorter placement videos compared to the longer ad-video pairs.
Table 12 reports the results. Further, Table 25 reports the results for the short length videos,
Table 26 reports the results for the medium length videos and Table 27 reports the results for the
long length videos.
33
Dependent variable:
Notes: The Table above reports results for the different lengths of the placement videos. Re-
gressions where an indicator for treatment (Incongruent)/control (Congruent) is interacted with
a categorical variable for the length of the placement video. The placement videos are further
divided into quartiles, Q1 , Q2 , Q3 & Q4 . Q1 denotes placement videos with length < 60s, Q2 de-
notes placement videos with length = 60s and Q3 denotes placement videos with length > 60s.
The number of videos in each quartile, Q1 , Q2 , Q3 is 8,9 and 10 respectively. Complete congruence
with length ∈ Q1 is taken as the reference group and all coefficients are reported with respect to
it.∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
34
In this study, we present the first field-based direct evidence of the effect of ad-video emotional
incongruence ad-video on user engagement with the ad. Our empirical analysis is based on a
field experiment that we designed and implemented on a video and advertising delivery platform.
The empirical results provide evidence for substantively greater engagement in the incongruent
condition. We find evidence consistent with greater attention to the ad when the ad is emotionally
incongruent or mismatched with the content video. We don’t find evidence for emotional regulation
being the mechanism driving the results.
The limitations of this study include our inability to study behavioral outcomes downstream
of ad engagement, such aa visits to the advertiser’s website or purchase of the advertised product
due to the nature of our partner platform, and the inability of tracking consumers after they have
finished viewing the ad. A second limitation is the nature of the platform, where the video and
ad might be playing without the consumer necessarily engaging with it actively. This leads to an
overestimation of engagement, though not of the treatment effect itself.
35
Athey, S., Bergstrom, K. A., Hadad, V., Jamison, J. C., Ozler, B., Parisotto, L., Sama, J. D.
et al. (2021), Shared decision-making: Can improved counseling increase willingness to pay for
modern contraceptives?, Technical report, The World Bank.
Athey, S. and Wager, S. (2019), ‘Estimating treatment effects with causal forests: An application’,
Observational Studies 5(2), 37–51.
Belanche, D., Flavián, C. and Pérez-Rueda, A. (2017), ‘Understanding interactive online advertis-
ing: Congruence and product involvement in highly and lowly arousing, skippable video ads’,
Journal of Interactive Marketing 37, 75–88.
Biswas, R., Riffe, D. and Zillmann, D. (1994), ‘Mood influence on the appeal of bad news’, Jour-
nalism Quarterly 71(3), 689–696.
Bleier, A. and Eisenbeiss, M. (2015), ‘Personalized online advertising effectiveness: The interplay
of what, when, and where’, Marketing Science 34(5), 669–688.
Bounie, D., Valérie, M. and Quinn, M. (2017), ‘Do you see what i see? ad viewability and the
economics of online advertising’, Ad Viewability and the Economics of Online Advertising (March
1, 2017) .
Chernozhukov, V., Demirer, M., Duflo, E. and Fernandez-Val, I. (2018), Generic machine learning
inference on heterogeneous treatment effects in randomized experiments, with an application to
immunization in india, Technical report, National Bureau of Economic Research.
Cornelis, E., Adams, L. and Cauberghe, V. (2012), ‘The effectiveness of regulatory (in) congruent
ads: The moderating role of an ad’s rational versus emotional tone’, International Journal of
Advertising 31(2), 397–420.
Coulter, K. S. (1998), ‘The effects of affective responses to media context on advertising evalua-
tions’, Journal of Advertising 27(4), 41–51.
Dahlén, M., Rosengren, S., Törn, F. and Öhman, N. (2008), ‘Could placing ads wrong be right?:
advertising effects of thematic incongruence’, Journal of Advertising 37(3), 57–67.
36
Furnham, A., Bergland, J. and Gunter, B. (2002), ‘Memory for television advertisements as a
function of advertisement–programme congruity’, Applied Cognitive Psychology: The Official
Journal of the Society for Applied Research in Memory and Cognition 16(5), 525–545.
Gelman, A. and Carlin, J. (2014), ‘Beyond power calculations: Assessing type s (sign) and type
m (magnitude) errors’, Perspectives on Psychological Science 9(6), 641–651.
Janssens, W., De Pelsmacker, P. and Geuens, M. (2012), ‘Online advertising and congruency
effects: It depends on how you look at it’, International Journal of Advertising 31(3), 579–604.
Kamins, M. A., Marks, L. J. and Skinner, D. (1991), ‘Television commercial evaluation in the con-
text of program induced mood: Congruency versus consistency effects’, Journal of Advertising
20(2), 1–14.
Karmarkar, U. R. and Tormala, Z. L. (2010), ‘Believe me, i have no idea what i’m talking about:
The effects of source certainty on consumer involvement and persuasion’, Journal of Consumer
Research 36(6), 1033–1049.
Kemp, E. and Kopp, S. W. (2011), ‘Emotion regulation consumption: When feeling better is the
aim’, Journal of Consumer Behaviour 10(1), 1–7.
Kirmani, A. and Yi, Y. (1991), ‘The effects of advertising context on consumer responses’, ACR
North American Advances .
Kononova, A. and Yuan, S. (2015), ‘Double-dipping effect? how combining youtube environmen-
tal psas with thematically congruent advertisements in different formats affects memory and
attitudes’, Journal of Interactive Advertising 15(1), 2–15.
37
Kwon, E. S., King, K. W., Nyilasy, G. and Reid, L. N. (2019), ‘Impact of media context on ad-
vertising memory: A meta-analysis of advertising effectiveness’, Journal of Advertising Research
59(1), 99–128.
Lambrecht, A. and Tucker, C. (2013), ‘When does retargeting work? information specificity in
online advertising’, Journal of Marketing research 50(5), 561–576.
Lau-Gesk, L. and Meyers-Levy, J. (2009), ‘Emotional persuasion: When the valence versus the
resource demands of emotions influence consumers’ attitudes’, Journal of Consumer Research
36(4), 585–599.
Lee, C. J., Andrade, E. B. and Palmer, S. E. (2013), ‘Interpersonal relationships and preferences
for mood-congruency in aesthetic experiences’, Journal of Consumer Research 40(2), 382–391.
McGranaghan, M., Liaukonyte, J. and Wilbur, K. C. (2022), ‘How viewer tuning, presence, and
attention respond to ad content and predict brand search lift’, Marketing Science .
Noseworthy, T. J., Di Muro, F. and Murray, K. B. (2014), ‘The role of arousal in congruity-based
product evaluation’, Journal of Consumer Research 41(4), 1108–1126.
Pieters, R. and Warlop, L. (1999), ‘Visual attention during brand choice: The impact of time
pressure and task motivation’, International Journal of research in Marketing 16(1), 1–16.
Puccinelli, N. M., Wilcox, K. and Grewal, D. (2015), ‘Consumers’ response to commercials: when
the energy level in the commercial conflicts with the media context’, Journal of Marketing
79(2), 1–18.
Rocklage, M. D. and Fazio, R. H. (2020), ‘The enhancing versus backfiring effects of positive
emotion in consumer reviews’, Journal of Marketing Research 57(2), 332–352.
Sahni, N. S., Wheeler, S. C. and Chintagunta, P. (2018), ‘Personalization in email marketing: The
role of noninformative advertising content’, Marketing Science 37(2), 236–258.
Tamir, M. (2009), ‘What do people want to feel and why? pleasure and utility in emotion regula-
tion’, Current Directions in Psychological Science 18(2), 101–105.
Tamir, M. (2016), ‘Why do people regulate their emotions? a taxonomy of motives in emotion
regulation’, Personality and social psychology review 20(3), 199–222.
38
Uhl, C., Nabout, N. A. and Miller, K. (2020), ‘How much ad viewability is enough? the effect of
display ad viewability on advertising effectiveness’, arXiv preprint arXiv:2008.12132 .
Wager, S. and Athey, S. (2018), ‘Estimation and inference of heterogeneous treatment effects using
random forests’, Journal of the American Statistical Association 113(523), 1228–1242.
39
Impression Definition
All ad impressions involved in the experiment are preceded by a video that was shown to the
end user. Each impression thus, is characterised by both its own emotion, and the emotion of
the video it followed. Since individual user level events are not logged, we resort to leveraging
IP addresses and user-agents (browser, devices) to reconstruct user actions. There are 15,16,553
unique IP addresses, and 21,22,598 unique combinations of IP addresses and user-agents indicating
that there are few IP addresses associated with more than one device. We define the ”viewer”
of an impression as the combination of user-agent and IP address that has contiguous logged
events associated with a particular video and related impression. For example, in Table [], IP
1.186.120.21 started watching a video (Video-ID: V1) on his iPhone at 7:03:25, which was then
followed by an advertisement(Ad-ID: A1) which the end user only watched till midpoint, since no
further events were logged. Most analysis is based on events logged after the user started watching
the advertisement. It should be noted that not all videos lead to ad impressions, since completing
the video is a prerequisite to the ad being shown. There are a total of 3.5 million ad impressions
while there are more than 12 million video impressions.
Notes: The Table above shows the activity of a particular IP address, IP 1.186.120.21,on his iPhone on 16th Au-
gust, 2021.
Outcome Variables
Our goal is to delineate unique ad impressions from raw data and analyse associated outcomes
for each impression and experiment group. For such analysis, we need a table with each unique
40
1. Event Logs: We have a large number of log files, which have accounts of all events logged by
the firm for a particular time frame.
2. Video and Ad characteristics: Emotion scores for all videos and ads part of the experiment.
B Randomization Checks
Figure 15. Quantile-plot of the Estimated p-values for Occurrence of Advertisements across videos
of Happy and Sad emotions
Notes: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider the null
hypothesis that the proportion of different advertisements shown is same irrespective of the video emotion being
happy or sad. We perform p-test for proportions on all the advertisements and obtain the p values. The figure shows
that the p-values are uniformly distributed between 0 and 1, which is what we expect if the proportion of different
advertisements displayed is not systematically different across the two video emotions. A Kolmogorov-Smirnov test
is unable to reject that the p-values are uniformly distributed (p-value=0.35)
41
Figure 16. Quantile-plot of the Estimated p-values for Occurrence of Advertisements across de-
vices(ks test p-value 0.09)
Notes: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider the null
hypothesis that the proportion of different advertisements shown is same irrespective of the device. We perform
p-test for proportions on all the advertisements and obtain the p values. The figure shows that the p-values
are uniformly distributed between 0 and 1, which is what we expect if the proportion of different advertisements
displayed is not systematically different across the device type. A Kolmogorov-Smirnov test is unable to reject that
the p-values are uniformly distributed (p-value=0.09)
42
Notes: Panel:A: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider
the null hypothesis that the proportion of different advertisements shown is same on Monday. We perform p-test
for proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is not
systematically different on Monday. A Kolmogorov-Smirnov test is unable to reject that the p-values are uniformly
distributed (p-value=0.06)
Panel:B: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider the null
hypothesis that the proportion of different advertisements shown is same on Wednesday. We perform p-test for
proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is
not systematically different on Wednesday. A Kolmogorov-Smirnov test is unable to reject that the p-values are
uniformly distributed (p-value=0.62)
43
Notes: Panel:A: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider
the null hypothesis that the proportion of different advertisements shown is same on Thursday. We perform p-test
for proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is not
systematically different on Monday. A Kolmogorov-Smirnov test is unable to reject that the p-values are uniformly
distributed (p-value=0.11)
Panel:B: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider the
null hypothesis that the proportion of different advertisements shown is same on Friday. We perform p-test for
proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is
not systematically different on Wednesday. A Kolmogorov-Smirnov test is unable to reject that the p-values are
uniformly distributed (p-value=0.13)
44
Notes: Panel:A: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider
the null hypothesis that the proportion of different advertisements shown is same on Thursday. We perform p-test
for proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is not
systematically different on Saturday. A Kolmogorov-Smirnov test is unable to reject that the p-values are uniformly
distributed (p-value=0.91)
Panel:B: The figure plots the estimated p-values against a uniform CDF (45 degree line). We consider the
null hypothesis that the proportion of different advertisements shown is same on Friday. We perform p-test for
proportions on all the advertisements and obtain the p values. The figure shows that the p-values are uniformly
distributed between 0 and 1, which is what we expect if the proportion of different advertisements displayed is not
systematically different on Sunday. A Kolmogorov-Smirnov test is unable to reject that the p-values are uniformly
distributed (p-value=0.27)
Notes: The Table reports on results from several regressions. For each regression, p-values from an F-test to re-
ject H0 : the covariate has no effect on the independent variable are reported. For example, for the first regression,
a very high p-value of 0.802 demonstrates that the null hypothesis cannot be rejected, and hence Ad Emotion is
45
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
46
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
47
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
48
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
49
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
50
Dependent variable:
Notes: The Table reports on results from several regressions. In the first regression, the dependent variable is a
0/1 indicator of if the video was watched till the first quartile. The standard errors are in parentheses next to the
coefficient estimates and are clustered by clustered by user identification for cookied impressions and by fingerprint
identification (unique IP number, OS, browser, language, and cookie-type combination). Additionally, below each
coefficient in square brackets, we report (1) 95% confidence interval of coefficient. In Column(2), the dependent
variable is a 0/1 indicator of if the video was watched till the second quartile. In Column(3), the dependent vari-
able is a 0/1 indicator of if the video was watched till the third quartile. In Column(4), the dependent variable is
a 0/1 indicator of if the video was watched till the fourth quartile. ∗ p<0.1; ∗∗ p<0.05; ∗∗∗ p<0.01
Dependent variable:
51
52
Dependent variable:
Dependent variable:
53
Dependent variable:
54