Behavioral Research in the Wild
On Amir
Rady School of Management
UC San Diego
Recent research…
Consumer Decision Making
Behavioral Change
Methods
Behavior in the wild
Consumer Decision Making
Stuck in a Rut: The Behavioral Entrenchment Effect (with Allie Lieberman and Ziv Carmon) –2nd round - OBHDP
A Threshold Value Theory of Sequential Choice (with Coby Morvinski and Evan Weingarten) – 2nd round JMR
Weingarten, et al. (2020). Experts Outperform Technology in Creative Markets, She Ji: The Journal of Design, Economics, and Innovation.
Kristen E. Duke & On Amir (2019). Guilt Dynamics: Consequences of Temporally Separating Decisions and Actions, JCR, Ferber Award.
How Product Unavailability Leads to Choice of Lesser Alternatives—The Role of Hedonic Aggregation (With Jean Zhang and Gal Smitizsky)
Reflecting on the Reflection Effect: The Role of Probability Representation in Risky Choice (with Kristen Duke and Daniel Mochon)
Kristen Duke, Kelly Goldsmith, & On Amir (2018). Is the Preference for Certainty Always so Certain? JACR
Coby Morvinski & On Amir (2018). Liking Goes with Liking: An Intuitive Congruence between Preference and Prominence, JEP:LMC.
Daniella Kupor, Wendy Liu, & On Amir (2017). The Effect of an Interruption on Risk Decisions, JCR
Coby Morvinski, On Amir, & Eitan Muller (2017). “Ten Million Readers Can’t Be Wrong!,” or Can They? On the Role of Information About Adoption Stock in New Product Trial, Marketing Science.
Behavioral change
Aharon, et al.(2022). Improving Cardiac Rehabilitation Patient Adherence via Personalized Interventions, PLOS-ONE.
Alicea Lieberman, Juliana Schroeder, & On Amir (2022). A Voice Inside My Head: The Psychological and Behavioral Consequences of Auditory Technologies, OBHDP.
Coby Morvinski, Silvia Saccardo, & On Amir (2022). Mis-Nudging Morality, Management Science, 1-11.
Alicea Lieberman, Andera Morales, & On Amir (2021). Tangential Immersion: Increasing Consumer Persistence, JCR.
Elanor Williams, Allie Lieberman, & On Amir (2021). Perspective Neglect: Inadequate Perspective Taking Limits Coordination, JDM.
Shuval, et al. (2017). Physical Activity Counseling in Primary Care: Insights from Public Health & Behavioral Economics, CA: A Cancer Journal for Clinicians.
Alicea Lieberman , Kristen Duke, & On Amir (2019). How Incentive Framing Can Harness the Power of Norms, OBHDP.
Goal Proximity, Social Information, and Giving: When Norms Backfire (with Coby Morvinski and Matt Lupoli) 3rd round, PLOS-One
The Impact of Linguistic Structure on Judgment Confidence and Source Attitude (with Giulia Maimone and Uma Karmarkar) – Under review at JMR
Skeptical but Still Indebted: Understanding How Reciprocity is Immune to Skepticism (with Freeman Wu and Andrea Morales)
Methods
What Methods Succeed at the Journal of Consumer Research? (with Andrea Morales and Evan Weingarten) – 2nd round JCR
Weingarten, et al. (2022). What Makes People Happy? Decoupling the Experiential-Material Continuum, JCP
Thomadsen, et al. (2018). How Context Affects Choice, Customer Needs and Solutions.
Andrea Morales, On Amir, & Leonard Lee (2017). A Tutorial in Consumer Research: Keeping it Real in Experimental Research – Understanding When, Where, and How to Enhance Realism and Measure Consumer Behavior, JCR.
In the wild…
Kristen Duke & On Amir (2022). The Importance of Selling Formats: When Integrating Purchase and Quantity Decisions Increases Sales, Marketing Science
Dominance in the Wild (with Ariel Fridman and Karsten Hansen) – 2nd round JCR
Consumer Coping: A Stage Model of Pandemic Response (with Jean Zhang) – 3rd round JACR
Price Contrast in the Wild (with Ariel Fridman and Karsten Hansen)
Consideration Sets in the Wild (with Ariel Fridman, Karsten Hansen, and Wendy Liu)
An unexpected bias: High levels of achievement change the biases women face at work (with Jean Zhang and Elizabeth Campbell)
Communication Style as an Impediment to Trade (with Gal Smitizsky and Kaidi Wu)
Tipping in a Digital Services Marketplace (with Seung Hyun Kim and Kenneth C. Wilbur)
Topics
The context of search results assortment
Dominance
Price contrast
Communication styles and global trade
The dark side of discrimination awareness
(Tipping online)
Dominance in the wild
(with Ariel Fridman and Karsten Hansen, JCR 2023)
(weak) Dominance
But… Frederick et al. (2014), Yang and Lynn (2014)
Dominating
Dominated
Research Questions
Is the dominance effect real?
(beyond the lab and 2-3 item choice sets; preference uncertainty)
Are dominance relationships prevalent?
(why would there be dominated items?)
How does the effect work?
Perceptual vs. Cognitive
Dominance Statuses
Dominating
(Weakly) Dominated
93% of assortments have at least one dominance relationship
No Dominance
64% of assortments
Dominated
6% of assortments
Dominating
29% of assortments
0.7%
0%
1.0%
40% increase
Experienceability of Categories – preference uncertainty
Experienceable
Non-Experienceable
Pre-test: Experienceability and Preference Uncertainty
DV: How certain are you that the choice you selected is the right one for you?
(0 = not at all certain, 10 = completely certain)
Pre-test: Experienceability and Preference Uncertainty
N = 347
DV: How certain are you that the choice you selected is the right one for you?
(0 = not at all certain, 10 = completely certain)
P = .036
Data
Baseline Regression Equation
Purchase ~
Dominance Status x Experienceable +
Fixed Effects:
Gig ID
Country
Date
Position
Controls: Median price, SD price, N gigs at same price, buyer type, etc.
Baseline Regression Equation
Controls: Median price, SD price, N gigs at same price, buyer type, etc.
Purchase ~
Dominance Status x Experienceable +
Fixed Effects:
Gig ID
Country
Date
Position
Neither dominated nor dominating
N = 51.6 million
Customers = 1.2 million
Dominance Effect
N = 51.6 million
Customers = 1.2 million
N = 51.6 million
Customers = 1.2 million
Magnitude of Dominance
>>
≳
Magnitude of Dominance
>>
≳
Perceptual
Simple, automatic processing
(Simonson 1989; Pocheptsova et al. 2009)
Similarity
Similarity drives cognitive assessments – pairwise comparisons
(Hamilton et al. 2007)
Magnitude of Dominance
Magnitude of Dominance
Magnitude of Dominance
Visual Distance & Count of Dominated Gigs
Robustness Checks
Linear probability models
Category-level models
Dominance on price
Clicks
Non-WEIRD populations
Lab experiment
Research Questions
Is the dominance effect real?
YES
Are dominance relationships prevalent?
YES
How does the effect work?
Perceptual vs. Cognitive
Price contrast in the wild
(with Ariel Fridman and Karsten Hansen)
95th percentile
Research Questions
Contrast or Assimilation?
Past research predicts contrast, but positive or negative?
Do the behavioral patterns shed light on the underlying psychology?
Trust/confidence vs. reference point shift?
(Affective/Holistic)
(Range-Frequency)
$5
$5
$5
$5
$10
$20
$10
$10
$20
$25
$30
$X
224 times
$5
$5
$5
$5
$10
$20
$10
$10
$20
$25
$30
$30
-
$150
147 times
2.7%
$5
$5
$5
$5
$10
$20
$10
$10
$20
$25
$30
$170
-
$875
77 times
0%
Data
Baseline Regression Equation
Purchase from assortment ~
Extreme Condition x Experienceable +
Fixed Effects:
Price Distribution (prices 1-11) x Category
Date
All data
Purchase from assortment ~
Extreme Condition x Experienceable +
Fixed Effects:
Price Distribution (prices 1-11) x Category
Date
All data
Clean data
Buyer experience
Buyer experience
Buyer experience
Extremeness
Extremeness
Extremeness
*** (P < .001)
P = .35
Different specifications
Purchase from assortment ~
Extreme Condition x Experienceable +
Median Price x Category
IQR Price x Category
Fixed Effects:
Date
Robustness Checks
Linear probability models
Excluding largest categories (Logo Design and Translation)
Entire dataset (median, IQR controls)
Outcome: purchase on prices 1-11
Non-WEIRD populations
What about average selling price?
*Excludes one outlier at $332 in non-experienceable
ASP�~-4%
LTV: Carryover or persistent effect
Data: all
Outcome:
any future purchase
SEs: Clustered by user
Data: all
SEs: Clustered by user
Outcome:
any future purchase
Data: all
SEs: Clustered by user
Outcome:
any future purchase
*** (P < .001)
P = .19
Data: future purchase
SEs: Clustered by user
Outcome:
log future value
Research Questions
Contrast or Assimilation?
Past research predicts contrast, but positive or negative?
As uncertainty increases – negative!
Do the behavioral patterns shed light on the underlying psychology?
Trust/confidence* vs. reference point shift?
* Heterogeneity
(Affective/Holistic)
(Range-Frequency)
Robustness – 2022 Data
All assortments
Only first 8
~500,000 assortments
“The wild” opens a window to cultural differences and other factors relevant for diversity, equity and inclusion
Communication style
(with Gal Smitizsky)
Communication style theories
Each country gets a score between 0 and 100 for different cultural factors, based on large amount of lab research
Some examples of cultural factors :
Communication style (direct / indirect)
Evaluation habits (direct / indirect)
Decision habits (consensual / top-down)
Trust habits (task / relationship based)
Direct communication
Good communication is precise, simple and clear. Messages are expressed and understood at face value. Repetition is appreciated if it helps clarify the communication.
Responsibility for understanding: The speaker.
Indirect communication
Good communication is sophisticated, nuanced and layered. Messages are both spoken and read between the lines. Messages are often implied but not plainly expressed.
Responsibility for understanding: The listener.
Indirect
Direct
Indirect
Good communication is sophisticated, nuanced and layered
Direct
Good communication is precise, simple and clear.
Sufficient variability
US
Canada
UK
Australia
94
89
67
94
France
Italy
Germany
Portugal
Brazil
Spain
Mexico
Netherlands
35
39
75
44
40
84
43
37
Cancellation rate
Buyer comm. score
AVG cancellation rate
Relational mobility – no significant effect
Buyer comm. score
Conversion rate
Conversion from conversation
Cancellations after Conversion from conversation
����We find support for the lab results even when aggregated by country: �Asking questions facilitates trade�(non-verbal cues absent in digital commerce)��Platforms should accommodate cultural differences
Dark side of discrimination awareness
(with Jean Zhang and Elizabeth Campbell)
Discrimination awareness
Gender discrimination is real
People are aware (especially women)
This drives expectations of successful women up!
Is this a good thing?
Expectancy disconfirmation theory - Revision requests
Buyers can reject a delivery and ask for a revision
The theory suggests this pattern would follow expectations
Carries real economic costs
Do gender-based expectations lead to differential revision requests?
Variable | (1) Logit, FE | (2) Logit, FE + Price |
Male Buyer | -0.043 | -0.041 |
(0.041) | (0.022) | |
Male Seller | 0.048 | 0.064* |
(0.039) | (0.028) | |
Pro | 0.810** | 0.687*** |
(0.256) | (0.197) | |
Extra Revisions Purchased | 0.444** | 0.430*** |
(0.166) | (0.067) | |
Rating Level | 0.010*** | 0.010*** |
(0.001) | (0.002) | |
Number of Past Rated Orders | 0.000 | 0.000 |
(0.000) | (0.000) | |
Male Buyer x Male Seller | -0.029 | -0.042 |
(0.048) | (0.026) | |
Male Buyer x Pro | -0.469* | -0.543* |
(0.238) | (0.236) | |
Male Seller x Pro | -1.147*** | -1.115*** |
(0.327) | (0.290) | |
Male Buyer x Male Seller x Pro | 0.768* | 0.878** |
(0.317) | (0.306) | |
| | |
Log-Likelihood | -236,361.1 | -235,576.5 |
Adj. Psuedo R2 | 0.077783 | 0.080045 |
BIC | 496,046.2 | 495,919.1 |
Squared Cor. | 0.102208 | 0.105509 |
Observations | 397,435 | 397,435 |
Variable | (3) Logit, FE | (4) Logit, FE+Price |
Male Buyer | -0.075*** | -0.075*** |
(0.021) | (0.026) | |
Male Seller | 0.215*** | 0.213*** |
(0.019) | (0.013) | |
Female-Male Sub-Category Ratio | -0.304*** | -0.344*** |
(0.027) | (0.023) | |
Extra Revisions Purchased | 0.556*** | 0.536*** |
(0.031) | (0.029) | |
Rating Level | 0.001*** | 0.001*** |
(0.000) | (0.000) | |
Male Buyer x Male Seller | -0.089*** | -0.110*** |
(0.018) | (0.024) | |
Male Buyer x Female-Male Sub-Category Ratio | 0.019 | 0.025 |
(0.021) | (0.027) | |
Male Seller x Female-Male Sub-Category Ratio | -0.482*** | -0.432*** |
(0.017) | (0.010) | |
Male Buyer x Male Seller x Female-Male Sub-Category Ratio | 0.195*** | 0.120*** |
(0.009) | (0.011) | |
| | |
Log-Likelihood | -256,964.8 | -255,507.4 |
Adj. Psuedo R2 | 0.017241 | 0.021991 |
BIC | 534,524.2 | 533,080.5 |
Squared Cor. | 0.028943 | 0.035466 |
Observations | 404,949 | 404,949 |
Expectancy disconfirmation theory - Revision requests
Buyers can reject a delivery and ask for a revision
The theory suggests this pattern would follow expectations
Carries real economic costs
Do gender-based expectations lead to differential revision requests?
YES
This is where the field started!
Collaboration across research methods
Thank you!
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping in a Digital Services Marketplace
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
University of California, San Diego
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Motivating Question
Why do People Tip?
Why do People Tip in the Gig Economy?
Motivation
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tips are voluntary, post-transaction payments that can motivate service effort, enable price discrimination and help retain talented workers (Lynn 2008)
16% of Americans surveyed have earned money from online gig platforms, 31% of whom said it was their main job within the past year (Pew Research 2021)
Digital services usually feature no face-to-face interaction or observational learning, so tipping practices are opaque.
Digital market design affects consumption and production behaviors, but design effects can be hard to discern, even in thick digital data
Research Questions
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
▶ Describe 341K tips among 4.1MM transactions
▶ Field experiment manipulated the standard tip suggestion: Reciprocity vs Norms
Empirical Context: �Online Marketplace for Freelancer Digital Services
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Pre-treatment Data
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
4.1MM transactions between 1.3MM buyers and 171K sellers
▶ $152MM in total spending, inc. $3.9MM in 341K tips
Period: 01/01/2019 ∼ 06/09/2019 Buyer, seller & transaction characteristics Tipping measures:
▶ Tipping Rate: the proportion of transactions that were tipped
▶ Average Tip Percentage: the average % of price among tips observed
Tipping Rate and Tip Percentages
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate (b) Tip Percentages
Buyers tipped sellers in 8.3% of transactions overall
Among tipped transactions, the average tip observed is 51.8% of the order price
Tipping and Other Factors
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping covaries with Price
▶ High ATP is mainly driven by the low prices
(a) Tipping Rate
(b) Average Tip Percentage
Tipping and Buyer & Seller Region
Observed Fact #1 - Online tipping behavior is associated with regional tipping culture
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate
(b) Average Tip Percentage
North Americans tipped in 11.5% of all transactions ATP varies much less across buyer regions
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping and Buyer & Seller Region
Observed Fact #1 - Online tipping behavior is associated with regional tipping culture
Little or no “Home-region Bias” or “Tourist Effect”
Tipping and Transaction Quality
Observed Fact #2 - Tipping covaries with the quality of the transaction
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate (b) Average Tip Percentage
TR was four times higher among 5-star than 3-star transactions ATP did not change as much with buyer rating
Tipping and Transaction Quality
Observed Fact #2 - Tipping covaries with the quality of the transaction
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate (b) Average Tip Percentage
TR increases with sellers’ average rating, but ATP did not vary
ATP did not vary across sellers’ average rating
Tipping and Buyers’ Characteristics
Observed Fact #3 - Buyer’s previous tipping behavior and previous platform behavior both predict tipping
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate (b) Average Tip Percentage
Buyers’ tipping history predicts tipping
ATP varies much less with buyer prior tips than tipping rate
Tipping and Buyers’ Characteristics
Observed Fact #3 - Buyer’s previous tipping behavior and previous platform behavior both predict tipping
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
(a) Tipping Rate (b) Average Tip Percentage
TR generally increases with buyer tenure up to a peak at 3 years
ATP is nearly flat among newer buyers but then rises with buyer tenure among older cohorts
Tipping and Other Factors
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping covaries with Price
▶ 5% of $5 transactions are tipped, rising to 11% of $30-$40 transactions, and then falling back to 9% of $150+ transactions
▶ ATP is nearly constant at the 20% default for prices above $25
Tipping covaries with Seller Tip Mention
▶ Tipping Rate increased with seller tip mentions from 8.1% to 13%
▶ ATP did not change
Tipping covaries with the type of service provided
▶ Tipping Rate was 10.6% in the Graphics Design category, the first-most frequent category
▶ Digital Marketing was the second-most frequent category had a Tipping Rate of just 2.6%
▶ ATP did not change across the service category
Field Experiment Design - Message
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Period: 06/10/2019-07/07/2019
Randomly assign 45,497 repeat buyers and 7,880 new buyers Treatment revealed after buyer rating, then held constant
Control Group: “Would You Like to Leave a Tip to (Sellername)?”
Implicit Reciprocity: “Show your appreciation to your seller by giving a tip” Reciprocity: “Leave (Sellername) a tip to show your appreciation for a job well done” Norm: “It’s customary to leave a tip for the seller’s service”
Field Experiment Design - Default Tip Scale
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Platform simultaneously changed the default tips in the Reciprocity and Norms conditions New default tip scale contained three options rather than two, with custom tip
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Treatment Examples
Pre-test: Testing the anticipated theoretical constructs of messages
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Method
Participants from Amazon Mechanical Turk were randomly assigned to one of eight conditions (N = 1000)
4 x Message (control [status-quo], implicit reciprocity, explicit reciprocity, and norms) by 2 x Seller name (present vs. not present)
Six reasons presented followed Cialdini’s six “weapons of influence” (Reciprocity/ Social Proof/ Commitment/ Authority/ Liking/ Scarcity)
All participants were asked to mark all explanations that fit how they interpreted each tip message.
Pre-test: Testing the anticipated theoretical constructs of messages
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Control (Status Quo) message motivates multiple interpretations, and no distinct one emerges Participants perceived the three treatment messages as intended
Pre-test: Testing the anticipated theoretical constructs of messages
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Prevalent perception of all messages does not change regardless of the inclusion of the seller’s
name
Does the language of the request for a tip make a difference?
New buyers vs. repeat buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Nonparametric Treatment Effect on New Buyers
Specification for New Buyers
Fixed Effect Regression (pooled)
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Specification for New Buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Fixed Effects Regression
Estimate treatment effects separately
New Buyers Tipping Regression
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Pool (1)
First (2)
Second (3)
Third (4)
Implicit Reciprocity
−0.00001
(0.006)
0.006
(0.006)
0.027∗∗∗ (0.006)
0.257∗∗∗ (0.023)
(0.009)
0.003
(0.009)
Reciprocity
Norms
−0.009
0.010 −0.017
(0.011)
0.006
(0.011)
0.069∗∗∗ (0.012)
0.010
(0.012)
0.013
(0.012)
0.016
(0.011)
0.361∗∗∗ (0.037)
(0.009)
0.318∗∗∗ (0.030)
0.003 −0.030
(0.041)
I previous tip
I previous tip:Implicit Reciprocity
−0.021
(0.033)
(0.053)
I previous tip:Reciprocity
−0.033
−0.036
−0.032
(0.032)
0.055
(0.031)
(0.041)
0.111∗∗∗ (0.039)
(0.052)
0.025
(0.050)
I previous tip:Norms
Buyer Country Seller Country # of Buyers
Observations R2
Adjusted R2
Y Y
7,880
26,436
0.172
0.162
Y Y
7,880
7,880
0.102
0.070
Y Y
7,258
7,258
0.264
0.235
Y Y
3,545
3,545
0.321
0.270
Note:
∗∗p<0.05; ∗∗∗p<0.01
Limitation
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Message treatment effect might be confounded with scale treatment effect
To address this concern, we use post-treatment period data
The platform back to show the Control messages with the new default tip scale for a few weeks
| Control | Treatment |
Test | Control Message + Three Default | Treatment Message + Four Default |
Post Test | Control Message + Four Default | Control Message + Four Default |
Table: Platform Design across Time
To identify the effect of the new default tip scale and the (reverse-)message effect, we use difference-in-differences with buyer fixed effects, seller fixed effects, and transaction characteristics
Specification for New Buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Specification for New Buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
β1 captures the effect of the new default tip scale among Status-Quo message
βMi captures the (reverse-)treatment effect by comparing the change in Tipping Rate
2
across time within treatment groups
Assume the additive specification (treatment effect = message + default tip scale)
Difference-in-Differences Result
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping Rate
Post
0.026∗∗ (0.012)
Implicit Reciprocity:Post
−0.015
(0.016)
Reciprocity:Post
−0.055∗∗∗
(0.017)
Norms:Post
−0.041∗∗
(0.017)
Observations
# of Buyers R2
Adjusted R2
20,023
2,928
0.786
0.458
Note:
∗∗p<0.05; ∗∗∗p<0.01
Effect of the new default tip scale among Status-Quo message was significantly positive to new buyers
Treated new buyers during test period (i.e., Reciprocity, Norm) were less likely to pay tip after platform gave
Status-Quo message
Both ‘message’ and ‘default tip scale’ design could change online tipping behavior of new buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Nonparametric Treatment Effect on Repeat Buyers
Difference-in-Differences
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Difference-in-Differences
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Estimate treatment effects separately
Repeat Buyer Tipping Regression
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Pool (1)
First (2)
Second (3)
Third (4)
Implicit Reciprocity:POST
−0.001
0.002
(0.004)
0.006
(0.004)
0.020∗∗∗ (0.004)
0.002
(0.003)
−0.006
−0.006
(0.002)
0.001
(0.002)
0.004
(0.002)
(0.004)
0.002
(0.004)
0.005
(0.004)
0.002
(0.003)
(0.005)
Reciprocity:POST
−0.004
(0.005)
Norms:POST
−0.005
(0.005)
0.0002
(0.004)
POST
−0.0002
(0.002)
Buyer FE | Y | Y | Y | Y |
Seller FE | Y | Y | Y | Y |
Category FE | Y | Y | Y | Y |
# of Buyers | 40,823 | 40,823 | 34,871 | 22,226 |
Observations | 773,979 | 640,714 | 584,083 | 466,511 |
R2 | 0.544 | 0.562 | 0.562 | 0.569 |
Adjusted R2 | 0.454 | 0.462 | 0.461 | 0.470 |
Residual Std. Error | 0.211 (df = 646783) | 0.210 (df = 521970) | 0.204 (df = 474938) | 0.195 (df = 378992) |
Note: | | | | ∗∗p<0.05; ∗∗∗p<0.01 |
Limitation
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Like New buyers case, message treatment effect might be confounded with scale treatment effect
To address this concern, we use post-treatment period data
| Control | Treatment |
Test | Control Message + Three Default | Treatment Message + Four Default |
Post Test | Control Message + Four Default | Control Message + Four Default |
Table: Platform Design across Time
To identify the effect of the new default tip scale and the (reverse-)message effect, we use difference-in-differences with buyer FE, seller FE, and transaction characteristics
Specification for Repeat Buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Difference-in-Differences Result
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Tipping Rate
POST
−0.003
(0.002)
Implicit Reciprocity:POST
−0.001
(0.003)
Reciprocity:POST
−0.003
(0.003)
Norms:POST
−0.003
(0.003)
Buyer FE Seller FE Observations
# of Buyers R2
Adjusted R2
Y Y
236,072
26,892
0.660
0.504
Note:
∗∗p<0.05; ∗∗∗p<0.01
Effect of the new default tip scale among Status-Quo message was negative but not significant
Tipping Rate was directionally negative for treated repeat buyers during the test period compared to untreated repeat buyers
The change in platform design is less likely to change market behaviors of repeat buyers
Potential Moderators
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
Study why and how the treatment effect increased tipping by interacting treatment effects with potential moderators related to observed facts
▶ Observed fact #1 showed that buyers from countries with established tipping culture are more likely to tip than those without
▶ This fact indicates that regional tipping norms could motivate online tipping
▶ Test whether the regional tipping culture could moderate the treatment effect
▶ Observed Fact #2 shows that a positive correlation between tipping and service quality
▶ It implies that buyers are more likely to reward sellers for the high-quality service
▶ Test whether the high service quality could moderate the treatment effect
How treatment effects vary with the two moderators for new buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
▶ The interaction between North American dummy and norm treatments are not significant, but positive
▶ The interaction between satisfaction and norm treatment is positive and significant
▶ Satisfaction moderates the effect of norm treatment
How treatment effects vary with the two moderators for repeat buyers
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
▶ The interaction between North American dummy and norm treatments are not significant, but positive
▶ The interaction between North American dummy and both Reciprocity treatments are not significant, and negative
▶ Two-way interaction terms with Norm treatment is not significant, but positive
▶ Two-way interaction terms with both Reciprocity treatments are not significant, and negative
Takeaways
Seung Hyun Kim, On Amir, Kenneth C. Wilbur
▶ Tipping likelihood is more situational than tip percentages
▶ Buyer factors drive tipping; Buyer Region, Buyer Satisfaction and Buyer previous tipping/platform behavior
▶ Yes. An injunctive Norm message increased both new buyers’ and repeat buyers’ tipping
▶ Analysis with post-treatment data shows that both the message’ and the ‘default tip scale’ design could change online tipping behavior of new buyers
▶ The Norms treatment effect increased with buyer satisfaction
▶ Other treatment messages (e.g., [Implicit] Reciprocity) did not change buyer behaviors
Thank you!
Lab Experiment
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
Methods
Choice set of songs
Recruited students from UCSD Behavioral Lab (N = 395 after exclusions)
Conditions: non-experienceable, experienceable
Data, analysis code, research materials: https://osf.io/8dasb/?view_only=973dbd5f9d204316ad0fcc7ffbe8e2ea
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
Song Attributes
A’ makes A dominating, B’ makes B dominating
| Song A’ | Song A | Song B | Song B’ |
price | 95 cents | 90 cents | 70 cents | 80 cents |
average rating (out of 5) | 4.8 | 4.9 | 4.7 | 4.6 |
number of ratings | 96 | 119 | 135 | 82 |
Randomized
Experienceable | | | | |
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
Song Attributes
A’ makes A dominating, B’ makes B dominating
| Song A’ | Song A | Song B | Song B’ |
price | 95 cents | 90 cents | 70 cents | 80 cents |
average rating (out of 5) | 4.8 | 4.9 | 4.7 | 4.6 |
number of ratings | 96 | 119 | 135 | 82 |
Randomized
Experienceable | | | | |
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
Song Attributes
A’ makes A dominating, B’ makes B dominating
| Song A’ | Song A | Song B | Song B’ |
price | 95 cents | 90 cents | 70 cents | 80 cents |
average rating (out of 5) | 4.8 | 4.9 | 4.7 | 4.6 |
number of ratings | 96 | 119 | 135 | 82 |
Randomized
Experienceable | | | | |
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
Results: Choice of Dominating Song
Field Evidence
Discussion
Lab Evidence
Questions
Motivation
P < .001
Discussion
Real and prevalent
Experienceability
Perceptual mechanism
Naturalistic setting
Large assortment
Actionable
Field Evidence
Lab Evidence
Discussion
Questions
Motivation
Clean data
Other Real-World Evidence
Three papers tested for dominance effects in field settings:
Retracted: Li, Sun, and Chen (2019)
Cannot isolate the dominance effect from other context effects
Cannot observe choice set, only market-level data
Lab-in-the-field evidence, small sample
Control Variables
Variable Name | Description |
Set Characteristics
| |
Median price | Median price of gigs in set |
SD price | Standard deviation of the prices of the gigs in set |
Minimum price | Minimum price of gigs in set |
Maximum price | Maximum price of gigs in set |
N gigs at same price | Number of gigs at the same price in the set |
Variable Name | Description |
Buyer Characteristics
| |
Total Orders Lifetime | Count of orders placed by the buyer |
Total Orders Cancellation | Count of orders canceled by the buyer |
Buyer Type | Type of buyer: guest, registered not converted (rnc), first time buyer (ftb), second time buyer (stb), repeat buyer, unknown |
Operating System | Operating system of buyer: Windows, Mac, Linux, Chrome OS, iOS, Anrdoid, other, unknown |
RFM Segmentation | Recency, Frequency and Monetary segmentation of buyer: A, B, C, D, E, OTB, unknown |
Count of Dominated Gigs &
Visual Distance
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Dominance on:
average rating, number of ratings
average rating, number of ratings, price
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Logit
Linear Probability
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Category-level Models
Top 6 Categories account for:
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Purchase
Click
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Advanced
Developing
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Lab Evidence
Discussion
Questions
Field Evidence
Motivation
Position
Matters
Clicks per Set
Conceptual Assortment
Shaded box: Dominated Options
Ideal Experiment
92% of sets contain a dominating gig
Dominated Relationship Only
Dominated Relationship Only
Preference Uncertainty Experienceability Assumption:
A Manipulation Check
Mixed Effects Models
Mirrored results from fixed-effects models
Choices Shares
Dominating option very rarely chosen in non-experienceable
Dominance Effect
Future Work
What about other context effects? E.g. Contrast Effect
Formation of consideration sets and demand
How much does the unstructured data for each gig matter?