SlideShare una empresa de Scribd logo
1 de 21
Hands-on experiences with
incentives and mechanism design
www.insemtives.eu 1
Roberta Cuel, University of Trento, IT; Markus
Rohde, University of Siegen, DE and
Germán Toro del Valle, Telefonica I+D, ES
ISWC 2010
How to design effective
incentives/rules
1. Analyze the domain
• What?
• Working environment
• Job descriptions
• Organization (tasks, hierarchy, compensation, social,
communication)
• How?
• Qualitative face-to-face interviews and questionnaires
• Observations with selected individuals
• Quantitative analysis (data collections)
1/30/2015 www.insemtives.eu 2
How to design effective
incentives/rules (2)
2. Identify the preferences and motivations that drive
users
Concentrate on every-day uses for those specific users
3. Formalize the existing reward system
Find yourself in the matrix
4. Design the simplest possible solution that can
effectively support those uses
Translate into a small number of alternative testable
hypothesis
5. Fine tune the rewarding system
1/30/2015 www.insemtives.eu 3
Fine-tuning incentives with
mechanism design: a step by step
procedure
• Mimic situation in the lab
• Set up experiment as close to real life situation as
possible
• Run experiment with volunteer subjects, with
random allocation of subjects
• Test alternative hypothesis regarding effect of
incentive schemes on behavior
• Check differences in outcome
If happy go to next slide,
otherwise re-design hypothesis and run a new trial.
Fine-tuning part II
• Start adding realism components:
– Move to real subjects (field test)
– Move to real tasks (with real subjects)
– Move to real subjects handling real tasks
– Move to real situation (field experiment)
• During the process you:
– Lose control over ability to manipulate variables
– Gain awareness of interaction between variables
• Let’s look at what we are doing with case studies!
Telefonica I+D case study
• Corporate portal
• What is the most obvious incentive from
economic point of view?
• What can we do with a small budget to be
dedicated to incentivize users?
• How do we know which system is the best for
our setting?
1/30/2015 www.insemtives.eu 6
Basic experiment
• Test Two rewarding/incentives systems
• Pay per click:
– 0,03 € per tag added (up to 3 € maximum).
• Winner takes all model:
– The person who adds the higher number of
tags/annotation wins 20€
What would you choose?
(Participation fee – 5 €)
1/30/2015 www.insemtives.eu 7
The experiment (setting)
• 36 students
– Random assignment to the two “treatments”
• Individual task: annotation of images
• Clear set of Instructions
• Training (guided) session to give basic
understanding of annotation tool
• 8 minutes clocked session (time pressure)
• Goal: produce maximum amount of tags in
allotted time on a random set of images
1/30/2015 www.insemtives.eu 8
The lab
1/30/2015 www.insemtives.eu 9
The experiment: screenshots
1/30/2015 www.insemtives.eu 10
1/30/2015 www.insemtives.eu 11
1/30/2015 www.insemtives.eu 12
1/30/2015 www.insemtives.eu 13
1/30/2015 www.insemtives.eu 14
ID N. Tags Reward Tot. €
2 48 1,5 6,5
3 41 1,5 6,5
4 78 2,5 7,5
5 54 2 7
6 51 2 7
7 36 1,5 6,5
8 44 1,5 6,5
9 54 2 7
10 64 2 7
11 41 1,5 6,5
12 63 2 7
13 58 2 7
14 60 2 7
15 25 1 6
16 43 1,5 6,5
17 50 1,5 6,5
18 24 1 6
19 30 1 6
20 37 1,5 6,5
tot. 901 31,5 126,5
av. 47,42 1,66 6,66 15
ID N. Tags Reward Tot. €
2 86 0 5
3 40 0 5
4 68 0 5
5 88 0 5
6 87 0 5
7 65 0 5
8 67 0 5
9 31 0 5
10 62 0 5
11 79 0 5
12 45 0 5
13 96 20 25
14 51 0 5
15 68 0 5
17 73 0 5
18 26 0 5
19 35 0 5
tot. 1067 20 105
av. 62,76 1,18 6,18
Number of tags
Pay per tag
(N=19)
– Total amount of tags: 901
– Max n. of tags: 78
– N. tags (avg.)= 47.42
– € (avg. per person)= 6.66
– € (avg. per tag) =0.1404
– € total = 126,5 € (31,5 €
flexible compensation)
Winner takes all model
(N=17)
– Total amount of tags: 1067
– Max n. of tags: 96
– N. tags (avg.)= 62.76 (32%
increase!)
– € (avg. per person)= 6.18
– € (avg. per tag)=0.098407
– € total = 105 € (20 €
flexible compensation)
1/30/2015 www.insemtives.eu 16
The results
1/30/2015
www.insemtives.eu
17
T-test and F-test are significant
Tags distribution: interface matters!
Pay per tag
• Tag “nature”  24 times
• “snow”  22 times
• “green”  20 times
• …
• 134 tags repeated only 2
times
• 437 unique tags
Winner take all
• Tag “green” 18 times
• “snow”  14 times
• “butterfly”  13 times
• …
• 118 tags repeated only
2 times
• 390 unique tags
1/30/2015 www.insemtives.eu 18
Some biases
• Students are
– Volunteers who are used participating in
experiments
– Strong web users and game players
– Paid to show up
• Quality of the tags
– Quality of tagging has been controlled for: no
obvious ‘mistakes’ or ‘cheating’
1/30/2015 www.insemtives.eu
19
Summary of results & next lab steps
• Basic hypothesis confirmed
• More work needed:
– Effort directed to producing a good (tags) that are
not consumed by users (used to achieve other
goals)  change structure of the game to let users
exploit tagging to achieve results (treasure hunt!)
– Re-run experiment with new structure. Now users
produce tags to get money and to use tags to
perform more tasks)
Next steps: Telefonica I+D
1/30/2015 www.insemtives.eu 21
• Replicate experiment with real users
– Main change 1: task becomes relevant in terms of
practical usefulness for users
– Main change 2: task has social implications
– Main change 3: expectations change dramatically
(workers vs. students  5 Euros to participate???)
• Add realism
– Mimic social structure in the company:
• Run experiment with teammates
• Use real tasks
• Try alternative pay for performance schemes

Más contenido relacionado

Similar a Incentives Experiment Yields 32% Tagging Increase

SemTech2011 - Employee-of-the-Month' Badge Unlocked
SemTech2011 - Employee-of-the-Month' Badge UnlockedSemTech2011 - Employee-of-the-Month' Badge Unlocked
SemTech2011 - Employee-of-the-Month' Badge UnlockedINSEMTIVES project
 
Digital analytics lecture1
Digital analytics lecture1Digital analytics lecture1
Digital analytics lecture1Joni Salminen
 
Introduction to Data Science
Introduction to Data ScienceIntroduction to Data Science
Introduction to Data ScienceLivePerson
 
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation Feedback
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation FeedbackHeuristic Stimuli Generation for Coverage Closure Exploiting Simulation Feedback
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation FeedbackDVClub
 
Machine Learning driven Quantum Optimization for Marketing
Machine Learning driven Quantum Optimization for MarketingMachine Learning driven Quantum Optimization for Marketing
Machine Learning driven Quantum Optimization for MarketingDeep Learning Italia
 
2010 bristol q1_heuristic-stimuli-generation
2010 bristol q1_heuristic-stimuli-generation2010 bristol q1_heuristic-stimuli-generation
2010 bristol q1_heuristic-stimuli-generationObsidian Software
 
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...Masanori HIRANO
 
Classifying intangible social innovation concepts using machine learning and ...
Classifying intangible social innovation concepts using machine learning and ...Classifying intangible social innovation concepts using machine learning and ...
Classifying intangible social innovation concepts using machine learning and ...Nikola Milosevic
 
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)DevGAMM Conference
 
Danske Bank HELP Universe
Danske Bank HELP UniverseDanske Bank HELP Universe
Danske Bank HELP UniverseIntranätverk
 
Planning and running usability tests
Planning and running usability testsPlanning and running usability tests
Planning and running usability testsChris Collingridge
 
How to Use Data for Product Decisions by YouTube Product Manager
How to Use Data for Product Decisions by YouTube Product ManagerHow to Use Data for Product Decisions by YouTube Product Manager
How to Use Data for Product Decisions by YouTube Product ManagerProduct School
 
Are your analytic tools really adding value?
Are your analytic tools really adding value?Are your analytic tools really adding value?
Are your analytic tools really adding value?Debby Sieradzki
 
Data Science in Online Marketing
Data Science in Online MarketingData Science in Online Marketing
Data Science in Online MarketingMarkus Ojala
 

Similar a Incentives Experiment Yields 32% Tagging Increase (20)

SemTech2011 - Employee-of-the-Month' Badge Unlocked
SemTech2011 - Employee-of-the-Month' Badge UnlockedSemTech2011 - Employee-of-the-Month' Badge Unlocked
SemTech2011 - Employee-of-the-Month' Badge Unlocked
 
Digital analytics lecture1
Digital analytics lecture1Digital analytics lecture1
Digital analytics lecture1
 
Introduction to Data Science
Introduction to Data ScienceIntroduction to Data Science
Introduction to Data Science
 
Tagauchi method
Tagauchi methodTagauchi method
Tagauchi method
 
Accounting Fundamentals
Accounting FundamentalsAccounting Fundamentals
Accounting Fundamentals
 
Quantitative Aptitude 7 October
Quantitative Aptitude 7 OctoberQuantitative Aptitude 7 October
Quantitative Aptitude 7 October
 
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation Feedback
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation FeedbackHeuristic Stimuli Generation for Coverage Closure Exploiting Simulation Feedback
Heuristic Stimuli Generation for Coverage Closure Exploiting Simulation Feedback
 
Machine Learning driven Quantum Optimization for Marketing
Machine Learning driven Quantum Optimization for MarketingMachine Learning driven Quantum Optimization for Marketing
Machine Learning driven Quantum Optimization for Marketing
 
2010 bristol q1_heuristic-stimuli-generation
2010 bristol q1_heuristic-stimuli-generation2010 bristol q1_heuristic-stimuli-generation
2010 bristol q1_heuristic-stimuli-generation
 
When no clicks are good news
When no clicks are good newsWhen no clicks are good news
When no clicks are good news
 
When no clicks are good news
When no clicks are good newsWhen no clicks are good news
When no clicks are good news
 
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...
2022/10/30 BESC2022: Parameter Tuning Method for Multi-agent Simulation using...
 
Classifying intangible social innovation concepts using machine learning and ...
Classifying intangible social innovation concepts using machine learning and ...Classifying intangible social innovation concepts using machine learning and ...
Classifying intangible social innovation concepts using machine learning and ...
 
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)
ML game metrics monitoring system launch / Aleksandr Tolmachev (Xsolla)
 
Danske Bank HELP Universe
Danske Bank HELP UniverseDanske Bank HELP Universe
Danske Bank HELP Universe
 
Planning and running usability tests
Planning and running usability testsPlanning and running usability tests
Planning and running usability tests
 
EDI - European Data Incubator - Public Presentation
EDI - European Data Incubator - Public PresentationEDI - European Data Incubator - Public Presentation
EDI - European Data Incubator - Public Presentation
 
How to Use Data for Product Decisions by YouTube Product Manager
How to Use Data for Product Decisions by YouTube Product ManagerHow to Use Data for Product Decisions by YouTube Product Manager
How to Use Data for Product Decisions by YouTube Product Manager
 
Are your analytic tools really adding value?
Are your analytic tools really adding value?Are your analytic tools really adding value?
Are your analytic tools really adding value?
 
Data Science in Online Marketing
Data Science in Online MarketingData Science in Online Marketing
Data Science in Online Marketing
 

Último

WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsMiki Katsuragi
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfSeasiaInfotech2
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii SoldatenkoFwdays
 

Último (20)

WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering Tips
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdf
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko
 

Incentives Experiment Yields 32% Tagging Increase

  • 1. Hands-on experiences with incentives and mechanism design www.insemtives.eu 1 Roberta Cuel, University of Trento, IT; Markus Rohde, University of Siegen, DE and Germán Toro del Valle, Telefonica I+D, ES ISWC 2010
  • 2. How to design effective incentives/rules 1. Analyze the domain • What? • Working environment • Job descriptions • Organization (tasks, hierarchy, compensation, social, communication) • How? • Qualitative face-to-face interviews and questionnaires • Observations with selected individuals • Quantitative analysis (data collections) 1/30/2015 www.insemtives.eu 2
  • 3. How to design effective incentives/rules (2) 2. Identify the preferences and motivations that drive users Concentrate on every-day uses for those specific users 3. Formalize the existing reward system Find yourself in the matrix 4. Design the simplest possible solution that can effectively support those uses Translate into a small number of alternative testable hypothesis 5. Fine tune the rewarding system 1/30/2015 www.insemtives.eu 3
  • 4. Fine-tuning incentives with mechanism design: a step by step procedure • Mimic situation in the lab • Set up experiment as close to real life situation as possible • Run experiment with volunteer subjects, with random allocation of subjects • Test alternative hypothesis regarding effect of incentive schemes on behavior • Check differences in outcome If happy go to next slide, otherwise re-design hypothesis and run a new trial.
  • 5. Fine-tuning part II • Start adding realism components: – Move to real subjects (field test) – Move to real tasks (with real subjects) – Move to real subjects handling real tasks – Move to real situation (field experiment) • During the process you: – Lose control over ability to manipulate variables – Gain awareness of interaction between variables • Let’s look at what we are doing with case studies!
  • 6. Telefonica I+D case study • Corporate portal • What is the most obvious incentive from economic point of view? • What can we do with a small budget to be dedicated to incentivize users? • How do we know which system is the best for our setting? 1/30/2015 www.insemtives.eu 6
  • 7. Basic experiment • Test Two rewarding/incentives systems • Pay per click: – 0,03 € per tag added (up to 3 € maximum). • Winner takes all model: – The person who adds the higher number of tags/annotation wins 20€ What would you choose? (Participation fee – 5 €) 1/30/2015 www.insemtives.eu 7
  • 8. The experiment (setting) • 36 students – Random assignment to the two “treatments” • Individual task: annotation of images • Clear set of Instructions • Training (guided) session to give basic understanding of annotation tool • 8 minutes clocked session (time pressure) • Goal: produce maximum amount of tags in allotted time on a random set of images 1/30/2015 www.insemtives.eu 8
  • 15. ID N. Tags Reward Tot. € 2 48 1,5 6,5 3 41 1,5 6,5 4 78 2,5 7,5 5 54 2 7 6 51 2 7 7 36 1,5 6,5 8 44 1,5 6,5 9 54 2 7 10 64 2 7 11 41 1,5 6,5 12 63 2 7 13 58 2 7 14 60 2 7 15 25 1 6 16 43 1,5 6,5 17 50 1,5 6,5 18 24 1 6 19 30 1 6 20 37 1,5 6,5 tot. 901 31,5 126,5 av. 47,42 1,66 6,66 15 ID N. Tags Reward Tot. € 2 86 0 5 3 40 0 5 4 68 0 5 5 88 0 5 6 87 0 5 7 65 0 5 8 67 0 5 9 31 0 5 10 62 0 5 11 79 0 5 12 45 0 5 13 96 20 25 14 51 0 5 15 68 0 5 17 73 0 5 18 26 0 5 19 35 0 5 tot. 1067 20 105 av. 62,76 1,18 6,18
  • 16. Number of tags Pay per tag (N=19) – Total amount of tags: 901 – Max n. of tags: 78 – N. tags (avg.)= 47.42 – € (avg. per person)= 6.66 – € (avg. per tag) =0.1404 – € total = 126,5 € (31,5 € flexible compensation) Winner takes all model (N=17) – Total amount of tags: 1067 – Max n. of tags: 96 – N. tags (avg.)= 62.76 (32% increase!) – € (avg. per person)= 6.18 – € (avg. per tag)=0.098407 – € total = 105 € (20 € flexible compensation) 1/30/2015 www.insemtives.eu 16
  • 18. Tags distribution: interface matters! Pay per tag • Tag “nature”  24 times • “snow”  22 times • “green”  20 times • … • 134 tags repeated only 2 times • 437 unique tags Winner take all • Tag “green” 18 times • “snow”  14 times • “butterfly”  13 times • … • 118 tags repeated only 2 times • 390 unique tags 1/30/2015 www.insemtives.eu 18
  • 19. Some biases • Students are – Volunteers who are used participating in experiments – Strong web users and game players – Paid to show up • Quality of the tags – Quality of tagging has been controlled for: no obvious ‘mistakes’ or ‘cheating’ 1/30/2015 www.insemtives.eu 19
  • 20. Summary of results & next lab steps • Basic hypothesis confirmed • More work needed: – Effort directed to producing a good (tags) that are not consumed by users (used to achieve other goals)  change structure of the game to let users exploit tagging to achieve results (treasure hunt!) – Re-run experiment with new structure. Now users produce tags to get money and to use tags to perform more tasks)
  • 21. Next steps: Telefonica I+D 1/30/2015 www.insemtives.eu 21 • Replicate experiment with real users – Main change 1: task becomes relevant in terms of practical usefulness for users – Main change 2: task has social implications – Main change 3: expectations change dramatically (workers vs. students  5 Euros to participate???) • Add realism – Mimic social structure in the company: • Run experiment with teammates • Use real tasks • Try alternative pay for performance schemes

Notas del editor

  1. What is the most obvious incentive for people from economic point of view? – MONEY!!!!!!! How much money an average project in semantic annotation normally has to dedicate to incentivize users? – Little What can we do with a little budget to be dedicated to incentivize users? – Give a little bit to every user or you can keep the whole sum and assign it to the most productive user, other How do we know which system is best for us? – We need to test it In the context of Telefonica use case, corporate environment, we test it with the help of experiment.
  2. We start with the laboratory experiment with University students Doing the experiment in experimental laboratory requires us to provide incentives to students to participate. They are not our friends that give us a favor to test the software or students in our course that earn course credit – it is not what we want. We want subjects that are neutral to us and to the task and that will reply only to the incentive structure that we provide. There is a fee that we need to pay to students no matter what just to maintain reputation of the laboratory to be sure that students will keep coming to the experiments organized also by other researchers. You can think of environments where you can run the test and don’t pay the participation fee but offer only the flexible part (Mechanicle Turk?) Don’t look at the €5 but concentrate on the flexible part of the payment
  3. 36 students – randomly assigned, no previous experience required, no knowledge of the tool We used the Telefonica annotation application to annotate images Notice, however, that for the first testing of our incentive system we do not really need the real tool. We can also do it with some other task that can be percieved as similar in terms of effort.
  4. For payment requirements we had to round rewards in pay per tag treatment to the highest 50 cents (without rounding the budget was 27 euro)
  5. With a budget of € 20 we get a 32% increase in the average number of annotations per person compared to a €31,5 budget (don’t look at totals!!! – we have different number of subjects in the sessions!) Notice that Winner takes all is realtively scale free. We could have had 40-50 subjects in the session(we had only 17 subjects due to constraints in the laboratory) and obtaine similar results with the same budget. While in the pay per tag scenario for this number of subjects our budget should be at least €60-€75.
  6. T-student = 2.58, p-value = 0.0089 assuming unequal variance F-test for the significance of the difference between the Variances of the two samples F=2.34, p-value = 0,042
  7. This experiment is based on the previous experience of Telefonica: failures The amount of money they gave to employees in ridicolous 300 € for blog entires? More work: the task of subjects in the experiment was just to annotate images. They didn’t care why and they didn’t need annotations for anything. In a real setting of Telefonica annotations are needed to users for their work and work of their collegues. We are in the setting of pubblic good: when you upload a document and annotate it (or finally find the document you were looking for) you need to spend time to do the good annotation, time that you can spend doing your job for which you are actually paid for. They are happy if the document is annotated so that they safe time when look for a document but they don’t want to annotate it themselves, to safe time. The next experiment should incorporate these features of the problem – annotations should be useful to the user for performing the main task, where the main task is not annotations Corporate portal is a specific case. Here we are not only concerned with the fact that people need to produce annotations but also with the fact that they don’t need to dedicate too much time to annotating. The problem here is to find the “right” set of incentives that makes users to produce the “right” amount of annotations. Here the mechanism design comes to help us. We model our problem and solve it to define incentive structure. Then we test the structure in the laboratory with the task that permits annotations to help with the main task