On Assigning Implicit Reputation Scores in an Online Labor Marketplace

Slides:



Advertisements
Similar presentations
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Advertisements

Suleyman Cetintas 1, Monica Rogati 2, Luo Si 1, Yi Fang 1 Identifying Similar People in Professional Social Networks with Discriminative Probabilistic.
Active Learning and Collaborative Filtering
CONCEPT OF SELECTION The next step after requirement is the selection of candidates for the vacant position from among the applicants. This is the most.
Chapter 30: The Labor Market Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin 13e.
Recruiting in Labor Markets Exercise
 The need  The priority  The confidence to take action The Financial Ca$e for Online Staffing Online Assessment plus Tracking delivers outstanding financial.
11 Copyright © Cengage Learning. All rights reserved. 11 Techniques of Differentiation with Applications.
Each One Refer More! Program Update SAP Global Delivery.
Communicating Your Value Marketing You Ch. 7. Wouldn’t it be convenient if employers recognized the contributions you can make? Unfortunately, they don’t.
Presenter: Libin Zheng, Yongqi Zhang Department of Computer Science and Engineering HKUST Date: 24/11/2015 Crowd-aided course selection on MOOC.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
CS-424 Gregory Dudek Lecture 10 Annealing (final comments) Adversary Search Genetic Algorithms (genetic search)
Chapter 23 – Managing Human Resources Human resources management (personnel management) – all activities involved with acquiring, developing, and compensating.
Recruitment, Selection & Employment Unit 8. Starter Activity Think about the definitions for the following words: Recruitment Part time workers Full time.
TrustRank. 2 Observation – Good pages tend to link good pages. – Human is the best spam detector Algorithm – Select a small subset of pages and let a.
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
Copyright © 2009 Pearson Education, Inc. 4.1 What Is Average? LEARNING GOAL Understand the difference between a mean, median, and mode and how each is.
© 2013 by Nelson Education1 Decision Making. Chapter Learning Outcomes  After reading this chapter you should:  Appreciate the complexity of decision.
Functions 2 Copyright © Cengage Learning. All rights reserved.
Reputation Systems For Open Collaboration, CACM 2010 Bo Adler, Luca de Alfaro et al. Nishith Agarwal
Unique Benefits of BBA Distance Education Course.
[ 5.5 ] The Labor Force.
How to Write a Reader-Oriented Letter of Application
Part I Project Initiation.
Stock Market Application: Review
Social networks that matter: Twitter under the microscope
Chapter 1: What is Economics? Section 1
Labor Economics, 4th edition
Recommendation in Scholarly Big Data
Dependent-Samples t-Test
Recommender Systems & Collaborative Filtering
CSC 594 Topics in AI – Natural Language Processing
Labour Supply Two measures reflect two aspects of the worker’s labour supply decision: (i) participation – that is, whether to work or not work; and (ii)
Mixed signal: Charity reporting when donations signal generosity and income Bracha A. and Verstelund L.
Psychology Unit Research Methods - Statistics
Unit 5: Hypothesis Testing
Measuring The Cost of Living
Chapter 11 Resource Markets © 2006 Thomson/South-Western.
Analytic Hierarchy Process (AHP)
User Joining Behavior in Online Forums
Supply & Demand Made Easy
LEVEL OF MEASUREMENT Data is generally represented as numbers, but the numbers do not always have the same meaning and cannot be used in the same way.
PART IV TRAINING THE SALES TEAM. PART IV TRAINING THE SALES TEAM.
Mortgage Broker - The Ultimate Home Buyer's Guide
CHAPTER 4 Test Design Techniques
Hire A Movers & Packers With Confidence
#1 Explain the Work ➢ For technical work, you will be needing remote employees for Long- Term Roles and Assignments. ➢ The job description should clearly.
Advanced Scientometrics Workshop
Effective Social Network Quarantine with Minimal Isolation Costs
#VisualHashtags Visual Summarization of Social Media Events using Mid-Level Visual Elements Sonal Goel (IIIT-Delhi), Sarthak Ahuja (IBM Research, India),
Postdoc, School of Information, University of Arizona
Machine Learning in Practice Lecture 11
Significance Tests: The Basics
4.1 What Is Average? LEARNING GOAL
How Wages are Determined
Labor Contracts and Nominal-Wage Rigidity
Sr. Manager, Global Talent Acquisition
Significance Tests: The Basics
Theory of Computability
Your EirQuest Workbook
COMP444 Human Computer Interaction Usability Engineering
Chapter 11 Resource Markets © 2006 Thomson/South-Western.
Thanks for taking part in this study
Chapter 5: Nonwage labor costs
I am ready for a job, now how do I find one?
Microeconomics Pt.3: What Is Supply?.
Retrieval Performance Evaluation - Measures
Theory of Computability
GhostLink: Latent Network Inference for Influence-aware Recommendation
Presentation transcript:

On Assigning Implicit Reputation Scores in an Online Labor Marketplace Maria Daltayanni, Luca de Alfaro (UCSC), Panagiotis Papadimitriou (oDesk), Panayiotis Tsaparas (UoI) {mariadal, luca}@cs.ucsc.edu, papadimitriou@odesk, tsap@cs.uoi.gr

Labor Marketplace In online labor marketplaces employers post job openings and receive applications by workers interested in them. Data about labor marketplaces are usually repressented with a bipartite graph as shown. The employers decide which applicant to hire and then they work with the selected worker to accomplish the job requirements. At the end of the contract, an employer can provide his worker with some feedback rating that becomes visible in the online worker profile and can guide future hiring decisions of other employers. Usually, the possible outcomes of an application, in order of decreasing employer satisfaction, are offer(hire) > interview > shortlist > ignore > hide > reject. The employer’s feedback rating is an explicit form of reputation for the workers. He selects a number of stars that reflects his belief about the worker quality on the accomplished job.

Which candidate to hire? Use Past Employer Ratings (Explicit Reputation) scores usually skewed towards high ratings very sparse, since a worker needs to apply, get hired and complete few jobs before he obtains a representative reputation score However, as shown in the figure, most employers tend to provide workers with 5 star ratings. Hence explicit feedback does not help discriminate between good workers. At the same time, explicit stars feedback is provided only for workers who eventually got hired and accomplished a job. We not receive any information, though, about the rest candidates who participate in the application process and who did not receive an offer in the last stage.

Which candidate to hire? WorkerRank: Use Past Jobs Application Data (Implicit Reputation) Which workers applied to which jobs? Did they get hired? Interviewed, shortlisted, ignored, hidden, rejected? User-profiling To deal with these two basic problems, the authors devise an algorithm which examines the application data, that is, who applied to which job and what was the hiring outcome. Did the applicant get hired, rejected, etc. This application data provides implicit information about the quality of applicants. For example, the quality of a candidate who got rejected at a job is worse than that of a candidate who was shortlisted by the employer.

WorkerRank ‘PageRank’ for workers and employers General score computation rule: score(worker) += weight * score(employer) , for all employers score(employer) += weight * score(worker) , for all workers Weights include info about the hiring outcome How to specify weights? The used algorithm is based on PageRank where the employers are the hubs and the workers are the authorities The general score computation rule is uses weights. Weights on the edges are used to represent the quality info propagated from worker to employer and vice versa. For example, if a worker got an offer from Google, that would give him higher score than a worker who got hired by a startup since the startup has much fewer incoming edges from applicants. Our algorithm uses HITS (the pagerank algorithm) on the application data. An rwthsei kanenas: (The employers are the hubs The workers are the authorities.)

Relativity k: rank of worker, if we sort them by hiring outcome An example job with 5 candidates can be: ---------- k hiring 1 offer 2 shortlist 3 hide 4 hide 5 reject We may allow for fixed weights, e.g. offer = 3, interview = 2, shortlist = 1, ignore = -1, hide = -2, reject = -3 Or compute relative weights for each job as shown in the figure It is important to consider Relativity, that is, what are the hiring outcomes for the two candidates – Essentially we want to measure “how much better” is one candidate versus the other E.g. offer versus shortlist or offer versus reject

Selectivity n: total number of candidates k: number of candidates who received offer Selectivity: (n-k) / n In addition it is important to consider selectivity (competitiveness): how many candidates applied to the job – it is more competitive to get an offer at a job of 100 candidates than at one with 10 candidates

Elo Ratings: Example Initial Hiring Elo Score Outcome Scores 100 rejected 96 win/ lose 100 ignored 98 In the very next step Elo ratings are applied, which cover both relativity and selectivity in a principled way. Elo ratings perform pairwise comparisons among players and assign implicit reputation scores based on Relativity and competitiveness(selectivity). draw 100 ignored 98 100 hired 108

Skills: Further Work Usage of Skills Derive skill-wise reputation scores For example, we may recommend candidates for a job that requires java Using their java scores As a further work, the authors consider usage of skills; jobs require particular skills and workers claim to have expertise on some skills. The graph can be expanded such that there is one node per (worker, skill) pair instead of worker, and (employer, skill) pair instead of employer. (In fact we use job nodes, which correspond to employers who posted the jobs). WorkerRank can run and export reputation results about worker,skill pairs. Thus instead of learning one score for each worker, we now learn one score for each skill of the worker. For example, an employer who posts a job requiring java, is interested in the java score of a candidate and he does not ask for his python score.

Skills: Further Work Expanding job applications graph to support skills Here is an example graph where we show how it is expanded when we add the skills information. It is obvious that the new graph can be considered as a set of skill-wise graphs, e.g. one java graph, one python graph, etc.

Set Up Dataset: by oDesk Real world application data 53 weeks between March 2012 – March 2013 ~17M applications ~0.5M workers ~0.8M jobs ~0.2 employers For evaluating of WorkerRank application data from the online labor marketplace company oDesk was used. The dataset spans 53 weeks and contains approximatelly ~17M applications ~0.5M workers ~0.8M jobs ~0.2 employers

Results Coverage of Worker Applications: +50.7% Cold Start In this image you can see on the Y-axis the percentage of workers for whom we have obtained reputation after x weeks, which is indicated on the x-axis of the image.The experiments show that the coverage of workers grows by 50.7%. The cold start improvement pertains to how fast we acquire reputation information with each method. Implicit reputation becomes available for about 100% of the new workers joining the platform after 12 weeks. It is very interesting that explicit feedback becomes available for almost 4% of the new workers after 12 weeks that they have joined the platform.

Results Lift(x) = Finally, the initial goal was to learn a reputation score for each worker so that we can sort the candidates of a job by quality and recommend the good workers to the employer. The lift metric is used as the hiring probability of the top x%(per cent) applicants divided by the hiring probability across all applicants This result shows that WorkerRank improves the ranking accuracy compared to explicit feedback. In particular it almost doubles the probability of ranking a good quality worker at the top 0.5% of the list. As a reminder the top ranked workers are of greater interest for our system since the employers will usually look only at the top ~20 candidates.

Thank you