Crowdsourcing Insights with Opinion Space Ken Goldberg, IEOR, School of Information, EECS, UC Berkeley.

Slides:



Advertisements
Similar presentations
Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Advertisements

Component Analysis (Review)
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
INTRODUCTION TO Machine Learning 2nd Edition
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Dimensionality Reduction PCA -- SVD
Dimension reduction (1)
Optimization for Radiology and Social Media
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Classification and Decision Boundaries
Discriminative and generative methods for bags of features
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
EigenTaste: A Constant Time Collaborative Filtering Algorithm Ken Goldberg Students: Theresa Roeder, Dhruv Gupta, Chris Perkins Industrial Engineering.
Principal Component Analysis
Dimensionality Reduction and Embeddings
Dimensional reduction, PCA
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
Clustering.
SVD and PCA COS 323. Dimensionality Reduction Map points in high-dimensional space to lower number of dimensionsMap points in high-dimensional space to.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Dimensionality Reduction
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Microarray Gene Expression Data Analysis A.Venkatesh CBBL Functional Genomics Chapter: 07.
Summarized by Soo-Jin Kim
Presented By Wanchen Lu 2/25/2013
Cao et al. ICML 2010 Presented by Danushka Bollegala.
Alignment Introduction Notes courtesy of Funk et al., SIGGRAPH 2004.
Principles of Pattern Recognition
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Whitening.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Optimization for Radiology and Social Media Ken Goldberg IEOR (EECS, School of Information, BCNM) UC Berkeley College of Engineering Research Council,
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Dimension Reduction in Workers Compensation CAS predictive Modeling Seminar Louise Francis, FCAS, MAAA Francis Analytics and Actuarial Data Mining, Inc.
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Ken goldberg, gail de kosnik, kimiko ryokai (+ students) uc berkeley Opinion Space.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
CS685 : Special Topics in Data Mining, UKY The UNIVERSITY of KENTUCKY Dimensionality Reduction CS 685: Special Topics in Data Mining Spring 2008 Jinze.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Dimensionality Reduction Part 1: Linear Methods Comp Spring 2007.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Opinion Space visualizing social media for insight and innovation Ken Goldberg, UC Berkeley
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Spectral Methods for Dimensionality
Visualizing High-Dimensional Data
Dimensionality Reduction
CS 2750: Machine Learning Dimensionality Reduction
Dimensionality Reduction
COSC 4335: Other Classification Techniques
Presentation transcript:

Crowdsourcing Insights with Opinion Space Ken Goldberg, IEOR, School of Information, EECS, UC Berkeley

“We’re moving from an Information Age to an Opinion Age.” - Warren Sack, UCSC

Motivation Goals Engage community Understand community – Solicit input – Understand the distribution of viewpoints – Discover insightful comments Goals of Community Members Understand relationship to other community members Participate, express ideas, and be heard Encounter a diversity of viewpoints

Motivation Classic approach: surveys, polls Drawbacks: limited samples, slow, doesn’t increase engagement Modern approach: online forums, comment lists Drawbacks: data deluge, cyberpolarization, hard to discover insights

Approach: Visualization

Approach: Level the Playing Field

Approach: Wisdom of Crowds

Related Work: Visualization Clockwise, starting from top left: Morningside Analytics, MusicBox, Starry Night

Related Work: Politics Clockwise, starting from top left: EU Profiler, Poligraph, The How Progressive Are You? quiz

Related Work: Opinion Sharing Polling & Opinion Mining – Fishkin, 1991: deliberative polling – Dahlgren, 2005: Internet & the public sphere – Berinsky, 1999: understanding public opinion – Pang & Lee, 2008: sentiment analysis Increasing Participation – Bishop, 2007: theoretical framework – Brandtzaeg & Heim: user study – Ludford et al, 2004: uniqueness & group dissimilarity

Related Work: Info Filtering K. Goldberg et al, 2001: Eigentaste E. Bitton, 2009: spatial model Polikar, 2006: ensemble learning

Opinion Space: Live Demonstration

Six 50-minute Learning Object Modules, preparation materials, slides for in-class lectures, discussion ideas, hand-on activities, and homework assignments.

To try it: google: “opinion space” contact us:

Dimensionality Reduction low variance projectionmaximal variance projection

Dimensionality Reduction Principal Component Analysis (PCA) Assumes independence and linearity Minimizes squared error Scalable: compute position of new user in constant time

Canonical Correlation Analysis 2-view PCA Assume: – Each data point has a latent low-dim canonical representation z – Observe two different representations of each data point (e.g. numerical ratings and text) Learn MLEs for low-rank projections A and B Equivalently, pick projection that maximizes correlation between views z z x x y y Graphical model for CCA x = Az + ε y = Bz + ε z = A -1 x = B -1 y

CCA on Opinion Space Each user is a data point – x i = user i’s responses to propositions – y i = vector representation of textual comment Run CCA to find A and B, use A -1 to find 2D representation Position of users reflects rating vector and textual response Ignores ratings that are not correlated with text, and vice versa Given text, can predict ratings (using B) z z x x y y Graphical model for CCA x = Az + ε y = Bz + ε z = A -1 x = B -1 y

Multidimensional Scaling Goal: rearrange objects in low dim space so as to reproduce distances in higher dim Strategy: Rearrange & compare solns, maximizing goodness of fit: Can use any kind of similarity function Pros – Data need not be normal, relationships need not be linear – Tends to yield fewer factors than FA Con: slow, not scalable δ ij i j d ij i j

Kernel-based Nonlinear PCA Intuition: in general, can’t linearly separate n points in d < n dim, but can almost always do so in d ≥ n dim Method: compute covariance matrix after transforming data into higher dim space Kernel trick used to improve complexity If Φ is the identity, Kernel PCA = PCA

Kernel-based Nonlinear PCA Pro: Good for finding clusters with arbitrary shape Cons: Need to choose appropriate kernel (no unique solution); does not preserve distance relationships Input dataKPCA output with Gaussian kernel

Stochastic Neighbor Embedding Converts Euclidean dists to conditional probabilities p j|i = Pr(x i would pick x j as its neighbor | neighbors picked according to their density under a Gaussian centered at x i ) Compute similar prob q j|i in lower dim space Goal: minimize mismatch between p j|i and q j|i : Cons: tends to crowd points in center of map; difficult to optimize

Metavid

Six 50-minute Learning Object Modules, preparation materials, slides for in-class lectures, discussion ideas, hand-on activities, and homework assignments.

Opinion Space: Crowdsourcing Insights Scalability: n Participants, n Viewpoints n 2 Peer to Peer Reviews Viewpoints are k-Dimensional Dim. Reduction: 2D Map of Affinity/Similarity Insight vs. Agreement: Nonlinear Scoring Ken Goldberg, UC Berkeley Alec Ross, U.S. State Dept

Opinion Space Wisdom of Crowds: Insights are Rare Scalable, Self-Organizing, Spatial Interface Visualize Diversity of Viewpoints Incorporate Position into Scoring Metrics Ken Goldberg UC Berkeley