Project 4 User and Movie 2018/11/10.

Slides:



Advertisements
Similar presentations
Sequential Minimal Optimization Advanced Machine Learning Course 2012 Fall Semester Tsinghua University.
Advertisements

Assessing and Comparing Classification Algorithms Introduction Resampling and Cross Validation Measuring Error Interval Estimation and Hypothesis Testing.
Credibility: Evaluating what’s been learned. Evaluation: the key to success How predictive is the model we learned? Error on the training data is not.
CMU’s TTO3 Tasks Analyze annotation manuals and annotated samples from all TTO3 teams –Status: Formal UML models have been constructed and merged together.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Evaluation of Results (classifiers, and beyond) Biplav Srivastava Sources: [Witten&Frank00] Witten, I.H. and Frank, E. Data Mining - Practical Machine.
Model Assessment and Selection Florian Markowetz & Rainer Spang Courses in Practical DNA Microarray Analysis.
Evaluating Classifiers
SVMLight SVMLight is an implementation of Support Vector Machine (SVM) in C. Download source from :
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
1 1 Slide Evaluation. 2 2 n Interactive decision tree construction Load segmentchallenge.arff; look at dataset Load segmentchallenge.arff; look at dataset.
While Loops Indefinite Iteration. Last lesson we looked at definite loops using the ‘For’ statement. The while loop keeps going while some condition is.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
Classification Performance Evaluation. How do you know that you have a good classifier? Is a feature contributing to overall performance? Is classifier.
 Conversation Level Constraints on Pedophile Detection in Chat Rooms PAN 2012 — Sexual Predator Identification Claudia Peersman, Frederik Vaassen, Vincent.
Manu Chandran. Outline Background and motivation Over view of techniques Cross validation Bootstrap method Setting up the problem Comparing AIC,BIC,Crossvalidation,Bootstrap.
Exploring Alternative Splicing Features using Support Vector Machines Feature for Alternative Splicing Alternative splicing is a mechanism for generating.
Introduction to Software Project Estimation I (Condensed) Barry Schrag Software Engineering Consultant MCSD, MCAD, MCDBA Bellevue.
A Content-Based Approach to Collaborative Filtering Brandon Douthit-Wood CS 470 – Final Presentation.
Weka Just do it Free and Open Source ML Suite Ian Witten & Eibe Frank University of Waikato New Zealand.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Closing Slides Mark Klerer – Jerry Upton Vice-Chairs C /10.
Evaluating Classifiers Reading: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)An introduction to ROC analysis.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
Information Organization: Evaluation of Classification Performance.
IR Homework #2 By J. H. Wang May 9, Programming Exercise #2: Text Classification Goal: to classify each document into predefined categories Input:
Big Data Processing of School Shooting Archives
Data Science Credibility: Evaluating What’s Been Learned
A Simple Approach for Author Profiling in MapReduce
Name: Sushmita Laila Khan Affiliation: Georgia Southern University
Evaluating Classifiers
Chapter 3: Measurement: Accuracy, Precision, and Error
Performance Evaluation 02/15/17
Evaluating Techniques for Image Classification
Can Computer Algorithms Guess Your Age and Gender?
SOCIAL COMPUTING Homework 3 Presentation
Project 2 k-NN 2018/11/10.
Roberto Battiti, Mauro Brunato
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
Mitchell Kossoris, Catelyn Scholl, Zhi Zheng
Title Research.
Project 1 Binary Classification
Evaluation and Its Methods
Scratch: selection / branching/ if / If…else / compound conditionals / error trapping by Mr. Clausen.
ERT 207 Analytical Chemistry
Final Project Description
Predict Failures with Developer Networks and Social Network Analysis
Text Categorization Rong Jin.
SASU manual: sampling issues
Learning Algorithm Evaluation
Movie Recommendation System
INTRODUCTION TO Machine Learning
A Framework for Benchmarking Entity-Annotation Systems
Lecture 6: Introduction to Machine Learning
CSCI N317 Computation for Scientific Applications Unit Weka
Model Evaluation and Selection
Intro to Machine Learning
Week 3 Lecture Notes PSYC2021: Winter 2019.
Model generalization Brief summary of methods
Lecture 10 – Introduction to Weka
Dynamic Category Profiling for Text Filtering and Classification
Project 3 Image Retrieval
Evaluation and Its Methods
Assignment 1: Classification by K Nearest Neighbors (KNN) technique
Machine Learning: Methodology Chapter
Evaluation and Its Methods
ECE – Pattern Recognition Lecture 8 – Performance Evaluation
Information Organization: Evaluation of Classification Performance
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Project 4 User and Movie 2018/11/10

Brief In this experiment, you can choose ONE of the following two tasks Background: MovieLens (http://www.movielens.org/) is a website where users can rate movies. The website released some data regarding user information, movie information, and the ratings Task 1: User profiling. You need to predict the gender and age of users based on their ratings on movies Task 2: Movie annotation. You need to predict the genres of movies based on their ratings given by users 2018/11/10

Data There are three types of data files users.dat: UserID::Gender::Age::Occupation::Zip-code Note: for task 1, users.dat is split into ten files named users.dat0, 1, …, 9. They are used for 10-fold cross validation movies.dat: MovieID::Title::Genres Note: for task 2, movies.dat is split into ten files named movies.dat0, 1, …, 9. They are used for 10-fold cross validation ratings.dat: UserID::MovieID::Rating::Timestamp For more information, please read the movielens_1m_readme All files can be opened with Text Editors (e.g. Notepad, Word, …) 2018/11/10

Task 1 In this task, you need to perform 10-fold cross validation. In each fold, users.datx (x=0,…,9) is used for test, and all other users.datx are used for training. You can use movies.dat and ratings.dat in both training and test and in whatever way The evaluation is To show the error rate of prediction of gender To show the weighted error rate of prediction of age (Remark: in this data set, user’s age is divided into 7 ranges; thus, if the predicted age and the true age are in the same range, error weight is 0; if they are in adjacent ranges, error weight is 1; if they are in range 3 and range 5, error weight is 2; …; if they are in range 1 and range 7, error weight is 6; the final weighted error rate is the average of error weights of all test samples) To show the runtime of training and test Evaluation results of each fold and the average results must be reported 2018/11/10

Task 2 In this task, you need to perform 10-fold cross validation. In each fold, movies.datx (x=0,…,9) is used for test, and all other movies.datx are used for training. You can use users.dat and ratings.dat in both training and test and in whatever way The evaluation is To show the precision, recall, F1-value of prediction of genres (Remark: you shall calculate the average precision and recall, and then calculate the F1- value; please refer to the next slide) To show the runtime of training and test Evaluation results of each fold and the average results must be reported 2018/11/10

Task 2: Average Precision and Recall and F1-value In this data set, each movie has >=1 genres, thus the prediction of genres is a multi-label classification, and the calculation of average precision and recall and F1-value is as follows For example, if there are 2 test samples, and the ground-truth genres of the test samples are {C, L} and {T, H, W}, respectively If the predicted genres are {C} and {T, F}, then the total true positive is 2 (C and T), the total TP+FN=5 (C,L,T,H,W), the total TP+FP=3 (C,T,F); thus the average precision is 2/3 and the average recall is 2/5 and the average F1-value is 1/2 If the predicted genres are {C, L, F} and {T, W, C}, then the average precision is 4/6 and the average recall is 4/5 and the average F1-value is 8/11 … 2018/11/10

Report .pdf is best, .doc and .ppt are acceptable English is encouraged List all references Source code is a plus 2018/11/10