Web Usage Mining Classification Fang Yao MEMS 2002 185029 Humboldt Uni zu Berlin.

Slides:



Advertisements
Similar presentations
Lecture 3: CBR Case-Base Indexing
Advertisements

COMP3740 CR32: Knowledge Management and Adaptive Systems
Machine Learning in Real World: C4.5
Decision Trees with Numeric Tests
Decision Tree Algorithm (C4.5)
ICS320-Foundations of Adaptive and Learning Systems
Classification Techniques: Decision Tree Learning
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 3 of Data Mining by I. H. Witten, E. Frank and M. A. Hall.
Decision Trees.
Decision Trees Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei Han.
Lazy Associative Classification By Adriano Veloso,Wagner Meira Jr., Mohammad J. Zaki Presented by: Fariba Mahdavifard Department of Computing Science University.
Constructing Decision Trees. A Decision Tree Example The weather data example. ID codeOutlookTemperatureHumidityWindyPlay abcdefghijklmnabcdefghijklmn.
Classification: Decision Trees
Induction of Decision Trees
1 Classification with Decision Trees I Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei.
Decision Trees an Introduction.
Decision Trees Chapter 18 From Data to Knowledge.
Classification: Decision Trees 2 Outline  Top-Down Decision Tree Construction  Choosing the Splitting Attribute  Information Gain and Gain Ratio.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Data Mining – Algorithms: OneR Chapter 4, Section 4.1.
Machine Learning Chapter 3. Decision Tree Learning
Classification II. 2 Numeric Attributes Numeric attributes can take many values –Creating branches for each value is not ideal The value range is usually.
Appendix: The WEKA Data Mining Software
Data Mining – Algorithms: Prism – Learning Rules via Separating and Covering Chapter 4, Section 4.4.
Data Mining Schemes in Practice. 2 Implementation: Real machine learning schemes  Decision trees: from ID3 to C4.5  missing values, numeric attributes,
Classification I. 2 The Task Input: Collection of instances with a set of attributes x and a special nominal attribute Y called class attribute Output:
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Decision-Tree Induction & Decision-Rule Induction
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.3: Decision Trees Rodney Nielsen Many of.
W E K A Waikato Environment for Knowledge Analysis Branko Kavšek MPŠ Jožef StefanNovember 2005.
Data Mining – Algorithms: Decision Trees - ID3 Chapter 4, Section 4.3.
CS690L Data Mining: Classification
1Weka Tutorial 5 - Association © 2009 – Mark Polczynski Weka Tutorial 5 – Association Technology Forge Version 0.1 ?
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.2 Statistical Modeling Rodney Nielsen Many.
 Classification 1. 2  Task: Given a set of pre-classified examples, build a model or classifier to classify new cases.  Supervised learning: classes.
Data Mining – Algorithms: Naïve Bayes Chapter 4, Section 4.2.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Sections 4.1 Inferring Rudimentary Rules Rodney Nielsen.
Slide 1 DSCI 4520/5240: Data Mining Fall 2013 – Dr. Nick Evangelopoulos Lecture 5: Decision Tree Algorithms Material based on: Witten & Frank 2000, Olson.
W E K A Waikato Environment for Knowledge Aquisition.
Data Management and Database Technologies 1 DATA MINING Extracting Knowledge From Data Petr Olmer CERN
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
An Exercise in Machine Learning
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.5: Mining Association Rules Rodney Nielsen.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.1: Decision Trees Rodney Nielsen Many /
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
Chapter 4: Algorithms CS 795. Inferring Rudimentary Rules 1R – Single rule – one level decision tree –Pick each attribute and form a single level tree.
Machine Learning Recitation 8 Oct 21, 2009 Oznur Tastan.
1 Decision Trees. 2 OutlookTemp (  F) Humidity (%) Windy?Class sunny7570true play sunny8090true don’t play sunny85 false don’t play sunny7295false don’t.
Decision Trees by Muhammad Owais Zahid
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Data Mining Practical Machine Learning Tools and Techniques Chapter 6.3: Association Rules Rodney Nielsen Many / most of these slides were adapted from:
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
Rodney Nielsen Many / most of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Mining Practical Machine Learning Tools and Techniques.
DECISION TREES An internal node represents a test on an attribute.
Decision Trees an introduction.
Data Science Algorithms: The Basic Methods
Classification Algorithms
Teori Keputusan (Decision Theory)
Prepared by: Mahmoud Rafeek Al-Farra
Data Science Algorithms: The Basic Methods
Decision Trees: Another Example
Artificial Intelligence
Data Science Algorithms: The Basic Methods
Decision Tree Saed Sayad 9/21/2018.
Advanced Artificial Intelligence
Machine Learning: Lecture 3
Dept. of Computer Science University of Liverpool
Data Mining CSCI 307, Spring 2019 Lecture 15
Data Mining CSCI 307, Spring 2019 Lecture 6
Data Mining CSCI 307, Spring 2019 Lecture 9
Presentation transcript:

Web Usage Mining Classification Fang Yao MEMS Humboldt Uni zu Berlin

Contents: Defination and the Usages Outputs of Classification Methods of Classification Application to EDOC Discussion on Imcomplete Data Discussion questions & Outlook Humboldt Uni zu Berlin

Classification A Major Data Mining Operation Give one attribute (e.g play), try to predict the value of new people ’ s behavior by means of some other available attributes. Humboldt Uni zu Berlin Usages: Behavior predictions improve Web design personal marketing …… “People with age less than 40 and salary > 40k trade on-line”

A Small Example Humboldt Uni zu Berlin Weather Data Source: Witten & Frank, table 1.2 outlooktemperaturehumiditywindyplay sunnyhothighfalseno sunnyhothightureno overcasthothighfalseyes rainymildhighfalseyes rainycoolnormalfalseyes rainycoolnormaltrue …. no Decision Tree outlook yes windy overcast rainy sunny false true …. humidity …. no high ….

Outputs of Classification Humboldt Uni zu Berlin Decision TreeClassification Rules If outlook = sunny and humidity = high then play = no If outlook = rainy and windy = true then play = no If outlook = overcast then play = yes If humidity = normal then play = yes outlook yes windy overcast rainy sunny false true …. humidity …. no high ….

Methods _ divide-and-conquer constructing decision trees Humboldt Uni zu Berlin Step 1: select a splitting attribute humidity Yes No high normal Yes No Gain(humidity) 0,152 bits windy Yes No true false Yes No Gain(windy) 0,048 bits temp. Yes No hot mild cool Yes yes No Yes No Gain(temperature): 0,029 bits > > outlook Yes overcast rainy sunny Yes No Yes No Gain(outlook): 0,247 bits >

Methods _ divide-and-conquer constructing decision trees Humboldt Uni zu Berlin Calculation information gain: Gain(outlook) = info([9,5]) – info([4,0],[3,2],[2,3])=0,247 bits info([4,0],[3,2],[2,3]) = (4/14)info([4,0]) + (5/14)info([3,2]) + (5/14)info([2,3]) Where: Informational value of creating a branch on the „ outlook “ outlook Yes overcast rainy sunny Yes No Yes No

Methods _ divide-and-conquer calculating information Humboldt Uni zu Berlin Formula for information value : nnn pppppppppentropylog...log ),...,,(  Logarithms are expressed in base 2. unit is ‚bits‘ argument p is expressed as fraction that add up to 1. Info([2,3])=entropy (2/5,3/5) = -2/5*log2/5 - 3/5*log3/5 =0,97 Example :

Methods _ divide-and-conquer calculating information Humboldt Uni zu Berlin

Methods _ divide-and-conquer constructing decision trees Humboldt Uni zu Berlin outlook yes overcast rainy sunny ??

Methods _ divide-and-conquer Humboldt Uni zu Berlin Step 2: select a daughter attribute___ outlook = sunny humidity No high normal Yes Gain(humidity) 0,971 bits temp. No hot mild cool Yes No Yes Gain(temperature): 0,571 bits > windy Yes No truefalse Yes No Gain(windy) 0,020 bits > Do this recursively !!!

Methods _ divide-and-conquer constructing decision trees Humboldt Uni zu Berlin outlook yes windy overcast rainy sunny humidity no high no true yes false yes normal Stop rules: stop when all leaf nodes are pure stop when no more attribute can be splited

Methods _ C 4.5 Humboldt Uni zu Berlin WHY C 4.5? The real -world data is more Complicated Numeric attributes Missing values Final solution need more Operations Pruning From trees to rules

Methods _ C 4.5 Humboldt Uni zu Berlin Numeric attributes: binary split with numeric thresholds halfway between the values Missing values: -- Ignoring leads to losing information -- Partial instances Pruning decision tree: -- subtree replacement -- subtree raising A C A B C A B

Application in WEKA Humboldt Uni zu Berlin

Application in WEKA Humboldt Uni zu Berlin Data: Clickstream from log of EDOC on 30 th March Method: J4.8 Algorithm Objective: Prediction of dissertation reading Attributes: HOME {1,0} AU-START {1,0} DSS-LOOKUP {1,0} SH-OTHER {1,0} OTHER {1,0} AUHINWEISE {1,0} DSS-RVK {1,0} AUTBERATUNG {1,0} DSS-ABSTR {1,0} HIST-DISS {1,0} OT-PUB-READ {1,0} OT-CONF {1,0} SH-START {1,0} SH-DOCSERV {1,0} SH-DISS {1,0} OT-BOOKS {1,0} SH-START-E {1,0}

Application in WEKA Humboldt Uni zu Berlin Result: DSS-ABSTR

Application in WEKA Humboldt Uni zu Berlin DSS-Lookup

Discussion on Incomplete data Humboldt Uni zu Berlin Idea: Site-centric data v.s. User-centric data Incomplete data are inferior to the one from Complete data. User-centric data : Site-centric data : Example: User1: Expedia1, Expedia2, Expedia3 User2: Expedia1, Expedia2, Expedia3, Expedia4 User1: Cheaptickets1, Cheaptickets2, Travelocity1, Travelocity2,Expedia1, Expedia2, Travelocity3, Travelocity4, Expedia3,Cheaptickets3 User2: Expedia1, Expedia2, Expedia3, Expedia4 Padmanabhan, B., Zheng, Z., and Kimbrough, S. (2001)

Humboldt Uni zu Berlin Results: Lift curve source: Padmanabhan, B., Zheng, Z., and Kimbrough, S. (2001) figrue Discussion on Incomplete data

Humboldt Uni zu Berlin Discussion Questions & Outlook What is the proper target attribute for an analysis of non-profit site? What data do we prefer to have? Which improvement could be made to the data?

Humboldt Uni zu Berlin References: Witten, I.H., & Frank, E.(2000). Data Mining. Practical Machine Learning Tools and Techniques with Java Implementations. San Diego, CA: Academic Press. Sections ; 4.3; 6.1 Padmanabhan, B., Zheng, Z., and Kimbrough, S. (2001). „Personalization from Incomplete Data: What you don’t know can hurt.“