Parametric calibration of speed–density relationships in mesoscopic traffic simulator with data mining Adviser: Yu-Chiang Li Speaker: Gung-Shian Lin Date:2009/10/20.

Slides:



Advertisements
Similar presentations
CLUSTERING.
Advertisements

Clustering II.
Clustering.
Cluster Analysis: Basic Concepts and Algorithms
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering. Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree-like diagram that.
Hierarchical Clustering, DBSCAN The EM Algorithm
Clustering Categorical Data The Case of Quran Verses
Data Mining Cluster Analysis: Advanced Concepts and Algorithms
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ What is Cluster Analysis? l Finding groups of objects such that the objects in a group will.
Data Mining Cluster Analysis: Basic Concepts and Algorithms
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
K Means Clustering , Nearest Cluster and Gaussian Mixture
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ What is Cluster Analysis? l Finding groups of objects such that the objects in a group will.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Clustering II.
Clustering II.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Data Mining Cluster Analysis: Advanced Concepts and Algorithms Lecture Notes for Chapter 9 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach,
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
© University of Minnesota Data Mining CSCI 8980 (Fall 2002) 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center.
Evaluating Hypotheses
Cluster Analysis: Basic Concepts and Algorithms
What is Cluster Analysis?
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
CHAMELEON : A Hierarchical Clustering Algorithm Using Dynamic Modeling
Fixed Parameter Complexity Algorithms and Networks.
Presented by Tienwei Tsai July, 2005
Cluster Analysis Part II. Learning Objectives Hierarchical Methods Density-Based Methods Grid-Based Methods Model-Based Clustering Methods Outlier Analysis.
Image Segmentation Seminar III Xiaofeng Fan. Today ’ s Presentation Problem Definition Problem Definition Approach Approach Segmentation Methods Segmentation.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 Converting Categories to Numbers for Approximate Nearest Neighbor Search 嘉義大學資工系 郭煌政 2004/10/20.
1 Lecture 10 Clustering. 2 Preview Introduction Partitioning methods Hierarchical methods Model-based methods Density-based methods.
1 CSE 980: Data Mining Lecture 17: Density-based and Other Clustering Algorithms.
Clustering Spatial Data Using Random Walk David Harel and Yehuda Koren KDD 2001.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Clustering COMP Research Seminar BCB 713 Module Spring 2011 Wei Wang.
A genetic approach to the automatic clustering problem Author : Lin Yu Tseng Shiueng Bien Yang Graduate : Chien-Ming Hsiao.
Cluster Analysis Cluster Analysis Cluster analysis is a class of techniques used to classify objects or cases into relatively homogeneous groups.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Clustering.
Prepared by: Mahmoud Rafeek Al-Farra
Data Mining Cluster Analysis: Advanced Concepts and Algorithms Lecture Notes for Chapter 9 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach,
Data Mining Cluster Analysis: Advanced Concepts and Algorithms
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Data Mining: Cluster Analysis This lecture node is modified based on Lecture Notes for Chapter.
Clustering Algorithms Sunida Ratanothayanon. What is Clustering?
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
The Science of Physics Mathematics. What We Want to Know… How do tables and graphs help understand data? How can we use graphs to understand the relationship.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
DATA MINING: CLUSTER ANALYSIS Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
CLUSTER ANALYSIS. Cluster Analysis  Cluster analysis is a major technique for classifying a ‘mountain’ of information into manageable meaningful piles.
Parametric calibration of speed–density relationships in mesoscopic traffic simulator with data mining Adviser: Yu-Chiang Li Speaker: Gung-Shian Lin Date:2009/10/20.
Semi-Supervised Clustering
Clustering CSC 600: Data Mining Class 21.
Data Mining K-means Algorithm
Posture Monitoring System for Context Awareness in Mobile Computing
CSE 5243 Intro. to Data Mining
DATA MINING Introductory and Advanced Topics Part II - Clustering
Cluster Analysis.
Text Categorization Berlin Chen 2003 Reference:
Clustering The process of grouping samples so that the samples are similar within each group.
EM Algorithm and its Applications
CS 685: Special Topics in Data Mining Jinze Liu
Clustering.
BIRCH: Balanced Iterative Reducing and Clustering using Hierarchies
Presentation transcript:

Parametric calibration of speed–density relationships in mesoscopic traffic simulator with data mining Adviser: Yu-Chiang Li Speaker: Gung-Shian Lin Date:2009/10/20 Information Sciences, vol.179, no.12, pp. 2002-2013, 2009

Experiments and results Outline 1 Introduction 2 Literature review 3 Data mining 4 Experiments and results 5 Conclusions

1.Introduction Calibrating speed–density relationship parameters using data mining techniques, and proposes a novel hierarchical clustering algorithm based on K-means clustering Mesoscopic simulators aim to model either a single vehicle or a group of vehicles in order to depict any responsive actions of different vehicles to the traffic information.

2.Literature review In the mesoscopic models which are used in DTA systems v0 k vu kjam k0

3.Data mining LWR(Locally weighted regression) Step 1: Take x (densities or both densities and flows make up the x) as a center to form a space. The width of the space isdescribed by the q = fn Step 2: Define the weights of all points in specific sectors. The weight of any point is the height of a weight function. The common weight function is selected: The weight for the observation (xi, yi) is: Step 3: Fit a polynomial for each point in an independent variable space by using the weighted least square algorithm Step 4: Acquire the value of yi.

p(x,xi)<d(x)→W(u)=(1-u3)3 3.Data mining x q=fn p(x,xi)<d(x)→W(u)=(1-u3)3 p(x,xi) ≧d(x)→W(u)=0

3.Data mining Agglomerative hierarchical clustering algorithm based on K-means The proposed algorithm can be summarized as follows Step 1: Use K-means to cluster the sensor data which is taken as training instances, and these k clusters are named as constraint- clusters. Densities, flows and speeds contain abundant information about the traffic status, so they are chosen to cluster.

3.Data mining K-means is executed in the following steps: 1. Randomly select k clustering centers from n training cases. 2. Find the nearest clustering center to each xi (density or both density and flow), then put xi in it. 3. Compute the objective function E. If the value of E is unchanged, we should consider that the results of the clustering are also unchanged. Then the iteration stops. 4. Otherwise, it will return to 2.

3.Data mining k=3

3.Data mining

3.Data mining Step 2:For each constraint-cluster, use the agglomerative hierarchical clustering to build a clustering tree. The basic steps of the complete-link algorithm are: 1. Place each instance in its own cluster. Then, compute the distances between these points. 2. Step thorough the sorted list of distances, forming for each distinct threshold value dk a graph of the samples where pairs of samples closer than dk are connected into a new cluster by a graph edge. If all the samples are members of a connects graph, stop. Otherwise, repeat this step. 3. The output of the algorithm is a nested hierarchy of graphs, which can be cut at the desired dissimilarity level forming a partition (clusters) identified by simple connected components in the corresponding subgraph.

3.Data mining

3.Data mining 樣本1 樣本2 樣本3 樣本4 樣本5 樣本6 樣本7

3.Data mining Step 3: These k clustering trees are combined as an integrated clustering tree by using the complete-link algorithm. After all samples are clustered, a separate local regression will be run for the observation in each cluster. Step 4: The new densities and flows are classified to the most appropriate cluster by using k-nearest neighbors. The k- nearest neighbor sorter uses Euclidean distance to search k densities and flows samples completed clustering.

4. Experiments and results The sensor data are preprocessed to eliminate erroneous data and repair missing ones. Step1: Define data in some cycles as data it is in some phase and scan the sending time of data one by one to find out the missing ones. Check it is erroneous or not according to the criteria in Table. Step 2: Repair the missing data and the erroneous data. The average value in the neighboring phase is used to repair these data.

4. Experiments and results

4. Experiments and results

4. Experiments and results Estimated speed by the classical speed–density relationship

4. Experiments and results

Table presents the RMSPE obtained by each approach.

5. Conclusions The proposed method overcomes the limitations of classic models of speed–density relationships. KHCA obtained the highest precision in capturing traffic dynamics compared to other existing clustering algorithms.

Thank You !