Fabricio Breve Department of Electrical & Computer Engineering

Slides:



Advertisements
Similar presentations
Unsupervised Learning
Advertisements

1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
An Illustrative Example
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Graph-Based Semi-Supervised Learning with a Generative Model Speaker: Jingrui He Advisor: Jaime Carbonell Machine Learning Department
Semi-Supervised Learning D. Zhou, O Bousquet, T. Navin Lan, J. Weston, B. Schokopf J. Weston, B. Schokopf Presents: Tal Babaioff.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Radial Basis Function Networks
Interactive Image Segmentation of Non-Contiguous Classes using Particle Competition and Cooperation Fabricio Breve São Paulo State University (UNESP)
Abstract - Many interactive image processing approaches are based on semi-supervised learning, which employ both labeled and unlabeled data in its training.
Semi-Supervised Learning with Concept Drift using Particle Dynamics applied to Network Intrusion Detection Data Fabricio Breve Institute of Geosciences.
Particle Competition and Cooperation to Prevent Error Propagation from Mislabeled Data in Semi- Supervised Learning Fabricio Breve 1,2
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
1 11 Subcarrier Allocation and Bit Loading Algorithms for OFDMA-Based Wireless Networks Gautam Kulkarni, Sachin Adlakha, Mani Srivastava UCLA IEEE Transactions.
Combined Active and Semi-Supervised Learning using Particle Walking Temporal Dynamics Fabricio Department of Statistics, Applied.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Particle Competition and Cooperation for Uncovering Network Overlap Community Structure Fabricio A. Breve 1, Liang Zhao 1, Marcos G. Quiles 2, Witold Pedrycz.
1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,
Department of Electrical Engineering, Southern Taiwan University Robotic Interaction Learning Lab 1 The optimization of the application of fuzzy ant colony.
Image Classification 영상분류
Uncovering Overlap Community Structure in Complex Networks using Particle Competition Fabricio A. Liang
Particle Competition and Cooperation in Networks for Semi-Supervised Learning with Concept Drift Fabricio Breve 1,2 Liang Zhao 2
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Query Rules Study on Active Semi-Supervised Learning using Particle Competition and Cooperation Fabricio Department of Statistics,
Chapter 6 Neural Network.
Online Social Networks and Media Absorbing random walks Label Propagation Opinion Formation.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
IEEE AI - BASED POWER SYSTEM TRANSIENT SECURITY ASSESSMENT Dr. Hossam Talaat Dept. of Electrical Power & Machines Faculty of Engineering - Ain Shams.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Learning and Removing Cast Shadows through a Multidistribution Approach Nicolas Martel-Brisson, Andre Zaccarin IEEE TRANSACTIONS ON PATTERN ANALYSIS AND.
CSE 4705 Artificial Intelligence
Big data classification using neural network
Chapter 5 Unsupervised learning
Fuzzy Logic in Pattern Recognition
Self-Organizing Network Model (SOM) Session 11
Semi-Supervised Clustering
Deep Learning Amin Sobhani.
Self Organizing Maps: Parametrization of Parton Distribution Functions
Constrained Clustering -Semi Supervised Clustering-
Outlier Processing via L1-Principal Subspaces
Network-Wide Bike Availability Clustering Using the College Admission Algorithm: A Case Study of San Francisco Bay Area Hesham Rakha, Ph.D., P.Eng. Samuel.
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
COMBINED UNSUPERVISED AND SEMI-SUPERVISED LEARNING FOR DATA CLASSIFICATION Fabricio Aparecido Breve, Daniel Carlos Guimarães Pedronette State University.
Hidden Markov Models Part 2: Algorithms
REMOTE SENSING Multispectral Image Classification
Jianping Fan Dept of Computer Science UNC-Charlotte
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Artificial Intelligence Chapter 3 Neural Networks
Emre O. Neftci  iScience  Volume 5, Pages (July 2018) DOI: /j.isci
IEEE World Conference on Computational Intelligence – WCCI 2010
Problem definition Given a data set X = {x1, x2, …, xn} Rm and a label set L = {1, 2, … , c}, some samples xi are labeled as yi  L and some are unlabeled.
Artificial Intelligence Chapter 3 Neural Networks
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
The Naïve Bayes (NB) Classifier
Generalized Locality Preserving Projections
Model generalization Brief summary of methods
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Semi-Supervised Learning
David Kauchak CS158 – Spring 2019
Keshav Balasubramanian
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

Particles Competition and Cooperation in Networks for Semi-Supervised Learning Fabricio Breve Department of Electrical & Computer Engineering University of Alberta Seminar - October 09, 2009

Contents Introduction Model Description Computer Simulations Semi-Supervised Learning Model Description Initial Configuration Node and Particle Dynamics Random-Deterministic Walk Algorithm Computer Simulations Synthetic Data Sets Real-World Data Sets Fuzzy Output and Outlier Detection Conclusions

Introduction Data sets under processing are becoming larger In many situations only a small subset of items can be effectively labeled Labeling process is often: Expensive Time consuming Requires intensive human involvement

Introduction Supervised Learning Unsupervised Learning Only labeled data items are used for training Unsupervised Learning All data items are unlabeled Semi-Supervised Learning Combines a few labeled data items with a large number of unlabeled data to produce betters classifiers

X. Zhu, “Semi-supervised learning literature survey,” Computer Sciences, University of Wisconsin-Madison, Tech. Rep. 1530, 2005.

Semi-Supervised Learning: Graph-Based Methods X. Zhu, Z. Ghahramani, and J. Lafferty, “Semi-supervised learning using gaussian fields and harmonic functions,” in Proceedings of the Twentieth International Conference on Machine Learning, 2003, pp. 912–919. D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Schölkopf, “Learning with local and global consistency,” in Advances in Neural Information Processing Systems, vol. 16. MIT Press, 2004, pp. 321–328. [Online]. Available: http://www.kyb.tuebingen.mpg.de/bs/people/weston/localglobal.pdf X. Zhu and Z. Ghahramani, “Learning from labeled and unlabeled data with label propagation,” Carnegie Mellon University, Pittsburgh, Tech. Rep. CMU-CALD-02-107, 2002. [Online]. Available: http://citeseer.ist.psu.edu/581346.html F. Wang and C. Zhang, “Label propagation through linear neighborhoods,” IEEE Transactions on Knowledge and Data Engineering, vol. 20, no. 1, pp. 55–67, Jan. 2008. A. Blum and S. Chawla, “Learning from labeled and unlabeled data using graph mincuts,” in Proceedings of the Eighteenth International Conference on Machine Learning. San Francisco: Morgan Kaufmann, 2001, pp. 19–26. M. Belkin, I. Matveeva, and P. Niyogi, “Regularization and semisupervised learning on large graphs,” in Conference on Learning Theory. Springer, 2004, pp. 624–638. M. Belkin, N. P., and V. Sindhwani, “On manifold regularization,” in Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics (AISTAT 2005). New Jersey: Society for Artificial Intelligence and Statistics, 2005, pp. 17–24. T. Joachims, “Transductive learning via spectral graph partitioning,” in Proceedings of International Conference on Machine Learning. AAAI Press, 2003, pp. 290–297.

Graph-Based Methods Advantage of identifying many different class distributions Most of them share the regularization framework, differing only in the particular choice of the loss function and the regularizer Most of them have high order of computational complexity (O(n3)), making their applicability limited to small or middle size data sets. X. Zhu, “Semi-supervised learning literature survey,” Computer Sciences, University of Wisconsin-Madison, Tech. Rep. 1530, 2005.

Particle Competition M. G. Quiles, L. Zhao, R. L. Alonso, and R. A. F. Romero, “Particle competition for complex network community detection,” Chaos, vol. 18, no. 3, p. 033107, 2008. [Online]. Available: http://link.aip.org/link/?CHAOEH/18/033107/ 1 Particles walk in the network and compete with each other in such a way that each of them tries to possess as many nodes as possible. Each particle prevents other particles to invade its territory. Finally, each particle is confined inside a network community.

Illustration of the community detection process by competitive particle walking. The total number if nodes is N=128, the number of communities is M=4. The proportion of out links is zout / k=0.2, and the average node degree is k=16. (a) Initial configuration. Four particles, represented by yellow the lightest gray, cyan the second lightest gray, orange the third lightest gray, and blue the second darkest gray, are randomly put in the network. Red the darkest gray represents free nodes. (b) A snapshot at iteration 250. (c) A snapshot at iteration 3500. (d) A snapshot at iteration 7000.

Proposed Method Particles competition and cooperation in networks Competition for possession of nodes of the network Cooperation among particles from the same team (label) Each team of particles tries to dominate as many nodes as possible in a cooperative way and at the same time prevent intrusion of particles of other teams. Random-deterministic walk

Initial Configuration A particle is generated for each labeled node of the network Particle’s home node Particles with same label play for the same team Nodes have an ownership vector Labeled nodes have ownership set to their respective teams. Ex: [ 1 0 0 0 ] (4 classes, node labeled as class A) Unlabeled nodes have levels set equally for each team Ex: [ 0.25 0.25 0.25 0.25 ] (4 classes, unlabeled node) Particles initial position is set to their respective home nodes.

Node and Particle Dynamics Node Dynamics When a particle selects a neighbor to visit: It decreases the domination level of the other teams in this same node It increases the domination level of its team in the target node Exception: Labeled nodes domination levels are fixed t t+1

Node and Particle Dynamics A particle will get: stronger when it is targeting a node being dominated by its team weaker when it is targeting a node dominated by other teams

Node and Particle Dynamics Distance table Keep the particle aware of how far it is from its home node Prevents the particle from losing all its strength when walking into enemies neighborhoods Keep them around to protect their own neighborhood. Updated dynamically with local information Does not require any prior calculation 1 1 2 4 2 3 3 4 ? 4

Particles Walk Shocks A particle really visits a target node only if the domination level of its team is higher than others; otherwise, a shock happens and the particle stays at the current node until next iteration. How a particle chooses a neighbor node to target? Random walk Deterministic walk

Random-deterministic walk Random walk The particle randomly chooses any neighbor to visit with no concern about domination levels or distance Deterministic walk The particle will prefer visiting nodes that its team already dominates and nodes that are closer to their home nodes The particles must exhibit both movements in order to achieve an equilibrium between exploratory and defensive behavior

Deterministic Moving Probabilities Random Moving Probabilities

Algorithm Build the adjacency matrix, Set nodes domination levels, Set initial positions of particles at their corresponding home nodes. Set particle strength and distance, Repeat steps 5 to 9 until convergence or until a predefined number of steps has been achieved, For each particle, complete steps 6 to 9, Select the target node by using the combined random- deterministic rule, Update target node domination levels, Update particle strength, Update particle distance table, Label each unlabeled data item by the team of maximum level of domination.

Computer Simulations SYNTHETIC DATA SETS

Fig. 1. Classification of the banana-shaped patterns Fig. 1. Classification of the banana-shaped patterns. (a) toy data set with 2; 000 samples divided in two classes, 20 samples are pre-labeled (red circles and blue squares). (b) classification achieved by the proposed algorithm.

Fig. 3. Time series for different values of pdet: (a) correct detection rate (b) nodes’ maximum domination level (c) average particle strength. Each point is the average of 200 realizations using a banana-shaped toy data set

Computer Simulations Real-world data sets

5620 amostras

Fuzzy Output and Outlier Detection There are common cases where some nodes in a network can belong to more than one community Example: In a social network of friendship, individuals often belong to several communities: their families, their colleagues, their classmates, etc These are called overlap nodes Most known community detection algorithms do not have a mechanism to detect them

Fuzzy Output and Outlier Detection Particle’s standard algorithm Final ownership levels define nodes labels Very volatile under certain conditions In overlap nodes the dominating team changes frequently Levels do not correspond to overlap measures Particle’s modified algorithm New variable: temporal averaged domination level for each team at each node Weighted by particle strength Considers only the random movements Now the champion is not the team who have won the last games, but rather the team who have won more games in the whole championship

Fig. 9. Fuzzy classification of two banana-shaped classes generated with different variance parameters: (a) s = 0.6 (b) s = 0.8 (c) s = 1.0. Nodes size and colors represent their respective overlap index detected by the proposed method.

Fig. 9. Fuzzy classification of two banana-shaped classes generated with different variance parameters: (a) s = 0.6 (b) s = 0.8 (c) s = 1.0. Nodes size and colors represent their respective overlap index detected by the proposed method.

Fig. 9. Fuzzy classification of two banana-shaped classes generated with different variance parameters: (a) s = 0.6 (b) s = 0.8 (c) s = 1.0. Nodes size and colors represent their respective overlap index detected by the proposed method.

Fig. 10. Classification of normally distributed classes (Gaussian distribution). (a) toy data set with 1,000 samples divided in four classes, 20 samples are labeled, 5 from each class (red squares, blue triangles, green lozenges and purple stars). (b) nodes size and colors represent their respective overlap index detected by the proposed method.

Fig. 10. Classification of normally distributed classes (Gaussian distribution). (a) toy data set with 1,000 samples divided in four classes, 20 samples are labeled, 5 from each class (red squares, blue triangles, green lozenges and purple stars). (b) nodes size and colors represent their respective overlap index detected by the proposed method.

Fig. 11. Comparative between the standard and the modified models: (a) artificial data set with some wrongly labeled nodes (b) classification by the standard particles method (c) classification by the modified particles method

Fig. 11. Comparative between the standard and the modified models: (a) artificial data set with some wrongly labeled nodes (b) classification by the standard particles method (c) classification by the modified particles method

Fig. 11. Comparative between the standard and the modified models: (a) artificial data set with some wrongly labeled nodes (b) classification by the standard particles method (c) classification by the modified particles method

Fig. 12. The karate club network Fig. 12. The karate club network. Nodes size and colors represent their respective overlap index detected by the proposed method.

Conclusions The main contributions of the proposed model can be outlined in the following way: unlike most other graph-based models, it does not rely on loss functions or regularizers; it can classify data with many different distribution, including linearly non-separable data; it has a lower order of complexity than other graph-based models, thus it can be used to classify large data sets; it can achieve better classification rate than other classical graph-based methods; it can detect overlap nodes and provide a fuzzy output for each of them; it can be used to detect outliers and, consequently, to stop error propagation.