Combined Active and Semi-Supervised Learning using Particle Walking Temporal Dynamics Fabricio Department of Statistics, Applied.

Slides:



Advertisements
Similar presentations
HOPS: Efficient Region Labeling using Higher Order Proxy Neighborhoods Albert Y. C. Chen 1, Jason J. Corso 1, and Le Wang 2 1 Dept. of Computer Science.
Advertisements

Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Using Parallel Genetic Algorithm in a Predictive Job Scheduling
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Modeling and Analysis of Random Walk Search Algorithms in P2P Networks Nabhendra Bisnik, Alhussein Abouzeid ECSE, Rensselaer Polytechnic Institute.
An Approach to Evaluate Data Trustworthiness Based on Data Provenance Department of Computer Science Purdue University.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Co-Training and Expansion: Towards Bridging Theory and Practice Maria-Florina Balcan, Avrim Blum, Ke Yang Carnegie Mellon University, Computer Science.
Semi-supervised learning and self-training LING 572 Fei Xia 02/14/06.
Active Learning of Binary Classifiers
Dept. of Computer Science & Engineering, CUHK1 Trust- and Clustering-Based Authentication Services in Mobile Ad Hoc Networks Edith Ngai and Michael R.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
Reinforcement Learning Rafy Michaeli Assaf Naor Supervisor: Yaakov Engel Visit project’s home page at: FOR.
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
On the Construction of Energy- Efficient Broadcast Tree with Hitch-hiking in Wireless Networks Source: 2004 International Performance Computing and Communications.
Ensemble Learning: An Introduction
Presented by Zeehasham Rasheed
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Semi-Supervised Learning D. Zhou, O Bousquet, T. Navin Lan, J. Weston, B. Schokopf J. Weston, B. Schokopf Presents: Tal Babaioff.
Multilayer Perceptron Classifier Combination for Identification of Materials on Noisy Soil Science Multispectral Images Fabricio A.
Interactive Image Segmentation of Non-Contiguous Classes using Particle Competition and Cooperation Fabricio Breve São Paulo State University (UNESP)
Abstract - Many interactive image processing approaches are based on semi-supervised learning, which employ both labeled and unlabeled data in its training.
Semi-Supervised Learning with Concept Drift using Particle Dynamics applied to Network Intrusion Detection Data Fabricio Breve Institute of Geosciences.
ENN: Extended Nearest Neighbor Method for Pattern Recognition
Active Learning for Class Imbalance Problem
Particle Competition and Cooperation to Prevent Error Propagation from Mislabeled Data in Semi- Supervised Learning Fabricio Breve 1,2
Quasar A Probabilistic Publish-Subscribe System for Social Networks over P2P Kademlia network David Arinzon Supervisor: Gil Einziger April
WALKING IN FACEBOOK: A CASE STUDY OF UNBIASED SAMPLING OF OSNS junction.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
Particle Competition and Cooperation for Uncovering Network Overlap Community Structure Fabricio A. Breve 1, Liang Zhao 1, Marcos G. Quiles 2, Witold Pedrycz.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
1 N -Queens via Relaxation Labeling Ilana Koreh ( ) Luba Rashkovsky ( )
Uncovering Overlap Community Structure in Complex Networks using Particle Competition Fabricio A. Liang
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Particle Competition and Cooperation in Networks for Semi-Supervised Learning with Concept Drift Fabricio Breve 1,2 Liang Zhao 2
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Click to edit Master subtitle style 2/23/10 Time and Space Optimization of Document Content Classifiers Dawei Yin, Henry S. Baird, and Chang An Computer.
+ DO NOW. + Chapter 8 Estimating with Confidence 8.1Confidence Intervals: The Basics 8.2Estimating a Population Proportion 8.3Estimating a Population.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Authors :
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Classification Ensemble Methods 1
Computer Science 1 Using Clustering Information for Sensor Network Localization Haowen Chan, Mark Luk, and Adrian Perrig Carnegie Mellon University
Biologically Inspired Computation Ant Colony Optimisation.
Query Rules Study on Active Semi-Supervised Learning using Particle Competition and Cooperation Fabricio Department of Statistics,
Particle Swarm Optimization (PSO)
CS Machine Learning Instance Based Learning (Adapted from various sources)
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Semi-Supervised Learning Jing xu. Slides From: Xiaojin (Jerry) Zhu ---An associate professor in the Department of Computer Sciences at University of Wisconsin-Madison.
Machine Learning: Ensemble Methods
Chapter 5 Unsupervised learning
Self-Organizing Network Model (SOM) Session 11
Neural Network Architecture Session 2
k-Nearest neighbors and decision tree
Whale Optimization Algorithm
Done Done Course Overview What is AI? What are the Major Challenges?
COMBINED UNSUPERVISED AND SEMI-SUPERVISED LEARNING FOR DATA CLASSIFICATION Fabricio Aparecido Breve, Daniel Carlos Guimarães Pedronette State University.
Fabricio Breve Department of Electrical & Computer Engineering
IEEE World Conference on Computational Intelligence – WCCI 2010
The Alpha-Beta Procedure
Problem definition Given a data set X = {x1, x2, …, xn} Rm and a label set L = {1, 2, … , c}, some samples xi are labeled as yi  L and some are unlabeled.
Spectrum Sharing in Cognitive Radio Networks
Presentation transcript:

Combined Active and Semi-Supervised Learning using Particle Walking Temporal Dynamics Fabricio Department of Statistics, Applied Mathematics and Computation (DEMAC), Institute of Geosciences and Exact Sciences (IGCE), São Paulo State University (UNESP), Rio Claro, SP, Brazil 1st BRICS Countries Congress (BRICS-CCI) and 11th Brazilian Congress (CBIC) on Computational Intelligence

Outline Active Learning and Semi-Supervised Learning The Proposed Method Computer Simulations Conclusions

Active Learning Learner is able to interactively query an human specialist (or some other information source) to obtain the labels of selected data points Key idea: greater accuracy with fewer labeled data points [4] B. Settles, “Active learning,” Synthesis Lectures on Artificial Intelligence and Machine Learning, vol. 6, no. 1, pp. 1–114, [5] F. Olsson, “A literature survey of active machine learning in the context of natural language processing,” Swedish Institute of Computer Science, Box 1263, SE Kista, Sweden, Tech. Rep. T2009:06, April 2009.

Semi-Supervised Learning Learns from both labeled and unlabeled data items.  Focus on problems where there are lots of easily acquired unlabeled data, but the labeling process is expensive, time consuming, and often requiring the work of human specialists. [1] X. Zhu, “Semi-supervised learning literature survey,” Computer Sciences, University of Wisconsin-Madison, Tech. Rep. 1530, [2] O. Chapelle, B. Schölkopf, and A. Zien, Eds., Semi-Supervised Learning, ser. Adaptive Computation and Machine Learning. Cambridge, MA: The MIT Press, [3] S. Abney, Semisupervised Learning for Computational Linguistics. CRC Press, 2008.

Semi-Supervised Learning and Active Learning comparison Semi-Supervised Learning Exploits what the learner thinks it knows about the unlabeled data Most confident labeled data used to retrain algorithm (self-learning methods) Relies on committee agreements (co-training methods) Active Learning Attempt to explore unknown aspects of the data Less confident labeled data have their labels queried (uncertainty sampling methods) Query according to committee disagreements (query by committee methods) [4] B. Settles, “Active learning,” Synthesis Lectures on Artificial Intelligence and Machine Learning, vol. 6, no. 1, pp. 1–114, [5] F. Olsson, “A literature survey of active machine learning in the context of natural language processing,” Swedish Institute of Computer Science, Box 1263, SE Kista, Sweden, Tech. Rep. T2009:06, April 2009.

Proposed Method Semi-Supervised Learning and Active Learning combined into a new nature-inspired method  Particles competition and cooperation in networks combined into an unique schema Cooperation:  Particles from the same class (team) walk in the network cooperatively, propagating their labels.  Goal: Dominate as many nodes as possible. Competition:  Particles from different classes (teams) compete against each other  Goal: Avoid invasion by other class particles in their territory

Initial Configuration 4

Nodes have a domination vector  Labeled nodes have ownership set to their respective teams (classes).  Unlabeled nodes have levels set equally for each team Ex: [ ] (4 classes, node labeled as class B) Ex: [ ] (4 classes, unlabeled node)

Node Dynamics When a particle selects a neighbor to visit:  It decreases the domination level of the other teams  It increases the domination level of its own team  Exception: labeled nodes domination levels are fixed

Particle Dynamics A particle gets:  Stronger when it selects a node being dominated by its own team  Weaker when it selects a node being dominated by another team

4? 24 Distance Table Each particle has its distance table. Keep the particle aware of how far it is from the closest labeled node of its team (class).  Prevents the particle from losing all its strength when walking into enemies neighborhoods.  Keeps the particle around to protect its own neighborhood. Updated dynamically with local information.  No prior calculation

Particles Walk Random-greedy walk Each particles randomly chooses a neighbor to visit at each iteration Probabilities of being chosen are higher to neighbors which are: Already dominated by the particle team. Closer to particle initial node.

Moving Probabilities

Particles Walk Shocks  A particle really visits the selected node only if the domination level of its team is higher than others;  Otherwise, a shock happens and the particle stays at the current node until next iteration.

Label Query When the nodes domination levels reach a fair level of stability, the system chooses a unlabeled node and queries its label.  A new particle is created to this new labeled node.  The iterations resume until stability is reached again, then a new node will be chosen.  The process is repeated until the defined amount of labeled nodes is reached.

Query Rule Two versions of the algorithm:  ASL-PCC A  ASL-PCC B They use different rules to select which node will be queried.

ASL-PCC A Uses temporal node domination information to select the unlabeled node which had more dispute over time.  The node the algorithm has less confidence on the label it is currently assigning.

AL-PCC B Chooses the unlabeled node which is currently more far away from any labeled node.  According to particles dynamic distance tables.

Computer Simulations Original PCC method  1% to 10% labeled nodes are randomly chosen. ASL-PCC A and ASL-PCC B  Only 1 labeled node from each class is randomly chosen.  New query each time the system stabilizes. Until it reaches 1% to 10% of labeled nodes.

Correct classification rate comparison when the methods are applied to the Iris data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the Wine data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the Digit1 data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the USPS data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the COIL 2 data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the BCI data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the g241c data set with different amounts of labeled nodes.

Correct classification rate comparison when the methods are applied to the COIL data set with different amounts of labeled nodes.

Conclusions Semi-supervised learning and active learning features combined into a single approach Inspired on the collective behavior of social animals  Protect their territories against intruding groups. No Retraining  New particles are created on the fly as unlabeled nodes become labeled nodes.  The algorithm naturally adapts itself to new situations.  Only nodes affected by the new particles are updated Equilibrium state is quickly reached again Saves execution time.

Conclusions Better classification accuracy than the only semi-supervised learning counterpart when the same amount of labeled data is used.  ASL-PCC A is indicated when: Classes are well separated. Frontiers do not have many outliers.  ASL-PCC B is indicated when: Frontiers are not well defined. There are overlapped regions. There are many outliers.

Combined Active and Semi-Supervised Learning using Particle Walking Temporal Dynamics Fabricio Department of Statistics, Applied Mathematics and Computation (DEMAC), Institute of Geosciences and Exact Sciences (IGCE), São Paulo State University (UNESP), Rio Claro, SP, Brazil 1st BRICS Countries Congress (BRICS-CCI) and 11th Brazilian Congress (CBIC) on Computational Intelligence