Implementation in C+CUDA of Multi-Label Text Categorizers In automated multi-label text categorization problems with large numbers of labels, the training.

Slides:



Advertisements
Similar presentations
An Introduction To Categorization Soam Acharya, PhD 1/15/2003.
Advertisements

Text Categorization.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Data Mining Classification: Alternative Techniques
ECE 8527 Homework Final: Common Evaluations By Andrew Powell.
A segmented principal component analysis applied to calorimetry information at ATLAS ACAT May 22-27, Zeuthen, Germany H. P. Lima Jr, J. M. de Seixas.
Learning for Text Categorization
2009/04/07 Yun-Yang Ma.  Overview  What is CUDA ◦ Architecture ◦ Programming Model ◦ Memory Model  H.264 Motion Estimation on CUDA ◦ Method ◦ Experimental.
1 Application of Metamorphic Testing to Supervised Classifiers Xiaoyuan Xie, Tsong Yueh Chen Swinburne University of Technology Christian Murphy, Gail.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
Gimme’ The Context: Context- driven Automatic Semantic Annotation with CPANKOW Philipp Cimiano et al.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ApMl (All Purpose Machine Learning) Toolkit David W. Miller and Helen Howell Semantic Web Final Project Spring 2002 Department of Computer Science University.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
Instance-Based Learners So far, the learning methods that we have seen all create a model based on a given representation and a training set. Once the.
Creating and Visualizing Document Classification J. Gelernter, D. Cao, R. Lu, E. Fink, J. Carbonell.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
A k-Nearest Neighbor Based Algorithm for Multi-Label Classification Min-Ling Zhang
AN OPTIMIZED APPROACH FOR kNN TEXT CATEGORIZATION USING P-TREES Imad Rahal and William Perrizo Computer Science Department North Dakota State University.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
We introduce the use of Confidence c as a weighted vote for the voting machine to avoid low confidence Result r of individual expert from affecting the.
Appendix B: An Example of Back-propagation algorithm
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
Personalized Web Search by Mapping User Queries to Categories Fang Liu Presented by Jing Zhang CS491CXZ February 26, 2004.
NEURAL NETWORKS FOR DATA MINING
Eric H. Huang, Richard Socher, Christopher D. Manning, Andrew Y. Ng Computer Science Department, Stanford University, Stanford, CA 94305, USA ImprovingWord.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Diane Marinkas CDA 6938 April 30, Outline Motivation Algorithm CPU Implementation GPU Implementation Performance Lessons Learned Future Work.
An informal description of artificial neural networks John MacCormick.
Special topics on text mining [ Part I: text classification ] Hugo Jair Escalante, Aurelio Lopez, Manuel Montes and Luis Villaseñor.
HPCLatAm 2013 HPCLatAm 2013 Permutation Index and GPU to Solve efficiently Many Queries AUTORES  Mariela Lopresti  Natalia Miranda  Fabiana Piccoli.
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
The Perceptron. Perceptron Pattern Classification One of the purposes that neural networks are used for is pattern classification. Once the neural network.
Project 1: Machine Learning Using Neural Networks Ver 1.1.
Fast Support Vector Machine Training and Classification on Graphics Processors Bryan Catanzaro Narayanan Sundaram Kurt Keutzer Parallel Computing Laboratory,
Map-Reduce for Machine Learning on Multicore C. Chu, S.K. Kim, Y. Lin, Y.Y. Yu, G. Bradski, A.Y. Ng, K. Olukotun (NIPS 2006) Shimin Chen Big Data Reading.
Work in Progress Presented by Ian Wood to CASCI, 4/24/13 Advisor: Luis Rocha.
A Neural Network Implementation on the GPU By Sean M. O’Connell CSC 7333 Spring 2008.
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Citation-Based Retrieval for Scholarly Publications 指導教授:郭建明 學生:蘇文正 M
Xiangnan Kong,Philip S. Yu An Ensemble-based Approach to Fast Classification of Multi-label Data Streams Dept. of Computer Science University of Illinois.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Essential components of the implementation are:  Formation of the network and weight initialization routine  Pixel analysis of images for symbol detection.
Implementation of a Biologically Inspired Stereoscopic Vision Model in C+CUDA Most of the depth perception processing is done in the visual cortex, mainly.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
How do you get here?
11 Automated multi-label text categorization with VG-RAM weightless neural networks Presenter: Guan-Yu Chen A. F. DeSouza, F. Pedroni, E. Oliveira, P.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Data Mining and Text Mining. The Standard Data Mining process.
Agglomerative clustering (AC)
Julián ALARTE DAVID INSA JOSEP SILVA
Neural Network Implementations on Parallel Architectures
Efficient Image Classification on Vertically Decomposed Data
A Deep Learning Technical Paper Recommender System
Genomic Data Clustering on FPGAs for Compression
Creating fuzzy rules from numerical data using a neural network
Classification with Perceptrons Reading:
Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan.
Efficient Image Classification on Vertically Decomposed Data
Text Categorization Assigning documents to a fixed set of categories
Face Recognition with Neural Networks
Word Embedding Word2Vec.
Presentation transcript:

Implementation in C+CUDA of Multi-Label Text Categorizers In automated multi-label text categorization problems with large numbers of labels, the training databases are large, which may render the categorization time prohibitive for online systems. In this work, we evaluate the parallel implementation in C+CUDA of two multi-label text categorizers: the first is based on the k-Nearest Neighbors (k-NN) algorithm [1] and the second is based on Probabilistic Neural Networks (PNN) [2]. We implemented these algorithms in three different ways: sequential in C, parallel in C+CUDA, and parallel using the C+CUBLAS library. where N k is the number of neurons of the pattern layer associated to ck. The categories c k ranked above a threshold are predicted to the input document d x. Lucas Veronese, Alberto F. De Souza, Claudine Badue, Elias Oliveira, Patrick M. Ciarelli Departamento de Informática – Laboratório de Computação de Alto Desempenho Universidade Federal do Espírito Santo, Av. F. Ferrari 514, Vitória-ES, Brazil {lucas.veronese, alberto, claudine, elias, pciarelli The PNN used in this work was proposed by Oliveira et. al [2] and is composed of two feed-forward layers: pattern layer and summation layer. In the training phase, for each document d i is created a set of neurons, one for each category, where each neuron n i stores the vector d i as a vector of term weights, w k,i. In the categorization phase, an input document d x is presented to the pattern layer. The i -th neuron, n i, associated to category c k in the pattern layer, calculates the activation function for document d x given by: [1] F. Sebastiani, “Machine Learning in Automated Text Categorization”, ACM Computing Surveys 34(1), 2002, pp [2] E. Oliveira, P. M. Ciarelli, A. F. De Souza, and C. Badue. Using a Probabilistic Neural Network for a Large Multi-Label Problem. Proceedings of the 10th Brazilian Symposium on Neural Networks (SBRN'08), pp , Salvador, Bahia, Brazil, October BibliographyBibliography IntroductionIntroduction k-Nearest Neighbors (k-NN) The k -NN categorizer finds the k nearest neighbors of an input document d x in the set of previously learned documents, TV, according to some given distance metric. We used the cosine of the angle between the floating-point vector that represents the input document d x (bag-of-words document representation [1]) and each document, cos(d x,d i ) : The k-NN categorizer (i) employs a function f(d x,c k ) that returns the highest value of cos(d x,d i ) for and, where C i is the set of pertinent categories for the document d i, and (ii) selects the k pairs from the top of the ranking derived from. Probabilistic Neural Network (PNN) k=1,..., |C|, i=1, …, |D k | where is a constant for all neurons (adjusted during training for best categorization performance [2]), C is the whole set of possible categories, and D k is the set of documents associated to category c k. In the summation layer, which has as many neurons as | C |, each neuron is associated with a category c k and computes the function f(d x,c k ): k=1,..., |C| Categ.C (s) C+CUDA (s) C+CUBLAS (s) Speed-up C+CUDA Speed-up C+CUBLAS k-NN PNN We ran the C, C+CUDA and C+CUBLAS versions of our categorizers in an AMD Athlon 64 X2 (Dual Core) 5,200+ of 2.7 GHz, with 3GB of 800 MHz DRAM DDR2, and video card NVIDIA GeForce GTX 285, with 1GB of DRAM GDDR3. The data set used is composed of 6,911 documents categorized into 105 different categories by specialists in the domain of the documents. Each one of these categories occurs in exactly 100 different documents, i.e., there are 100 documents of each category. Each document is represented by a vector of single precision floats of size 3,764 (the number of relevant terms in the system vocabulary). Experimental Setup To evaluate the performance of our categorizers in terms of time, we selected 6,910 documents of the data set for training, and a single one for testing the categorizers. Each categorizer was executed 100 times and the average was used to compare them. Table 1 shows the average times for each categorizer (rows) and categorizer implementation (columns), in addition to the speed-ups over the sequential implementation (last two columns). As the table shows, we achieved speed-ups of about 60 for the C+CUDA version and about 45 for the C+CUBLAS version. These results show that, with CUDA, it is possible to implement on-line text categorization and that, in some cases, it is worth implementing the whole code instead of using C+CUBLAS. ResultsResults dxdx w 1,1 w |c1|,i … … f(dx,c1)f(dx,c1) f(dx,ck)f(dx,ck) pattern layer summation layer w |ck|,1 w |ck|,i