Kansas State University Department of Computing and Information Sciences Real-Time Bayesian Network Inference for Decision Support in Personnel Management:

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Prof. Carolina Ruiz Computer Science Department Bioinformatics and Computational Biology Program WPI WELCOME TO BCB4003/CS4803 BCB503/CS583 BIOLOGICAL.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Review: Bayesian learning and inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, March 6, 2000 William.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
CS157A Spring 05 Data Mining Professor Sin-Min Lee.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
Computing & Information Sciences Kansas State University Lecture 28 of 42 CIS 530 / 730 Artificial Intelligence Lecture 28 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, March 1, 2000 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, March 27, 2000 William.
Computing & Information Sciences Kansas State University Lecture 27 of 42 CIS 530 / 730 Artificial Intelligence Lecture 27 of 42 William H. Hsu Department.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Kansas State University Department of Computing and Information Sciences Kansas State University KDD Lab ( Permutation.
Kansas State University Department of Computing and Information Sciences Kansas State University KDD Lab ( cDNA.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Tuesday 08 October 2002 William.
Kansas State University Department of Computing and Information Sciences Ben Perry – M.S. thesis defense Benjamin B. Perry Laboratory for Knowledge Discovery.
Aprendizagem Computacional Gladys Castillo, UA Bayesian Networks Classifiers Gladys Castillo University of Aveiro.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, October 7, 1999 William.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, March 15, 2000.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, March 10, 2000 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Classification Techniques: Bayesian Classification
Estimating Component Availability by Dempster-Shafer Belief Networks Estimating Component Availability by Dempster-Shafer Belief Networks Lan Guo Lane.
LEAP-KMC Workshop 2006 Visualization of KMC Simulation Data and Evolutionary Computation: The LEAP Infrastructure and Content Management System William.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 24, 2001.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, December 7, 1999 William.
Computing & Information Sciences Kansas State University Data Sciences Summer Institute Multimodal Information Access and Synthesis Learning and Reasoning.
Computing & Information Sciences Kansas State University Monday, 06 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 31 of 42 Monday, 06 November.
Chapter 6 Bayesian Learning
K2 Algorithm Presentation KDD Lab, CIS Department, KSU
Kansas State University Department of Computing and Information Sciences CIS 690: Data Mining Systems Lab 0 Monday, May 15, 2000 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 14 November 2003 William.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 27 of 42 Friday, 27 October.
Quiz 3: Mean: 9.2 Median: 9.75 Go over problem 1.
Data Mining and Decision Support
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Graphical Models of Probability.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Monday, 08 November 2004 William.
1 Machine Learning: Lecture 6 Bayesian Learning (Based on Chapter 6 of Mitchell T.., Machine Learning, 1997)
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 24 of 41 Monday, 18 October.
Bayesian Learning Bayes Theorem MAP, ML hypotheses MAP learners
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, March 8, 2000 Jincheng.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
School of Computer Science & Engineering
Spring 2003 Dr. Susan Bridges
Data Mining Lecture 11.
Classification Techniques: Bayesian Classification
CSCI 5822 Probabilistic Models of Human and Machine Learning
Data Warehousing and Data Mining
Overview of Machine Learning
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Chapter 14 February 26, 2004.
Presentation transcript:

Kansas State University Department of Computing and Information Sciences Real-Time Bayesian Network Inference for Decision Support in Personnel Management: Report on Research Activities William H. Hsu, Computing and Information Sciences Haipeng Guo, Computing and Information Sciences Shing I Chang, Industrial and Manufacturing Systems Engineering Kansas State University This presentation is:

Kansas State University Department of Computing and Information Sciences Overview Knowledge Discovery in Databases (KDD) –Towards scalable data mining –Applications of KDD: learning and reasoning Building Causal Models for Decision Support Time Series and Model Integration –Prognostic (prediction and monitoring) applications –Crisis monitoring and simulation –Anomaly, intrusion, fraud detection –Web log analysis –Applying high-performance neural, genetic, Bayesian computation Information Retrieval: Document Categorization, Text Mining –Business intelligence applications (e.g., patents) –“Web mining”: dynamic indexing and document analysis High-Performance KDD Program at K-State

Kansas State University Department of Computing and Information Sciences High-Performance Database Mining and KDD: Current Research Programs at K-State Laboratory for Knowledge Discovery in Databases (KDD) –Research emphases: machine learning, reasoning under uncertainty –Applications Decision support Digital libraries and information retrieval Remote sensing, robot vision and control Human-Computer Interaction (HCI) - e.g., simulation-based training Computational science and engineering (CSE) Curriculum and Research Development –Real-time automated reasoning (inference) –Machine learning –Probabilistic models for multi-objective optimization –Intelligent displays: visualization of diagrammatic models –Knowledge-based expert systems, data modeling for KDD

Kansas State University Department of Computing and Information Sciences Stages of Data Mining and Knowledge Discovery in Databases

Kansas State University Department of Computing and Information Sciences Visual Programming: Java-Based Software Development Platform D2K © 2002 National Center for Supercomputing Applications (NCSA) Used with permission.

Kansas State University Department of Computing and Information Sciences X1X1 X2X2 X3X3 X4X4 Season: Spring Summer Fall Winter Sprinkler: On, Off Rain: None, Drizzle, Steady, Downpour Ground: Wet, Dry X5X5 Ground: Slippery, Not-Slippery P(Summer, Off, Drizzle, Wet, Not-Slippery) = P(S) · P(O | S) · P(D | S) · P(W | O, D) · P(N | W) Conditional Independence –X is conditionally independent (CI) from Y given Z (sometimes written X  Y | Z) iff P(X | Y, Z) = P(X | Z) for all values of X, Y, and Z –Example: P(Thunder | Rain, Lightning) = P(Thunder | Lightning)  T  R | L Bayesian Network –Directed graph model of conditional dependence assertions (or CI assumptions) –Vertices (nodes): denote events (each a random variable) –Edges (arcs, links): denote conditional dependencies General Product (Chain) Rule for BBNs Example (“Sprinkler” BBN) Bayesian Belief Networks (BBNS): Definition

Kansas State University Department of Computing and Information Sciences Bayesian Networks and Recommender Systems Current Research –Efficient BBN inference (parallel, multi-threaded Lauritzen-Spiegelhalter in D2K) –Hybrid quantitative and qualitative inference (“simulation”) –Continuous variables and hybrid (discrete/continuous) BBNs –Induction of hidden variables –Local structure: localized constraints and assumptions, e.g., Noisy-OR BBNs –Online learning Incrementality (aka lifelong, situated, in vivo learning) Ability to change network structure during inferential process –Polytree structure learning (tree decomposition): alternatives to Chow-Liu –Complexity of learning, inference in restricted classes of BBNs Future Work –Decision networks aka influence diagrams (BBN + utility) –Anytime / real-time BBN inference for time-constrained decision support –Some temporal models: Dynamic Bayesian Networks (DBNs)

Kansas State University Department of Computing and Information Sciences Data Mining: Development Cycle Model Identification –Queries: classification, assignment –Specification of data model –Grouping of attributes by type Prediction Objective Identification –Assignment specification –Identification of metrics Reduction –Refinement of data model –Selection of relevant data (quantitative, qualitative) Synthesis: New Attributes Integration: Multiple Data Sources (e.g., Enlisted Master File, Surveys) Environment (Data Model) Learning Element Knowledge Base Decision Support System

Kansas State University Department of Computing and Information Sciences Learning Bayesian Networks: Gradient Ascent Algorithm Train-BN (D) –Let w ijk denote one entry in the CPT for variable Y i in the network w ijk = P(Y i = y ij | parents(Y i ) = ) e.g., if Y i  Campfire, then (for example) u ik  –WHILE termination condition not met DO// perform gradient ascent Update all CPT entries w ijk using training data D Renormalize w ijk to assure invariants: Applying Train-BN –Learns CPT values –Useful in case of known structure –Key problems: learning structure from data, approximate inference Bus TourGroup Storm LightningCampfire ForestFireThunder

Kansas State University Department of Computing and Information Sciences General-Case BBN Structure Learning: Use Inference to Compute Scores Recall: Bayesian Inference aka Bayesian Reasoning –Assumption: h  H are mutually exclusive and exhaustive –Optimal strategy: combine predictions of hypotheses in proportion to likelihood Compute conditional probability of hypothesis h given observed data D i.e., compute expectation over unknown h for unseen cases Let h  structure, parameters   CPTs Scores for Learning Structure: The Role of Inference Posterior ScoreMarginal Likelihood Prior over StructuresLikelihood Prior over Parameters

Kansas State University Department of Computing and Information Sciences Learning Structure: K2 Algorithm and ALARM Algorithm Learn-BBN-Structure-K2 (D, Max-Parents) FOR i  1 to n DO// arbitrary ordering of variables {x 1, x 2, …, x n } WHILE (Parents[x i ].Size < Max-Parents) DO// find best candidate parent Best  argmax j>i (P(D | x j  Parents[x i ])// max Dirichlet score IF (Parents[x i ] + Best).Score > Parents[x i ].Score) THEN Parents[x i ] += Best RETURN ({Parents [x i ] | i  {1, 2, …, n}}) A Logical Alarm Reduction Mechanism [Beinlich et al, 1989] –BBN model for patient monitoring in surgical anesthesia –Vertices (37): findings (e.g., esophageal intubation), intermediates, observables –K2: found BBN different in only 1 edge from gold standard (elicited from expert)

Kansas State University Department of Computing and Information Sciences Major Software Releases, FY 2002 Bayesian Network Tools in Java (BNJ) –v1.0a released Wed 08 May 2002 to –Key features Standardized data format (XML) Existing algorithms: inference, structure learning, data generation –Experimental results Improved structure learning using K2, inference-based validation Adaptive importance sampling (AIS) inference competitive with best published algorithms Machine Learning in Java (MLJ) –v1.0a released Fri 10 May 2002 to –Key features: (3) inductive learning algorithms from MLC++, (2) inductive learning wrappers (1 from MLC++, 1 from GA literature) –Experimental results Genetic wrappers for feature subset selection: Jenesis, MLJ-CHC Overfitting control in supervised inductive learning for classification

Kansas State University Department of Computing and Information Sciences About BNJ –v1.0a, 08 May 2002: lines of Java code, GNU Public License (GPL) – –Key features [Perry, Stilson, Guo, Hsu, 2002] XML BN Interchange Format (XBN) converter – to serve 7 client formats (MSBN, Hugin, SPI, IDEAL, Ergo, TETRAD, Bayesware) Full exact inference: Lauritzen-Spiegelhalter (Hugin) algorithm Five (5) importance sampling algorithms: forward simulation (likelihood weighting) [Shachter and Peot, 1990], probabilistic logic sampling [Henrion, 1986], backward sampling [Fung and del Favero, 1995] self- importance sampling [Shachter and Peot, 1990], adaptive importance sampling [Cheng and Druzdzel, 2000] Data generator Published Research with Applications to Personnel Science –Recent work GA for improved structure learning: results in [HGPS02a; HGPS02b] Real-time inference framework – multifractal analysis [GH02b] –Current work: prediction – migration trends (EMF); Sparse Candidate –Planned continuation: (dynamic) decision networks; continuous BNs Bayesian Network Tools in Java (BNJ)

Kansas State University Department of Computing and Information Sciences Change of Representation and Inductive Bias Control [B] Representation Evaluator for Learning Problems D: Training Data : Inference Specification D train (Inductive Learning) D val (Inference) [A] Genetic Algorithm Optimized Representation α Candidate Representation f(α) Representation Fitness GA for BN Structure Learning [Hsu, Guo, Perry, Stilson, GECCO-2002]

Kansas State University Department of Computing and Information Sciences [B] Representation Evaluator for Input Specifications : Evidence Specification D train (Model Training) D val (Model Validation by Inference) f(α) Specification Fitness (Inferential Loss) [ii] Validation (Measurement of Inferential Loss) h Hypothesis [i] Inductive Learning (Parameter Estimation from Training Data) α Candidate Input Specification Model-Based Validation [Hsu, Guo, Perry, Stilson, GECCO-2002]

Kansas State University Department of Computing and Information Sciences BNJ: Integrated Tool for Bayesian Network Learning and Inference XML Bayesian Network Learned from Data using K2 in BNJ

Kansas State University Department of Computing and Information Sciences About MLJ –v1.0a, 10 May 2002: lines of Java code, GNU Public License (GPL) – –Key features [Hsu, Schmidt, Louis, 2002] Conformant to MLC++ input-output specification Three (3) inductive learning algorithms: ID3, C4.5, discrete Naïve Bayes Two (2) wrapper inducers: feature subset selection [Kohavi and John, 1997], CHC [Eshelman, 1990; Guerra-Salcedo and Whitley, 1999] Published Research with Applications to Personnel Science –Recent work Multi-agent learning [GH01, GH02a] Genetic feature selection wrappers [HSL02, HWRC02, HS02] –Current work: WEKA compatibility, parallel online continuous arcing –Planned continuations New inducers: instance-based (k-nearest-neighbor), sequential rule covering, feedforward artificial neural network (multi-layer perceptron) New wrappers: theory-guided constructive induction, boosting (Arc-x4, AdaBoost.M1, POCA) Integration of reinforcement learning (RL) inducers Machine Learning in Java (MLJ)

Kansas State University Department of Computing and Information Sciences Infrastructure for High-Performance Computation in Data Mining Rapid KDD Development Environment: Operational Overview

Kansas State University Department of Computing and Information Sciences National Center for Supercomputing Applications (NCSA) D2K

Kansas State University Department of Computing and Information Sciences Visual Programming Interface (Java): Parallel Genetic Algorithms

Kansas State University Department of Computing and Information Sciences Time Series Modeling and Prediction: Integration with Information Visualization New Time Series Visualization System (Java3D)

Kansas State University Department of Computing and Information Sciences Demographics-Based Clustering for Prediction (Continuing Research) Cluster Formation and Segmentation Algorithm (Sketch) Dimensionality- Reducing Projection (x’) Clusters of Similar Records Delaunay Triangulation Voronoi (Nearest Neighbor) Diagram (y)

Kansas State University Department of Computing and Information Sciences Data Clustering in Interactive Real-Time Decision Support 15 × 15 Self-Organizing Map (U-Matrix Output) Cluster Map (Personnel Database)

Kansas State University Department of Computing and Information Sciences Laboratory for Knowledge Discovery in Databases (KDD) –Applications: interdisciplinary research programs at K-State, FY 2002 Decision support, optimization (Hsu, CIS; Chang, IMSE) (NSF EPSCoR) Bioinformatics – gene expression modeling (Hsu, CIS; Welch, Agronomy; Roe, Biology; Das, EECE) Digital libraries, info retrieval (Hsu, CIS; Zollman, Physics; Math, Art) Human-Computer Interaction (HCI) - e.g., simulation-based training Curriculum Development –Real-time intelligent systems (Chang, Hsu, Neilsen, Singh) –Machine learning and artificial intelligence; info visualization (Hsu) –Other: bioinformatics, digital libraries, robotics, DBMS Research Partnerships –NCSA: National Computational Science Alliance, National Center for Supercomputing Applications –Defense (ONR, ARL, DARPA), Industry (Raytheon) Publications, More Info: Summary: State of High-Performance KDD at KSU-CIS