Machine Learning 참고 자료 2 Learning Definition Learning is the improvement of performance in some environment through the acquisition of knowledge resulting.

Slides:



Advertisements
Similar presentations
Numbers Treasure Hunt Following each question, click on the answer. If correct, the next page will load with a graphic first – these can be used to check.
Advertisements

Machine learning Overview
Angstrom Care 培苗社 Quadratic Equation II
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
AP STUDY SESSION 2.
1
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 6 Author: Julia Richards and R. Scott Hawley.
Objectives: Generate and describe sequences. Vocabulary:
Fixture Measurements Doug Rytting.
David Burdett May 11, 2004 Package Binding for WS CDL.
Local Customization Chapter 2. Local Customization 2-2 Objectives Customization Considerations Types of Data Elements Location for Locally Defined Data.
Process a Customer Chapter 2. Process a Customer 2-2 Objectives Understand what defines a Customer Learn how to check for an existing Customer Learn how.
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt BlendsDigraphsShort.
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt RhymesMapsMathInsects.
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt Wants.
1 Click here to End Presentation Software: Installation and Updates Internet Download CD release NACIS Updates.
The 5S numbers game..
Photo Slideshow Instructions (delete before presenting or this page will show when slideshow loops) 1.Set PowerPoint to work in Outline. View/Normal click.
1 Combination Symbols A supplement to Greenleafs QR Text Compiled by Samuel Marateck ©2009.
Break Time Remaining 10:00.
1 Machine Learning: Lecture 1 Overview of Machine Learning (Based on Chapter 1 of Mitchell T.., Machine Learning, 1997)
Turing Machines.
Table 12.1: Cash Flows to a Cash and Carry Trading Strategy.
PP Test Review Sections 6-1 to 6-6
EIS Bridge Tool and Staging Tables September 1, 2009 Instructor: Way Poteat Slide: 1.
Bellwork Do the following problem on a ½ sheet of paper and turn in.
Operating Systems Operating Systems - Winter 2010 Chapter 3 – Input/Output Vrije Universiteit Amsterdam.
Exarte Bezoek aan de Mediacampus Bachelor in de grafische en digitale media April 2014.
Copyright © 2013, 2009, 2006 Pearson Education, Inc. 1 Section 5.5 Dividing Polynomials Copyright © 2013, 2009, 2006 Pearson Education, Inc. 1.
Copyright © 2012, Elsevier Inc. All rights Reserved. 1 Chapter 7 Modeling Structure with Blocks.
1 RA III - Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Buenos Aires, Argentina, 25 – 27 October 2006 Status of observing programmes in RA.
Basel-ICU-Journal Challenge18/20/ Basel-ICU-Journal Challenge8/20/2014.
1..
CONTROL VISION Set-up. Step 1 Step 2 Step 3 Step 5 Step 4.
Adding Up In Chunks.
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt Synthetic.
Artificial Intelligence
1 Using Bayesian Network for combining classifiers Leonardo Nogueira Matos Departamento de Computação Universidade Federal de Sergipe.
Subtraction: Adding UP
: 3 00.
5 minutes.
1 hi at no doifpi me be go we of at be do go hi if me no of pi we Inorder Traversal Inorder traversal. n Visit the left subtree. n Visit the node. n Visit.
Analyzing Genes and Genomes
Prof.ir. Klaas H.J. Robers, 14 July Graduation: a process organised by YOU.
Essential Cell Biology
Converting a Fraction to %
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
Clock will move after 1 minute
PSSA Preparation.
Copyright © 2013 Pearson Education, Inc. All rights reserved Chapter 11 Simple Linear Regression.
Essential Cell Biology
Physics for Scientists & Engineers, 3rd Edition
Energy Generation in Mitochondria and Chlorplasts
Select a time to count down from the clock above
1 Dr. Scott Schaefer Least Squares Curves, Rational Representations, Splines and Continuity.
1 Decidability continued…. 2 Theorem: For a recursively enumerable language it is undecidable to determine whether is finite Proof: We will reduce the.
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
1 Some rules  No make-up exams ! If you miss with an official excuse, you get average of your scores in the other exams – at most once.  WP only-if you.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CpSc 810: Machine Learning Design a learning system.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Chapter 1: Introduction. 2 목 차목 차 t Definition and Applications of Machine t Designing a Learning System  Choosing the Training Experience  Choosing.
1 Four Components of a Learning System t Performance system  S olve the given performance task  U se the learned target function  N ew problem -> trace.
1 Introduction to Machine Learning Chapter 1. cont.
Introduction Machine Learning: Chapter 1. Contents Types of learning Applications of machine learning Disciplines related with machine learning Well-posed.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Presentation transcript:

Machine Learning 참고 자료

2 Learning Definition Learning is the improvement of performance in some environment through the acquisition of knowledge resulting from experience in that environment.

3 Machine Learning: Tasks Supervised Learning Learn f w from training set D={(x,y)} s.t. Classification: y is discrete Regression: y is continuous Unsupervised Learning Learn f w from D={(x)} s.t. Density Estimation Compression, Clustering

4 Machine Learning: Methods Symbolic Learning Version Space Learning Neural Learning Multilayer Perceptrons (MLPs) Evolutionary Learning Genetic Algorithms Probabilistic Learning Bayesian Networks (BNs) Other Machine Learning Methods Decision Trees (DTs)

5 Applications of Machine Learning Driving an autonomous vehicle 무인 자동차 운전, 센서기반 제어 등에도 응용 Classifying new astronomical structures 천체 물체 분류, Decision tree learning 기법 사용 Playing world-class Backgammon 실제 게임을 통해서 전략을 학습, 탐색공간 문제에 응용

6 A Definition of Learning : Well-posed Learning Problems Definition A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E. A class of tasks T Experience E Performance measure P

7 Checkers Problem (1/2) 말은 대각선으로만 움직일 수 있다. 맞은편 끝까지 가기 전에는 앞으로만 진행할 수 있다. 대각선에 상대편 말이 있을 경우 그 말을 없앨수 있다. 게임은 한편 말이 모두 없어지면 끝난다.

8 Checkers Problem (2/2) homepage

9 A Checkers Learning Problem Three Features: 학습문제의 정의 The class of tasks The measure of performance to be improved The source of experience Example Task T: playing checkers Performance measure P: percent of games won against opponent Training experience E: playing practice games against itself

10 Designing a Learning System Choosing the Training Experience Choosing the Target Function Choosing a Representation for the Target Function Choosing a Function Approximation Algorithm

11 Choosing the Training Experience (1/2) Key Attributes Direct/indirect feedback Direct feedback: checkers state and correct move Indirect feedback: move sequence and final outcomes Degree of controlling the sequence of training example Learner 가 학습 정보를 얻을 때 teacher 의 도움을 받는 정도

12 Choosing the Training Experience (2/2) Distribution of examples 시스템의 성능을 평가하는 테스트의 예제 분포 를 잘 반영해야 함

13 Choosing the Target Function (1/2) A function that chooses the best move M for any B ChooseMove : B  M Difficult to learn It is useful to reduce the problem of improving performance P at task T to the problem of learning some particular target function.

14 Choosing the Target Function (2/2) An evaluation function that assigns a numerical score to any B V : B  R

15 Target Function for the Checkers Problem Algorithm If b is a final state that is won, then V(b) = 100 ……. that is lost, then V(b)=-100 ……. that is drawn, then V(b)=0 If b is not a final state, then V(b)=V(b’), where b’ is the best final board state

16 Choosing a Representation for the Target Function Describing the function Tables Rules Polynomial functions Neural nets Trade-off in choice Expressive power Size of training data

17 Linear Combination as Representation (b) = w 0 + w 1 x 1 + w 2 x 2 + w 3 x 3 +w 4 x 4 + w 5 x 5 + w 6 x 6 x 1 : # of black pieces on the board x 2 : # of red pieces on the board x 3 : # of black kings on the board x 4 : # of red kings on the board x 5 : # of black pieces threatened by red x 6 : # of red pieces threatened by black w 1 - w 6 : weights

18 Partial Design of a Checkers Learning Program Task T: playing checkers Performance measure P: Percent of games won in the world tournament Training experience E: games played against itself Target function V: Board  R Target function representation (b) = w 0 + w 1 x 1 + w 2 x 2 + w 3 x 3 + w 4 x 4 + w 5 x 5 + w 6 x 6

19 Choosing a Function Approximation Algorithm A training example is represented as an ordered pair b: board state V train (b): training value for b Instance: “black has won the game (x 2 = 0), +100>

20 Choosing a Function Approximation Algorithm Estimating training values for intermediate board states V train (b)  (Successor(b)) : current approximation to V Successor(b): the next board state

21 Adjusting the Weights (1/2) Choosing w i to best fit the training examples Minimize the squared error

22 Adjusting the Weights (2/2) LMS Weight Update Rule For each training example 1. Use the current weights to calculate V’(b) 2. For each weight w i, update it as

23 Sequence of Design Choices Determine Type of Training Experience Determine Target Function Determine Representation Of Learned Function Determine Learning Algorithm Table of correct moves Games against experts Games against self Board  move Board  value Polynomial Linear function of six features Arfiticial NN Gradient descent Complete Design Linear Programming

24 Perspectives in ML “Learning as search in a space of possible hypotheses” Representations for hypotheses Linear functions Logical descriptions Decision trees Neural networks

25 Perspectives in ML Learning methods are characterized by their search strategies and by the underlying structure of the search spaces.

26 Summary 기계학습은 다양한 응용분야에서 실용적 가치 가 크다. 많은 데이터로부터 규칙성을 발견하는 문제 (data mining) 문제의 성격 규명이 어려워 효과적인 알고리즘을 개발할 지식이 없는 문제 영역 (human face recognition) 변화하는 환경에 동적으로 적응하여야 하는 문제 영역 (manufacturing process control)

27 Summary 기계학습은 다양한 다른 학문 분야와 밀접히 관련된다. 인공지능, 확률통계, 정보이론, 계산이론, 심리학, 신경과학, 제어이론, 철학 잘 정의된 학습 문제는 다음을 요구한다. 문제 (task) 의 명확한 기술, 성능평가 기준, 훈련경험 을 위한 사례

28 Summary 기계학습 시스템의 설계 시에는 다음 사항을 고려 하여야 한다. 훈련경험의 유형 선택 학습할 목표함수 목표함수에 대한 표현 훈련 예로부터 목표함수를 학습하기 위한 알고리 즘

29 Summary 학습은 가능한 가설 공간에서 주어진 훈련 예 와 다른 배경지식을 가장 잘 반영하는 하나의 가설을 탐색하는 탐색이다. 다양한 학습 방법은 서로 다른 가설공간의 형태와 이 공간 내에서 탐색을 수행하는 전략에 의해 규정 지어진다.

Neural Networks

31 Biological motivation Neuron receives signals from other neurons through its dendrites Transmits signals generated by its cell body along the axon Network of Neuron

32 Neural Network Representations The primitive unit(e.g. perceptron) N input signals  weighted sum  threshold function  generate an output A learning process in the ANN Learning process involves choosing values for the weights w 0, …, w n Learning rules How network weights are updated?

33 Gradient descent and the delta rule The delta rule Linear unit for which the output o is given by Measure for the training error of a hypothesis d : the set of traing examples t d : the target output for training example d o d : the output of the linear unit for training example d We can characterize E as a function of

34 Gradient descent and the delta rule

35 Gradient descent and the delta rule Derivation of the gradient descent rule Direction of steepest descent along the error space Derivative E with respect to each component of The negative of this vector therefore gives the direction of steepest decrease

36 Gradient descent and the delta rule Training rule for gradient descent w i ← w i +  w i where, Efficient way of calculating the gradient So,

37 Gradient descent and the delta rule If  is too large, the gradient descent search runs the risk of overstepping the minimum gradually reduce the value of 

38 Multilayer Networks Why multilayer network? Single perceptrons can only express linear decision surfaces So, add an extra(hidden) layer between the inputs and outputs E.g.) the speech recognition task

39 Multilayer Networks Sigmoid function

40 E defined as a sum of the squared errors over all the output units k for all the training examples d. Error Function for BP

41 BP Algorithm

42 After a fixed number of iterations (epochs) Once the error falls below some threshold Once the validation error meets some criterion Learning Until…

Self Organizing Map

44 Introduction Unsupervised Learning SOM (Self Organizing Map) Visualization Abstraction

45 SOM structures Neighborhood Input Layer Output Layer

46 Data to be clustered

47 After 100 iterations

48 After 500 iterations

49 After 2000 iterations

50 After iterations