Advanced Higher Computing Based on Heriot-Watt University Scholar Materials Applications of AI – Vision and Languages 1.

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

Review of AI from Chapter 3. Journal May 13  What advantages and disadvantages do you see with using Expert Systems in real world applications like business,
What Makes Up a PC?. Objectives At the end of this session, you will be able to: List the various input devices.
Speech Recognition Part 3 Back end processing. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
Grammars.
1 SIMS 290-2: Applied Natural Language Processing Marti Hearst Sept 20, 2004.
CS 561, Sessions 27 1 Towards intelligent machines Thanks to CSCI561, we now know how to… - Search (and play games) - Build a knowledge base using FOL.
Processing Digital Images. Filtering Analysis –Recognition Transmission.
Natural Language Processing AI - Weeks 19 & 20 Natural Language Processing Lee McCluskey, room 2/07
CSE (c) S. Tanimoto, 2007 Segmentation and Labeling 1 Segmentation and Labeling Outline: Edge detection Chain coding of curves Segmentation into.
November 29, 2004AI: Chapter 24: Perception1 Artificial Intelligence Chapter 24: Perception Michael Scherger Department of Computer Science Kent State.
Natural Language Understanding
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Recap Low Level Vision –Input: pixel values from the imaging device –Data structure: 2D array, homogeneous –Processing: 2D neighborhood operations Histogram.
Data Handling Topic 1. Data Data is the raw material entered into a computer system. This raw material could be text, numbers, graphics, audio, animation.
9/8/20151 Natural Language Processing Lecture Notes 1.
Grammars.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Shape from Contour Recover 3D information. Muller-Lyer illusion.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
Introduction to CL & NLP CMSC April 1, 2003.
Graphical Models in Vision. Alan L. Yuille. UCLA. Dept. Statistics.
MACHINE VISION Machine Vision System Components ENT 273 Ms. HEMA C.R. Lecture 1.
3D Shape Inference Computer Vision No.2-1. Pinhole Camera Model the camera center Principal axis the image plane.
Computational linguistics A brief overview. Computational Linguistics might be considered as a synonym of automatic processing of natural language, since.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
Grammars Grammars can get quite complex, but are essential. Syntax: the form of the text that is valid Semantics: the meaning of the form – Sometimes semantics.
Artificial Intelligence: Natural Language
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
Higher Vision, language and movement. Strong AI Is the belief that AI will eventually lead to the development of an autonomous intelligent machine. Some.
Parsing and Code Generation Set 24. Parser Construction Most of the work involved in constructing a parser is carried out automatically by a program,
End of Key Stage 1 Statutory Assessments Teacher Assessment End of KS1 assessment will be based on judgements made by teachers throughout the key.
Natural Language Processing (NLP)
CIS Intro to AI 1 Interpreting Line Drawings (An Introduction to Constraint Satisfaction) Mitch Marcus CIS 391 Fall 2008.
Perception Problems Vision understanding, natural language processing and speech recognition all have several things in common –each problem is so complex.
NATURAL LANGUAGE PROCESSING
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
IIS for Speech Processing Michael J. Watts
Visual Information Processing. Human Perception V.S. Machine Perception  Human perception: pictorial information improvement for human interpretation.
Year 2 Stay and Play!.
Machine Vision ENT 273 Lecture 4 Hema C.R.
An Information Evening for Parents
BCT 2083 DISCRETE STRUCTURE AND APPLICATIONS
Basic Parsing with Context Free Grammars Chapter 13
Natural Language Processing (NLP)
Part I: Basics and Constituency
Note to teachers: anything in italics should be read aloud
Mean Shift Segmentation
Top-Down Parsing.
Machine Vision Acquisition of image data, followed by the processing and interpretation of these data by computer for some useful application like inspection,
Kocaeli University Introduction to Engineering Applications
BBI 3212 ENGLISH SYNTAX AND MORPHOLOGY
Natural Language - General
R.Rajkumar Asst.Professor CSE
Natural Language Processing
CS246: Information Retrieval
Teori Bahasa dan Automata Lecture 9: Contex-Free Grammars
Applied Linguistics Chapter Four: Corpus Linguistics
3D Shape Inference Computer Vision No.2-1.
IT523 Digital Image Processing
Natural Language Processing (NLP)
An Information Evening for Parents
David Kauchak CS159 – Spring 2019
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
Theoretical approaches to helping children to learn to read:
Natural Language Processing (NLP)
Presentation transcript:

Advanced Higher Computing Based on Heriot-Watt University Scholar Materials Applications of AI – Vision and Languages 1

Lesson Objectives Computer Vision Natural Language Understanding Speech Recognition Stages of NLU Ambiguity Simple Grammar 2

Stages in Vision Image acquisition - Signal processing Edge detection Capturing and understanding an image is a 5 stage process: Image acquisition - Signal processing Edge detection Object recognition Image understanding Each stage is complex. This sub-topic looks at the object recognition stage 3

Challenges in Computer Vision Humans take their vision system fro granted. We can capture and make sense of complex scenes in a fraction of a second. Computer vision programs need resolve ambiguity – where an object, or parts of an object can be classified from a range of possible objects. If we have a circle, should it be interpreted as the top surface of a cylinder, the bottom face of a cone, the base of a hemisphere or the outline of a sphere?

Challenges in Computer Vision II Another challenge lies with optical illusions that produce two different interpretations depending on how you look at the picture. In the picture above, can you see either a white vase or the side profile of two faces?

Digitisation (Higher revision) From an image source such as a digital camera, a scanner or even a satellite image, the source is represented as an array of pixels (bitmap). Problems at this stage – quality of the image (poor lighting, dust on lens or electronic distortion. Depending on quality, the edge of an object (see line above) may appear pixelated ie as a saw-tooth line.

Signal Processing (Higher revision) Calculations are performed to improve the quality of the image. Signals of interest can include: 1. Time-varying measurement values. 2. Sensor data, for example biological data such as electrocardiograms,control system signals, telecommunication transmission signals such as radio signals, and many others.

Edge and region detection (Higher revision) Stages the computer defines and locates objects and areas. The values of adjacent pixels are compared and so edges are detected. In both these stages there are a number of mathematical methods employed to analyse and interpret the pixels. BUT – artificial intelligence only begins to be applied at the next two stages: Object recognition Image understanding

Concave or convex? How do we now get the computer to decide on concave or convex edges? 1. The boundary lines of the object obscure the background and so all of these lines can be marked by a set of clockwise arrows 2. The lines that form the faces meet at vertices where three faces have a common point - hence the name trihedral vertex. There are 18 possible trihedral vertices that can be categorised as T junctions, arrows, forks (Y junctions) and L junctions.

Categorising the trihedral vertices Inspect the currently labelled lines at each vertex and decide which of the above 18 possible patterns our vertex fits. It may not be possible to categorise a vertex at first, but another vertex will supply information to categorise on a second or later repetition Try the examples on Page 119 of Scholar notes.

Natural Language Understanding (NLU) Humans use their ears to input sound waves and the brain processes this data to try to make sense of the sound. You decide a response and then use the throat, lips and tongue to create sound waves to convey your response. A computerised natural language understanding system must be able to mimic the human natural language understanding system as closely as is possible. Speech Recognition Natural Language Understanding Natural Language Generation Speech Synthesis Speech in Speech out

Speech recognition (Higher revision) NLP systems can provide input in the form of text (typed or written) and this form of data does not need to be split up into ‘words’ as it already has been. Input in the form of a sound wave WILL need to separated into recognisable words that are formed by the sounds. Stages Analyse the digital sound data (frequency spectrograph) Break continuous data stream into individual sounds) Human language has around 50 of these called phonemes

Activities Do all Review questions in Scholar pp112 – 122 and check solutions

Stages of NLU Having a list of probable words is just the first stage. Need to determine what the list of words mean. Natural Language Understanding has the following stages: Syntactic Analysis – words fit together into allowed structures. A simple approach is Eliza which uses simple pattern matching – need something more sophisticated (grammar checks and Semantic Analysis – extracting meaning (the boy ate the chocolate, etc.) Pragmatic Analysis – deciding meaning based on context (what went before and what comes after)

Dealing with Ambiguity Ambiguity is where a sentence can have more than one meaning and this can happen at any of the three stages Let’s discuss Q7 – Q14 on page 124 of Scholar notes

A simple grammar Restrict the problem to a limited vocabulary with a limited grammar Problem with word matching: 1. The boy ate the chocolate 2. The chocolate ate the boy In sentence 1, the verb = ‘ate’ and subject= ‘boy’ - acceptable In sentence 2, the verb = ‘ate’ and subject= ‘chocolate’ - not acceptable Word match exactly but need to examine the structure to determine if the sentence makes sense

Building Blocks Let’s look at the task: Identifying word types on page 125 of Scholar notes

Combining words together In our simplified grammar, a sentence will consist of a noun phrase followed by a verb phrase eg ‘boy jumps wall’: The following sentences fit our simple grammar: 1. Rome is a city. 2. Rome is a beautiful city. 3. The tall man wrote a long letter. 4. The computer works. Simple Prolog rule: sentence(X,Y) :- noun_phrase(X), verb_phrase(Y).

Limited vocabulary of simple grammar Let’s explore page 127-128 of Scholar Notes Do activity Creating a parse tree on page 129 of Scholar notes