Multimedia – Sound Dr. Lina A. Nimri Lebanese University Faculty of Economic Sciences and Business Administration 1 st branch.

Slides:



Advertisements
Similar presentations
Audio Compression ADPCM ATRAC (Minidisk) MPEG Audio –3 layers referred to as layers I, II, and III –The third layer is mp3.
Advertisements

Department of Computer Engineering University of California at Santa Cruz MPEG Audio Compression Layer 3 (MP3) Hai Tao.
Guerino Mazzola (Fall 2014 © ): Introduction to Music Technology IIIDigital Audio III.6 (Fr Oct 24) The MP3 algorithm with PAC.
CS335 Principles of Multimedia Systems Audio Hao Jiang Computer Science Department Boston College Oct. 11, 2007.
Digital Signal Processing
MPEG-1 MUMT-614 Jan.23, 2002 Wes Hatch. Purpose of MPEG encoding To decrease data rate How? –two choices: could decrease sample rate, but this would cause.
Sound can make multimedia presentations dynamic and interesting.
Time-Frequency Analysis Analyzing sounds as a sequence of frames
Dale & Lewis Chapter 3 Data Representation Analog and digital information The real world is continuous and finite, data on computers are finite  need.
Part A Multimedia Production Rico Yu. Part A Multimedia Production Ch.1 Text Ch.2 Graphics Ch.3 Sound Ch.4 Animations Ch.5 Video.
Digital Audio Coding – Dr. T. Collins Standard MIDI Files Perceptual Audio Coding MPEG-1 layers 1, 2 & 3 MPEG-4.
I Power Higher Computing Multimedia technology Audio.
Speech Compression. Introduction Use of multimedia in personal computers Requirement of more disk space Also telephone system requires compression Topics.
Foundations of Physics
Digital Representation of Audio Information Kevin D. Donohue Electrical Engineering University of Kentucky.
SIMS-201 Characteristics of Audio Signals Sampling of Audio Signals Introduction to Audio Information.
IT-101 Section 001 Lecture #8 Introduction to Information Technology.
1 Digital Audio Compression. 2 Formats  There are many different formats for storing and communicating digital audio:  CD audio  Wav  Aiff  Au 
Digital Audio.
Chapter 15 Sound 15.1 Properties of Sound 15.2 Sound Waves
Chapter 16: Sound 16-3 Intensity of Sound: Decibels
Howell Istance School of Computing De Montfort University
1 Multimedia Systems 1 Dr Paul Newbury School of Engineering and Information Technology ENGG II - 3A11 Ext: 2615.
PH 105 Dr. Cecilia Vogel Lecture 24. OUTLINE  Electronic music  Theramin  Analog vs digital  components in electronic music.
Chapter 6: The Human Ear and Voice
© 2010 Pearson Education, Inc. Conceptual Physics 11 th Edition Chapter 21: MUSICAL SOUNDS Noise and Music Musical Sounds Pitch Sound Intensity and Loudness.
Digital Audio Multimedia Systems (Module 1 Lesson 1)
1 Audio Compression Multimedia Systems (Module 4 Lesson 4) Summary: r Simple Audio Compression: m Lossy: Prediction based r Psychoacoustic Model r MPEG.
Audio Fundamentals Lesson 2 Sound, Sound Wave and Sound Perception
Digital audio. In digital audio, the purpose of binary numbers is to express the values of samples that represent analog sound. (contrasted to MIDI binary.
Digital Audio What do we mean by “digital”? How do we produce, process, and playback? Why is physics important? What are the limitations and possibilities?
Fall 2004EE 3563 Digital Systems Design Audio Basics  Analog to Digital Conversion  Sampling Rate  Quantization  Aliasing  Digital to Analog Conversion.
Digital Audio Watermarking: Properties, characteristics of audio signals, and measuring the performance of a watermarking system نيما خادمي کلانتري
Introduction to Interactive Media 10: Audio in Interactive Digital Media.
DTC 354 Digital Storytelling Rebecca Goodrich. Wave made up of changes in air pressure by an object vibrating in a medium—water or air.
Computing with Digital Media: A Study of Humans and Technology Mark Guzdial, School of Interactive Computing.
Chapter 12 Preview Objectives The Production of Sound Waves
COMP Representing Sound in a ComputerSound Course book - pages
Ch. 16 Sound and Light. Sound  Sound wave: Caused by vibrations and carry energy through a medium.
CSC361/661 Digital Media Spring 2002
CMPT 365 Multimedia Systems
Computer Some basic concepts. Binary number Why binary? Look at a decimal number: 3511 Look at a binary number: 1011 counting decimal binary
Media Representations - Audio
15.1 Properties of Sound  If you could see atoms, the difference between high and low pressure is not as great.  The image below is exaggerated to show.
Multimedia Data Speech and Audio Dr Sandra I. Woolley Electronic, Electrical and Computer Engineering.
Sound and audio. Table of Content 1.Introduction 2.Properties of sound 3.Characteristics of digital sound 4.Calculate audio data size 5.Benefits of using.
Images, Sound, and Multimedia. No Surprises Images, Sound, Music, and Movies – It’s all numbers – Binary Numbers Today we’ll discuss how multimedia is.
Overview of Multimedia A multimedia presentation might contain: –Text –Animation –Digital Sound Effects –Voices –Video Clips –Photographic Stills –Music.
Chapter 12 Sound.
Digital Multimedia, 2nd edition Nigel Chapman & Jenny Chapman Chapter 9 This presentation © 2004, MacAvon Media Productions Sound.
Introduction to SOUND.
1 Audio Compression. 2 Digital Audio  Human auditory system is much more sensitive to quality degradation then is the human visual system  redundancy.
Audio / Sound INTRODUCTION TO MULTIMEDIA SYSTEMS Lect. No 3: AUDIO TECHNOLOGY.
© Houghton Mifflin Harcourt Publishing Company Preview Objectives The Production of Sound Waves Frequency of Sound Waves The Doppler Effect Chapter 12.
Digital Recording. Digital recording is different from analog in that it doesn’t operate in a continuous way; it breaks a continuously varying waveform.
Sound in Multimedia Psychology of sound what do you use it for? what techniques for its communication exist? Science of sound why does it exist? how it.
MPEG-1Standard By Alejandro Mendoza. Introduction The major goal of video compression is to represent a video source with as few bits as possible while.
Chapter 2 Basic Science: Analog and Digital Audio.
Interactive Multimedia Sound Mikael Fernström. Data sources Microphones and transducers –Sample acoustic reality Synthesis –Simulate reality (and beyond.
Multimedia Sound. What is Sound? Sound, sound wave, acoustics Sound is a continuous wave that travels through a medium Sound wave: energy causes disturbance.
Digital Audio I. Acknowledgement Some part of this lecture note has been taken from multimedia course made by Asst.Prof.Dr. William Bares and from Paul.
Chapter 12 Preview Objectives The Production of Sound Waves
The physics of hearing and other odds and ends. The range of human hearing The range of average human hearing is from about 15Hz to about 15,000Hz. Though.
Multimedia Systems and Applications
Sound Digital Multimedia, 2nd edition Nigel Chapman & Jenny Chapman
Chapter 16: Sound HW problems: Chapter 16: Pb.2, Pb.18, Pb.24, Pb.35, Pb.40, Pb.62: Due Wed., Jan. 17.
C-15 Sound Physics 1.
Assist. Lecturer Safeen H. Rasool Collage of SCIENCE IT Dept.
Govt. Polytechnic Dhangar(Fatehabad)
Recap In previous lessons we have looked at how numbers can be stored as binary. We have also seen how images are stored as binary. This lesson we are.
Presentation transcript:

Multimedia – Sound Dr. Lina A. Nimri Lebanese University Faculty of Economic Sciences and Business Administration 1 st branch

What is sound? Sound is vibrations in the air; that is, a series of rising and falling pressures in the air, ▫deviating from the average, which is represented by atmospheric pressure. To prove this, you could place something loud (like an alarm clock) inside a vacuum chamber, and notice that the initially noisy object no longer makes a sound if it isn't surrounded by air anymore. The simplest way to create a sound is to make an object vibrate. ▫In this manner, a violin makes a sound when the bow makes its strikes vibrate, ▫a piano sounds a note when a key is struck, because a hammer struck a string and made it vibrate.

What is sound? This is how sound waves are produced; they can be represented in a diagram as changes in air pressure (or in the electricity level of the magnet) as a function of time. This can be represented as:

Sonogram A sonogram, on the other hand, depicts sound frequencies as a function of time. It should be noted that a sonogram shows fundamental frequency, on top of which higher frequencies, called harmonics, are superimposed. This is what allows us to distinguish between different sources of sound: low notes have low frequencies, while high notes have higher frequencies.

Sonogram

Sound sampling To play sound on a computer, it must be converted into a digital format, as this is the only kind of information computers can work with. ▫A computer program intersperses small samples of the sound (which amount to differences in pressure) at specific intervals of time. This is called sampling or digitizing sound. The period of time between two samples is called the sampling rate.

Sound sampling To reproduce audio which sounds continuous to the ear, it is required to sample at least once every few 100,000 ths of a second (10 micro sec) ▫it is more practical to go by the number of samples per second, expressed in Hertz (Hz).

Nyquist – Shannon theorem Sampling frequency must be high enough to preserve the form of the signal ▫the sampling rate must be equal to or greater than twice the maximum frequency contained in the signal. ▫Our ears can hear sounds up to about 20,000 Hz. Therefore, for a satisfactory level of sound quality, the sampling rate must be at least on the order of 40,000 Hz. There are several standardized sampling rates in use: ▫32 kHz: for digital FM radio (band-limited to 15 kHz) ▫44.1 kHz: for professional audio and compact discs ▫48 kHz: for professional digital multi-track recording, and consumer recording equipment (like DAT or MiniDisc)

Sound for computers Each sample (corresponding to an interval of time) is associated with a value, ▫The sample determines the air pressure value at that moment.

Sound for computers A computer works with bits, so the number of possible values that the sample could have must be determined. ▫This is done by setting the number of bits on which the sample values are encoded. With 8-bit coding, there are 2 8 (= 256) possible values. With 16-bit coding, there are 2 16 (= 65536) possible values. The second option clearly offers higher sound fidelity, but at the cost of using more computer memory.

Sound for computers Stereo sound requires two channels, with sound recorded individually on each one. One channel is fed into the left speaker, while the other is broadcast from the right speaker. In computer processing, a sound is therefore represented by several parameters: ▫The sampling rate ▫The number of bits in a sample ▫The number of channels (one for mono, two for stereo, and four for Quadraphonic sound)

Memory size to store a sound file It is easy to calculate what size an uncompressed audio sequence will be. ▫By knowing how many bits are used to code the sample, you know its size (as the sample size is the number of bits) To find out the size of a channel, all you need to know is the sample rate, and thus the number of samples per second, and from that the amount of space taken up by one second of music. ▫This comes to: Sampling rate x Number of bits

Memory size to store a sound file To find out how much space a sound extract several seconds long will take up, just multiply the preceding value by the number of seconds: ▫Sampling rate x Number of bits x Number of seconds Finally, to determine the actual file size of the extract, the above figure should be multiplied by the number of channels (it will be twice as large for stereo as for mono). The size, in bits, of a sound extract is equal to: ▫Sampling rate x Number of bits x Number of seconds x Number of channels

Audio files FrequencyQuality 48,000 HzDAT, DVD-audio 44,100 HzCD 22,000 HzRadio 8,000 HzTelephone CD stereo audio of one hour: 44,100 x 16 x 2 x 3600 = 635 Mb DVD quadraphonic audio of one hour: 48,000 x 16 x 4 x 3600 = 1,382 Gb

Psychoacoustics Human ear Audition - The ability to hear comes from acoustic process, mechanical, hydraulic, nervous and mental from the outer ear to the brain Apparatus - The main hearing are: ▫the ear canal (outer ear) ▫ossicular chain (middle ear) ▫the cochlea (inner ear)

Human ear

Cochlea ▫Basilar membrane: mechanical frequency analyzer ▫Corti Organ: vibration sensor membrane

Perception of sounds Frequency Range ▫The human ear, through the basilar membrane, collects a continuous range: [20 Hz - 20, 000 Hz] Sensitive frequencies ▫By the effects of filtering and carried resonance in the ear canal and the ossicles, the most sensitive frequencies are : [1, 000 Hz - 5, 000 Hz]

Perception of sounds Fletcher-Munson Curves: ▫Contours measuring perceived strength (loudness) of sound depending on the frequency  The lowest frequency is the Listening threshold and highest frequency is the threshold of pain  The curves evolve according to age

Perception of sounds Bass sounds ▫The first “La” of the piano: 27 Hz ▫The perception is disproportionally reduced

Perception of sounds Medium sounds ▫The standard “La” of the piano : 440 Hz ▫The hearing is very high in this range

Perception of sounds Treble Sounds ▫The last “La” of the piano: 3480 Hz ▫The hearing is highly variable and decreases with age

Masking Masking of frequencies ▫Two sounds of frequencies very close together are merged in a single frequency  Same time: one sound  Closer moments: beating

Masking Temporal Masking ▫A high amplitude sound will be perceived longer and hides the sounds that follow  Echo phenomenon  Exponentially decreases ( ms)

Sound Compression – general method Division ▫The audio signal is divided into 12 samples per second Conversion ▫Each sample is converted into a spectrum of 32 bands of frequencies Filtering ▫The frequency spectrum is analyzed with the psychoacoustics model  Threshold: any frequency below the threshold of listening is eliminated  Masking: any frequency obscured by another is eliminated Coding ▫Determine the number of bits needed to quantify the amplitudes while minimizing the quantization noise

Sound Compression – MP3 MP3 encoding is a lossy method Level I ▫Uniform Quantization ▫Only the frequency masking is used Level II ▫Uniform Quantization ▫Frequency and temporal masking Level III ▫Adaptive Quantization ▫Frequency and temporal masking ▫Huffman coding

Sound Compression – MP3 MP3 (“MPEG1 Audio layer 3”) is a lossy audio data compression format, developed by the International Standardization Organization (ISO). ▫This format is used to compress normal audio formats (WAV or audio CD) at a rate of 1:12. MP3 format permits the equivalent of files of twelve CD music albums to take up the same space as one CD-ROM. Moreover, mp3 barely alters the sound quality perceptible by human ear.

MP3 - Context MP3 compression involves removing the data corresponding to inaudible frequencies by the average person under normal listening conditions. ▫This compression analyses the spectrometric components of an audio signal, and applies a psychoacoustic model to them so as to preserve only "audible" sound. ▫The average human ear is able to recognize sounds between 0.02 kHz and 20 kHz, with sensitivity being at its peak for frequencies between 2 and 5 kHz MPEG compression involves determining which sounds go unheard and can be deleted; it is therefore “lossy compression”

MP3 – The masking effect Gabriel Bouvigne explains : "When you look at the sun and a bird passes in front of it, you don't see it, because the light from the sun is too bright. Acoustics are like that. When there are loud sounds, you don't hear the quiet sounds. For example: When an organist isn't playing, you can hear whistling in the pipes, and when he is playing, you can't hear it anymore, because it's masked. This is why it isn't necessary to record every sound, and this is the main principle used in the MP3 format to save space.“

Huffman code The Huffman algorithm is an encoding (not compression) algorithm, which takes effect at the end of the compression process, by creating variable-length codes over a large number of bits. ▫The codes have the advantage of a unique prefix, but they may be correctly decoded despite their variable length, and this can be done quickly with the use of tables. ▫This type of encoding saves, on average, a little under 20% of the space taken up. When sounds are “pure” (that is, there is no masking), the Huffman algorithm is very effective, as digital audio contains many redundant sounds.

MIDI - Definition MIDI is a protocol for communication and control between electronic musical instruments ▫one or more of these "instruments" can be a computer An instrument that supports a standard MIDI has a MIDI-IN port and a MIDI-OUT port ▫instruments can be linked in cascade (or daisy-chain) Sometimes the device has a MIDI-THRU, which is a direct copy of the MIDI-IN, without the delay caused by the copy on the MIDI- OUT port