Preview only show first 10 pages with watermark. For full document please download

Uncompressed Audio Formats

   EMBED


Share

Transcript

 Time varying media  Form of energy that propagates through a certain medium in the form of mechanical waves.  When an object oscillates through a medium, the vibrations are converted to sound.  Acoustics is a branch of science that deals with generation, propagation, transmission of vibrations, sound, ultra sound and infra sound.  Ex. Aeroacoustics, Audio signal processing, psychoacoustics etc.  Sound is one kind of longitudinal wave, in which the particles oscillate to and fro in the same direction of wave propagation. Sound waves cannot be transmitted through vacuum. The transmission of sound requires at least a medium, which can be solid, liquid, or gas.   condensation rarefaction  The wavelength, l is the distance between two successive rarefactions or condensations.  Positive peak in sine wave denotes compression, and the negative one denotes rarefaction.  Amplitude  Frequency  Wavelength  Speed  Waveform  Sound pressure  The amplitude of a periodic variable is a measure of its change over a single period.  Peak-to-peak amplitude is the change between peak (highest amplitude value) and trough (lowest amplitude value, which can be negative).  Peak amplitude  This is the maximum absolute value of the signal. 1: peak 2: peak to peak 3: RMS  For cyclical processes, such as rotation, oscillations, or waves, frequency is defined as a number of cycles per unit time.  Unit: Hertz(Hz).  Range of audible frequency for humans: 20Hz to 20 kHz: sonic frequency.  Ultrasonic > sonic >infraSonic  The period, usually denoted by T, is the duration of one cycle, and is the reciprocal of the frequency f: 1/T.  The distance over which the wave's shape repeats.  Distance between two successive troughs/ crests.  Higher frequency : shorter wavelength.  Indicated by Lambda  The speed of sound is the distance travelled per unit of time by a sound wave propagating through an elastic medium.  In dry air, at 20 °C (68 °F), the speed of sound is 343.2 metres per second (1,126 ft/s).  This is 1,236 kilometres per hour (667 kn; 768 mph), or a kilometre in 2.914 seconds or a mile in 4.689 seconds.  Sound pressure or acoustic pressure is the local pressure deviation from the ambient (average, or equilibrium) atmospheric pressure, caused by a sound wave.  In air, sound pressure can be measured using a microphone, and in water with a hydrophone.  Unit: Pascal (Pa).  Deals with human listening capabilities, psychological responses, and physiological responses, and physiological impact of music and sound upon the human nervous system.   Threshold of hearing Masking  Range of audible frequency for humans: 20Hz to 20 kHz: sonic frequency. As age of the person increases, upper limit decreases.  Adults (on an avg.) cannot hear sound of frequency above 16kHz.  Human hearing is most sensitive for frequency range 1kHz to 5kHz.  Lossy audio data compression methods, such as MPEG/Audio encoding, remove some sounds which are masked anyway  1. The general situation in regard to masking is as follows: A lower tone can eectively mask (make us unable to hear) a higher tone 2. The reverse is not true { a higher tone does not mask a lower tone well 3. The greater the power in the masking tone, the wider is its influence { the broader the range of frequencies it can mask. 4. As a consequence, if two tones are widely separated in frequency then little masking occurs MIDI is a technical standard that describes a protocol, digital interface and connectors and allows a wide variety of electronic musical instruments, computers and other related devices to connect and communicate with one another.  Each MIDI-capable instrument is equipped with a transmitter and receiver, though certain peripheral devices, such as signal processors, may have only a receiver or a transmitter.  The interface operates at a 31.25 Kbaud asynchronous (serial) transmission rate. Each byte consists of a start bit, eight data bits, and a stop bit, for a total duration of 320 microseconds per byte.  It is a standard adopted by the electronic music industry for controlling devices, such as synthesizers and sound cards, that emit music.  At minimum, a MIDI representation of a sound includes values for the note's pitch, length, and volume.  When you play a MIDI instrument, it produces data that can be captured by a MIDI “sequencer.”  When the sequencer plays back the data, your original performance is recreated.  The General MIDI sound set allows you to instantly play back pre-recorded music on many MIDI instruments, or on a computer.  One MIDI device can control another, letting you use the most convenient tool for any job 5 pin DIN connectors: IN, OUT, THRU.  IN: I/p signals, OUT: o/p signals,  THRU: daisy chaining multiple instruments.  THRU 2nd of 1st instrument is connected to IN of and so on.  Suppose you’ve got a MIDI keyboard connected to a MIDI synthesizer box, or “module,” and you want to play the module from the keyboard  When you hold down the Middle C on the keyboard, the keyboard sends out a Middle C “key down” MIDI message to the module. The module says to itself, “Ah, right, Middle C,” and plays Middle C  When you let go of Middle C, the keyboard sends a Middle C “note off” message to the  Note-on and note-off messages are just the beginning, though. Most MIDI controllers can also sense how hard you play a note, transmitting a MIDI “velocity” value to the receiving device so the volume or brightness of its note responds to your touch.  Instructional based music Channel Messages System Messages specific Channel Voice System common change, Channel mode System real time description language.  Channel message applied to : channel.  Voice: control program change, note on, note off, change etc. pitch bend System exclusive  Commands that are not channel specific, such as timing signals for synchronization, positioning information in pre-recorded MIDI sequences, and detailed setup information for the destination device.  Common: meant for receivers of the system. Ex.: song select message is used to store songs and recall them.  Real time: synchronize all the clock equipments of MIDI system.  Exclusive: send data between MIDI devices. based  Digital audio is technology that can be used to record, store, generate, manipulate, and reproduce sound using audio signals encoded in digital form. A microphone converts sound to an analog electrical signal, then ADC—typically using pulse-code modulation—converts the analog signal into a digital signal. A digital-to-analog converter performs the reverse process, converting a digital signal back into an analog signal, which analog circuits amplify and send to a loudspeaker.  Digital audio compression, systems storage, may include processing and transmission components.  Conversion convenient to a digital format manipulation, allows storage, transmission and retrieval of an audio signal.  Sampling is the method of converting an analogue signal to digital data.  Audio sampling is based on the Nyquist-Shannon sampling theorem, which states that for accurate signal reconstruction the sampling frequency should be at least twice the bandwidth of the source signal. Therefore, if a source signal has a bandwidth of 0 1KHz, then the sampling rate needs to be at least 2KHz.  The vast majority of digital audio systems encode the numbers which represent the original audio waveform as binary data, by means of a process known as pulse code modulation ( PCM ).  The logical format of an audio CD (officially Compact Disc Digital Audio or CD-DA) is described in a document produced in 1980 by the format's joint creators, Sony and Philips.  The selection of the sample rate was primarily based on the need to reproduce the audible frequency range of 20Hz - 20kHz.  The Nyquist–Shannon sampling theorem states that a sampling rate of double the maximum frequency to be recorded is needed, resulting in a 40 kHz rate. CD Formats CD-digital audio CD-ROM CDInteractive CDRecordable CDRewritable Data stored in form of sectors /blocks.  1 sec sound= 75 blocks & 1 block= 2352 bytes.  Contents can be ripped and copied on PC.  Super Audio CD gives better playback than CD-DA.  Follows red book standards.  MODE 1: 2352 bytes=2048(original data)+304(error detection & correction).  MODE 2: error detection &correction codes are absent. Can play 75 blocks in 1 sec.  Follows yellow book standards.  Includes text, image, audio, graphics, video, comp. programs.  H/w to store and process the data by connecting to TV(direct plug in + remote).  H/w used: decoder, CD player, joystick, controller and mouse interface.   2nd part of orange book.  Write once and read many times.  Multi session recordings.  3rd part of orange book.  Data stored can be erased and rewritten.  Incompatible with drives which do not support multi-session. White book standard.  Capable to hold 75 mins of compressed video with resolution of 352*288 or 352*240.  Audio is stored at 224 Kbits/sec & video at 1150 Kbits/sec.  Xing MPEG player and WMP can play them.  Super VCD: better quality A/V.  Uncompressed audio formats — formats that use no compression. This means all the data is available, at the risk of large file sizes. A WAV audio file is an example of an uncompressed audio file.  Lossless compression applies compression to an uncompressed audio file, but it doesn't lose information or degrade the quality of the digital audio file. The WMA audio file format uses lossless compression.  Lossy compression will result in some loss of data as the compression algorithm eliminates redundant or unnecessary information — basically it losses what it sees as irrelevant information, small file size, it is easier to transmit over the Internet.  MP3 and Real Audio files uses a lossy compression.   The most popular sound file format these days is .MP3 for both Mac and PCs.  Other file formats include .AIFF (for Mac); .AU for Mac and UNIX; .WAV for the PC; and .RA for Real Audio, a proprietary system for delivering streaming audio on the Web. and playing  The goal of the MP3 format is to compress a CD-quality song by a factor of 10 to 14 without losing the CD quality of the sound.  With MP3, a 32 megabyte song on a CD compresses down to 3 megabytes or so.  This lets you download a song in minutes rather than hours, and it lets you store hundreds of songs on your computer's hard disk without taking up that much space.  MP3 offers exceptional quality at 1:12 compression. This format is best played with Winamp, although RealAudio Quicktime, and the new Windows Media Player, and many others also support the format.  The size of the MP3 file depends on which compression routine was used, near- CD or CD quality.  It can be broken up into pieces, and each piece is still playable. We can now easily transmit a song over the internet.  Main disadvantage of the MP3 format is that it needs to be decompressed when played.  Format used for storing sound in files developed jointly by Microsoft and IBM.  Support for WAV files was built into windows 95 making it the de facto standard for sound on PCs.  WAV sound files end with a .Wav extension and can be played by nearly all windows applications that support sound.  PCM is widely used CODEC since it uses uncompressed lossless storage method that keeps audio track intact & gives maximum audio quality.  When a PC plays back a WAV file, it converts numbers in the file into audio signals for the PC's speakers.  Strengths:  WAV files are simple and widely used, especially on PCs.  Later versions of Netscape Navigator (3+) and Microsoft Internet Explorer (2+) support the WAV format.  Weaknesses:  WAV is seen as a proprietary Windows format, although conversion tools are available to play WAV files on other platforms.  The AIFF format is associated with Apple QuickTime Player on Mac and Windows PC.  Even if you run under Windows, you will have no problem playing these files so long as you have Apple QuickTime installed.  It uses uncompressed PCM CODEC.  Windows Media Audio, WMA is a Microsoft file format for encoding digital audio files similar to MP3 though can compress files at a higher rate than MP3.  WMA files, which use the ".wma" file extension, can be of any size compressed to match many different connection speeds, or bandwidths. MPEG is a working group in a subcommittee of ISO/IEC in charge of developing international standards for compression, decompression, processing, and coded representation of moving pictures, audio, and their combination.  Applies a filter bank to the input to break it into its frequency components  In parallel, a psychoacoustic model is applied to the data for bit allocation block  The number of bits allocated are used to quantize the info from the lter bank { providing the compression  MPEG audio offers three compatible layers :  Each succeeding layer able to understand the lower layers  Each succeeding layer oering more complexity in the psychoacoustic model and better compression for a given level of audio quality  Each succeeding layer, with increased compression effectiveness, accompanied by extra delay.  The objective of MPEG layers: a good tradeoff between quality and bit-rate C:\MPEG PDF  AAC LC/LTP  HE-ACCv1  HE-ACCv2  AMR-NB  AMR-WB  FLAC  MP3  MIDI  Vorbis  PCM/WAVE