Web audio codec guide - Web media technologies 编辑

Even modest quality, high-fidelity stereo sound can use a substantial amount of disk space. For web developers, an even bigger concern is the network bandwidth needed in order to transfer audio, whether for streaming or to download it for use during gameplay. The processing of audio data to encode and decode it is handled by an audio codec (COder/DECoder). In this article, we look at audio codecs used on the web to compress and decompress audio, what their capabilities and use cases are, and offer guidance when choosing audio codecs to use for your content.

Additionally, WebRTC implementations generally use a subset of these codecs for their encoding and decoding of media, and may support additional codecs as well, for optimal cross-platform support of video and audio conferencing, and to integrate better with legacy telecommunication solutions. See Codecs used by WebRTC for details.

For information about the fundamental concepts behind how digital audio works, see the article Digital audio concepts.

Common codecs

The list below denotes the codecs most commonly used on the web and which containers (file types) support them. If all you need to know is which codecs are even possible to use, this is for you. Of course, individual browsers may or may not choose to support all of these codecs, and their support for which container types can use them may vary as well. In addition, browsers may choose to support additional codecs not included on this list.

Codec name (short)Full codec nameContainer support
AACAdvanced Audio CodingMP4, ADTS, 3GP
ALACApple Lossless Audio CodecMP4, QuickTime (MOV)
AMRAdaptive Multi-Rate3GP
FLACFree Lossless Audio CodecMP4, Ogg, FLAC
G.711Pulse Code Modulation (PCM) of Voice FrequenciesRTP / WebRTC
G.7227 kHz Audio Coding Within 64 kbps (for telephony/VoIP)RTP / WebRTC
MP3MPEG-1 Audio Layer IIIMP4, ADTS, MPEG1, 3GP
OpusOpusWebM, MP4, Ogg
VorbisVorbisWebM, Ogg

[1] When MPEG-1 Audio Layer III codec data is stored in an MPEG file, and there is no video track on the file, the file is typically referred to as an MP3 file, even though it's still an MPEG format file.

Factors affecting the encoded audio

There are two general categories of factors that affect the encoded audio which is output by an audio codec's encoder: details about the source audio's format and contents, and the codec and its configuration during the encoding process.

For each factor that affects the encoded audio, there is a simple rule that is nearly always true: because the fidelity of digital audio is determined by the granularity and precision of the samples taken to convert it into a data stream, the more data used to represent the digital version of the audio, the more closely the sampled sound will match the source material.

The effect of source audio format on encoded audio output

Because encoded audio inherently uses fewer bits to represent each sample, the source audio format may actually have less impact on the encoded audio size than one might expect. However, a number of factors do still affect the encoded audio quality and size. The table below lists a number of key source audio file format factors and their impact on the encoded audio.

The effect of source audio format and contents on the encoded audio quality and size
FeatureEffect on qualityEffect on size
Channel countThe number of channels affects only the perception of directionality, not the quality.Each channel may substantially increase the encoded audio size, depending on contents and encoder settings.
Noise / HissUnwanted background noise or hiss tends to reduce audio quality both directly (by masking details of the foreground audio) and indirectly (by making the audio waveform more complicated and therefore difficult to reduce in size while maintaining precision).Hiss, static, or background noise increases the audio complexity, which generally reduces the amount of compression which is possible.
Sample rateThe more samples available per second, the higher the resulting encoded audio fidelity is likely to be.Increasing the sample rate increases the encoded audio file's size.
Sample sizeThe larger the samples, the more detail each sample can contain, resulting in more accurate representation of each sample.Depends on the codec; codecs typically have an internal sample format that may or may not be the same as the original sample size. But more source detail may make the encoded file larger; it will never make it smaller.

Of course, these effects can be altered by decisions made while encoding the audio. For example, if the encoder is configured to reduce the sample rate, the sample rate's effect on the output file will be reduced in kind.

For more information about these and other features of audio data, see Audio data format and structure in Digital audio concepts.

The effect of codec configuration on encoded audio output

Audio codecs typically employ cleverly-designed and highly-complex mathematical algorithms to take source audio data and compress it to take substantially less space in memory or network bandwidth. In addition to choosing the type of encoder to use, you may have the opportunity to adjust the encoder using parameters that choose specific algorithms, tune those algorithms, and specify how many passes to apply while encoding.

Audio encoder configuration effects on quality and size
FeatureEffect on qualityEffect on size
Lossless compressionNo loss of fidelityUnlikely to get more than 40-50% compression
Lossy compressionAlways some loss of fidelity; the higher the compression, the more the lossCompression of up to 80-95% possible
Quality settingThe higher the quality, the better the fidelity of the encoded audioThe higher the fidelity, the larger the resulting file becomes, though the amount of change varies from codec to codec
Bit rateThe higher the bit rate, the higher the quality can beThe higher the bit rate, the larger the encoded file is likely to be
Audio frequency bandwidthIf there is any audio in the removed frequency band(s), there may be a noticeable loss of fidelityRemoving frequency bands means less data to encode, thus smaller encoded files
Stereo codingSimple stereo and mid-side stereo coding don't affect quality; intensity stereo coding introduces loss of detail, however.Joint stereo can reduce the size of the encoded audio to some extent

The parameters available—and the range of possible values—varies from codec to codec, and even among different encoding utilities for the same codec, so read the documentation that comes with the encoding software you use to learn more.

Features affecting encoded audio size

Several factors affect the size of the encoded audio. Some of these are a matter of the form of the source audio; others are related to decisions made while encoding the audio.

Lossless versus lossy codecs

There are two basic categories of audio compression. Lossless compression algorithms reduce the size of the audio without compromising the quality or fidelity of the sound. Upon decoding audio compressed with a lossless codec such as FLAC or ALAC, the result is identical in every way to the original sound, down to the bit.

Lossy codecs, on the other hand, take advantage of the fact that the human ear is not a perfect interpreter of audio, and of the fact that the human brain can pluck the important information out of imperfect or noisy audio. They strip away audio frequencies that aren't used much, tolerate loss of precision in the decoded output, and use other methods to lose audio content, quality, and fidelity to produce smaller encoded media. Upon decoding, the output is, to varying degrees, still understandable. The specific codec used—and the compression configuration selected—determine how close to the original, uncompressed audio signal the output seems to be when heard by the human ear.

Because of the differences in how lossy codecs work compared to lossless ones, especially the fact that lossless ones have to be much more conservative with their compression, lossy codecs nearly always result in significantly smaller compressed audio than lossless codecs do.

Generally speaking, the most common reasons to choose lossless audio are because you require archival-quality storage, or because the audio samples will be remixed and recompressed, and you wish to avoid the amplification of artifacts in the audio due to recompression. For real-time streaming of audio, a lossy codec is usually required in order to ensure the flow of data can keep up with the audio playback rate regardless of network performance.

Maximum number of channels

The audio delivered to each speaker in a sound system is provided by one audio channel in a stream. Monaural sound is a single channel. Stereo sound is two. 5.1 surround sound has five audio channels, plus one Low Frequency Enhancement (LFE) channel.

LFE channels are specifically designed to store low-frequency audio data, and are commonly used to provide audio data for subwoofers, for example. When you see the number of audio channels written in the form X.Y (such as 2.1 or 5.1), the number after the decimal point, Y, is the number of LFE channels. For example, MP3 supports one LFE channel, while AAC supports up to 16.

In addition to providing audio for specific speakers in a sound system, some codecs may allow audio channels to be used to provide alternative audio, such as vocals in different languages or descriptive audio for visually impaired people.

Audio frequency bandwidth

The audio frequency bandwidth of a codec indicates the range of audio frequencies that can be represented using the codec. Some codecs operate specifically by eliminating audio that falls outside a given frequency range. There is a correlation between the sample rate and the maxium sound frequency that can be represented by a waveform represented by a codec. At a theoretical level, the maximum frequency a codec can represent is the sample rate divided by two; this frequency is called the Nyquist frequency. In reality, the maximum is slightly lower, but it's close.

The audio frequency bandwidth comes into play especially vividly when a codec is designed or configured to represent human speech rather than a broad range of sounds. Human speech generally resides within the audio frequency range of 300 Hz to 18 kHz. However, the vast majority of human vocalizations exist in the range 300 Hz to 8 kHz, and you can capture enough of human vocalizations in the frequency range 500 Hz to 3 kHz to still be understandable.

For that reason, speech-specific codecs often begin by dropping sound that falls outside a set range. That range is the audio frequency bandwidth. G.722, for example, strips away sounds outside the audio frequency bandwidth of 50 Hz to 7 kHz. This reduces the amount of data that needs to be encoded from the outset.

Codec details

Below we take a brief look at each of these codecs, looking at their basic capabilities and their primary use cases.

AAC (Advanced Audio Coding)

The Advanced Audio Coding (AAC) codec is defined as part of the MPEG-4 (H.264) standard; specifically, as part of MPEG-4 Part 3 and MPEG-2 Part 7. Designed to be able to provide more compression with higher audio fidelity than MP3, AAC has become a popular choice, and is the standard format for audio in many types of media, including Blu-Ray discs and HDTV, as well as being the format used for songs purchased from online vendors including iTunes.

AAC has a number of profiles that define methods of compressing audio for specific use cases, including everything from high quality surround sound to low-fidelity audio for speech-only use.

As a patent-encumbered format, AAC support is somewhat less predictable. For example, Firefox only supports AAC if support is provided by the operating system or an external library.

Supported bit ratesArbitrary, up to 512 kbps
Variable Bit Rate (VBR) supportYes
Supported sample formats32-bit integer
Supported sample rates8 kHz - 96 kHz
Recommended minimum bit rate for stereo sound96 kbps at 48 kHz sample rate
CompressionLossy
Maximum audio channels48 (plus 16 Low Frequency Enhancement channels)
Audio frequency  bandwidth0 Hz - 96 kHz (standard audio channels)
0 Hz - 120 Hz (LFE channels)
Latency20 ms to 405 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
AAC supportYes[2]YesYes[1]9Yes3.1
Container supportMP4, ADTS, 3GP
RTP / WebRTC compatibleYes
LicensingFor streaming or distributing AAC-encoded content: no license required; developers of codecs are required to obtain a patent license through VIA Licensing

[1] Due to patent issues, Firefox does not directly support AAC. Instead, Firefox relies upon a platform's native support for AAC. This capability was introduced on each platform in different Firefox releases:

AAC support in Firefox using external library, by platform
PlatformFirst Firefox version
with AAC support
Windows (Vista and later)22
Android20
Linux (depends on GStreamer)26
macOS35

[2] Chrome supports AAC only in MP4 containers, and only supports AAC's Main Profile. In addition, AAC is not available in Chromium builds.

ALAC (Apple Lossless Audio Codec)

The Apple Lossless Audio Codec (ALAC or Apple Lossless) is a lossless codec developed by Apple. After initially being a closed format, Apple opened it up under an Apache license.

Cross-platform and browser support for ALAC is not very strong, making it a less than ideal choice for general usage. However, if your target is primarily macOS and iOS users, it may be worth considering, as the operating systems have integrated support for ALAC. Otherwise, FLAC is likely a better choice, if you must use a lossless codec.

Keep in mind, however, that lossless codecs require substantially more bandwidth and storage capacity, and may not be a good choice outside very specific use cases.

Supported bit ratesBased on the sample format and sample rate, as well as the compression level
Variable Bit Rate (VBR) supportNo
Supported sample formats16-bit, 20-bit, 24-bit, and 32-bit integer
Supported sample rates1 Hz to 384,000 Hz
Recommended minimum bit rate for stereo soundn/a
CompressionLossless; up to 45-60%
Maximum audio channels8 (up to 7.1 surround)
Audio frequency  bandwidth?
Latency?
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
ALAC supportNoNoNoNoNoYes
Container supportMP4
RTP / WebRTC compatibleNo
LicensingOpen license (Apache License 2.0); source code available on GitHub

AMR (Adaptive Multi-Rate)

The Adaptive Multi-Rate audio codec is optimized for encoding human speech efficiently. It was standardized in 1999 as part of the 3GPP audio standard used for both GSM and UMTS cellular telephony, and uses a multi-rate narrowband algorithm to encode audio frequencies at a telephony-grade quality level at around 7.4 kbps. In addition to being used for real-time telephony, AMR audio may be used for voicemail and other short audio recordings.

AMR audio which is stored in files may be typed .amr, but may also be encapsulated in .3gp files.

As a speech-specific codec, AMR is essentially useless for any other content, including audio containing only singing voices. Additionally, because AMR is designed to minimize capacity requirements, it only captures the portion of the full audio frequency bandwidth of human speech which is absolutely necessary to understand what's being said, so quality is reduced accordingly. If you need the ability to record audio at a minimal impact to network and/or storage capacity, AMR can be a great choice. However, if you need high-fidelity reproduction of human speech—or even low-quality music reproduction—you need to choose another format.

Supported bit ratesHalf Rate (HR) and Full Rate (FR): 1.8 kbps, 4.75 kbps, 5.15 kbpz, 5.9 kbps, 6.7 kbps, 7.4 kbps, 7.95 kbps
Full Rate (FR) only: 10.2 kbps and 12.2 kbps
Variable Bit Rate (VBR) supportNo
Supported sample formats13-bit integer
Supported sample rates8 kHz
Recommended minimum bit rate for stereo soundn/a
CompressionLossy
Maximum audio channels1
Audio frequency  bandwidth200 Hz to 3,400 Hz
Latency25 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
AMR supportNo[1]?No[2]?No?
Container supportAMR, 3GPP
RTP / WebRTC compatibleNo
LicensingNon-free; license fees and annual royalties apply. See VoiceAge licensing for details

[1] While the Chrome browser does not support AMR, Chrome OS supports AMR-NB (narrowband) and AMR-WB (wideband).

[2] The Firefox browser doesn't support AMR; however, the Boot to Gecko platform does support both narrow and wideband AMR in 3GP containers.

FLAC (Free Lossless Audio Codec)

FLAC (Free Lossless Audio Codec) is a lossless audio codec published by the Xiph.org Foundation. It provides good compression rates with no loss of audio fidelity; that is, the decompressed audio is identical to the original. Because the compression algorithm is specifically designed for audio, it gives better results than would be achieved using a general-purpose compression algorithm.

FLAC is a great choice for smaller audio effects files where pristine quality and tonal accuracy are desired, as well as for archival of music.

Supported bit rates
Variable Bit Rate (VBR) supportNo
Supported sample formats4-bit to 24-bit integers
Supported sample rates1 Hz to 65,535 Hz (in 1 Hz increments) or 10 Hz to 655,350 Hz in 10 Hz increments
Recommended minimum bit rate for stereo sound
CompressionLossless; up to 40-50% size reduction
Maximum audio channels8
Audio frequency  bandwidthFull-spectrum
Latency4.3 ms to 92 ms with 46.4 ms being the typical average
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
FLAC supportYesYes51 (desktop)
58 (mobile)
NoNo11
Container supportMP4, Ogg, FLAC
RTP / WebRTC compatibleNo
LicensingFully open and free of any licensing requirements

G.711 (Pulse Code Modulation of Voice Frequencies)

The G.711 specification, published by the International Telecommunications Union (ITU), was initially created in 1972 to define standards for encoding the voice frequencies on telephone lines into digital form. The audio is encoded using Pulse Code Modulation (PCM) with either µ-law or A-law encoding.

The primary benefit to µ-law encoding is that the byte 0x00 never appears in the encoded audio, which may be beneficial in certain applications or networking environments. µ-law is primarily used in North America and Japan, with A-law being predominant elsewhere.

The drawbacks to G.711 are obvious: small 8-bit samples and narrow audio frequency bandwidth. That having been said, it's been in use for decades, is "good enough" for voice applications, and is vital if you may need to interface with a standard telephone network. This codec is required to be supported by all WebRTC solutions because it's a simple, easy to implement, and widely-used and compatible codec.

Typically, most modern applications will use G.711 only as a fallback option, unless the limitations have specific value to their use cases. Opus, for example, offers a narrowband encoding at a 40 kbps bit rate. Still, G.711 provides a least common denominator which can be fallen back upon when needed.

Supported bit rates64 kbps
Variable Bit Rate (VBR) supportNo
Supported sample formats14-bit signed integer (µ-law) or 13-bit unsigned integer (A-law); encoded audio is 8 bits per sample
Supported sample rates8 kHz (± 50 parts per million
Recommended minimum bit rate for stereo soundn/a (stereo not supported)
CompressionNone
Maximum audio channels1
Audio frequency bandwidth300 Hz – 3400 Hz
Latency0.125 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
G.711 support23[1]15[1]22[1]No43[1]11[1]
Container support3GP
RTP / WebRTC compatibleYes
LicensingAll applicable patents have expired, so G.711 is free to use without restriction

[1] G.711 is supported only for WebRTC connections.

G.722 (64 kbps (7 kHz) audio coding)

Published by the International Telecommunications Union (ITU), the G.722 codec is designed specifically for voice compression. Its audio coding bandwidth is limited to the 50 Hz to 7,000 Hz range, which covers most of the frequency range of typical human vocalization. This makes it ill-suited for handling any audio that might range outside the human speech range, such as music.

G.722 audio is encoded using Adaptive Differential Pulse Code Modulation (ADPCM), in which each sample is represented not by its absolute value, but as a value indicating how much the new sample differs from the previous sample.

G.722 is primarily used with WebRTC connections, as it's one of the audio codecs mandated by the WebRTC specification.

Supported bit ratesG.722: 48 kbps, 56 kbps, and 64 kbps; however, in practice 64 kbps is always used
G.722 Annex B Super Wide-Band: 64 kbps, 80 kbps, and 96 kbps
G.722 Annex D Stereo Wide-Band: 64 kbps and 80 kbps
G.722 Annex D Stereo Super Wide-Band: 80 kbps, 96 kbps, 112 kbps, and 128 kbps
Variable Bit Rate (VBR) supportNo
Supported sample formats14-bit integer
Supported sample rates16 kHz (ADPCM is specified to allow 8 kHz, 11.025 kHz, 22.05 kHz, 44.1 kHz, but G.722 uses 16 kHz)
Recommended minimum bit rate for stereo sound128 kbps at 44.1 kHz sample rate
CompressionLossy
Maximum audio channels2
Audio frequency bandwidth50 Hz - 7 kHz
Latency4 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
G.722 supportYes[1]Yes[1]Yes[1]NoYes[1]Yes[1]
Container support3GP, AMR-WB
RTP / WebRTC compatibleYes
LicensingAll applicable patents have expired; G.722 is free to use without restriction

[1] WebRTC only.

MP3 (MPEG-1 Audio Layer III)

Of the audio formats specified by the MPEG/MPEG-2 standards, MPEG-1 Audio Layer III—otherwise known as MP3—is by far the most widely used and best-known. The MP3 codec is defined by MPEG-1 Part 3 and MPEG-2 Part 3, and was introduced in 1991 (and finalized in 1992).

When MP3-format audio is stored inside an MPEG container, the resulting file is also referred to as just an "MP3 file" or "MP3." Files with the ubiquitous .mp3 extension are stored in what is perhaps the most widely distributed audio file format in the world, which is in large part responsible for the digital audio revolution of the late 1990s and early 2000s.

MPEG-1 MP3 audio supports higher bit rates as well as higher sampling rates than MP3 audio in MPEG-2 files. MPEG-1 format MP3 is generally best for music or other complex audio, while MPEG-2 mode MP3 audio is acceptable for speech and other simpler sounds.

The patents behind MP3 have expired, removing many or most licensing concerns around using MP3 files in your projects. That makes them a good choice for many projects.

Supported bit ratesMPEG-1 mode: 32 kbps, 40 kbps, 48 kbps, 56 kbps, 64 kbps, 80 kbps, 96 kbps, 112 kbps, 128 kbps, 160 kbps, 192 kbps, 224 kbps, 256 kbps, 320 kbps
MPEG-2 mode: 8 kbps, 16 kbps, 24 kbps, 32 kbps, 40 kbps, 48 kbps, 56 kbps, 64 kbps, 80 kbps, 96 kbps, 112 kbps, 128 kbps, 144 kbps, 160 kbps
Variable Bit Rate (VBR) supportYes
Supported sample formats16-bit integer
Supported sample ratesMPEG-1 mode: 32000 Hz, 44100 Hz, 48000 Hz
MPEG-2 mode: 16000 Hz, 22050 Hz, 24000 Hz (Half the frequency of the MPEG-1 supported modes)
Recommended minimum bit rate for stereo sound128 kbps at 48 kHz sample rate
CompressionLossy
Maximum audio channelsMPEG-1 mode: 2 [2.0]
MPEG-2 mode: 5 (plus 1 optional Low Frequency Enhancement channel) [5.1]
Audio frequency bandwidthVaries, depending on bit rate and psychoacoustic analysis
LatencyAt least 100 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
MP3 supportYesYesYes[1]9Yes3.1
Container supportMPEG-1, MPEG-2, MP4, ADTS, 3GP
RTP / WebRTC compatibleNo
LicensingPatent-free in the EU as of 2012; patent-free in the United States as of April 16, 2017; now free to use

[1] For patent reasons, Firefox did not directly support MP3 prior to version 71; instead, platform-native libraries were used to support MP3. This capability was introduced on each platform in different Firefox releases:

MP3 support with external library, by platform, in Firefox
 PlatformFirst Firefox version
with MP3 support
Windows (Vista and later)22
Android20
Linux (depends on GStreamer)26
macOS35

Opus

The Opus audio format was created by the Xiph.org Foundation as a fully open audio format; it has been standardized by IETF as RFC 6716. It's a good general-purpose audio codec that can efficiently handle both low-complexity audio such as speech as well as music and other high-complexity sounds.

Opus supports multiple compression algorithms, and can even use more than one algorithm in the same audio file, since the encoder can choose the bit rate, audio bandwidth, algorithm, and other details of the compression settings for each frame of audio.

Opus is a good all-around audio codec for use in your web applications, and can be used for any audio tasks you have in mind.

Supported bit rates6 kbps - 510 kbps
Variable Bit Rate (VBR) supportYes
Supported sample formats16-bit integer and 32-bit float (-1.0 to 1.0)
Supported sample rates
ProfileEffective sample rate
Narrowband (NB)8 kHz
Medium band (MB)12 kHz
Wideband (WB)16 kHz
Super Wideband (SWB)24 kHz
Fullband (FB)48 kHz[1]
Recommended minimum bit rate for stereo sound96 kbps at 48 kHz sample rate
CompressionLossy
Maximum audio channels255 (up to 1 LFE channel)
Audio frequency bandwidth
ProfileAudio bandwidth
Narrowband (NB)4 kHz
Medium band (MB)6 kHz
Wideband (WB)8 kHz
Super Wideband (SWB)12 kHz
Fullband (FB)20 kHz[3]
Latency5 ms to 66.5 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
Opus support331415No2011[2]

This information refers to support for Opus in HTML <audio> and <video> elements, and not to WebRTC.

Container supportOgg, WebM, MPEG-TS, MP4
RTP / WebRTC compatibleYes
LicensingFully open and free of any licensing requirements

[1] The specified sample rates are effective sample rates. Opus uses an algorithm based on audio bandwidths rather than sample rates. See RFC 6716, section 2 for details. In addition, there is an optional part of the Opus specification (Opus Custom) that does allow for non-standard sample rates, but the use of this feature is discouraged.

[2] Safari supports Opus in the <audio> element only when packaged in a CAF file, and only on macOS High Sierra (10.13) or iOS 11.

[3] Although the Nyquist–Shannon sampling theorem shows that audio bandwidth can be as much as one half of the sampling rate, Opus doesn't allow encoding outside a maximum 20 kHz audio frequency band, since the human ear can't percieve anything above the 20 kHz point anyway. This saves some space in the encoded audio.

Vorbis

Vorbis is an open format from the Xiph.org Foundation which supports a wide range of channel combinations, including monaural, stereo, polyphonic, quadraphonic, 5.1 surround, ambisonic, or up to 255 discrete audio channels. Depending on the quality setting used during encoding, the resulting bit rate can vary from around 45 kbps to 500 kbps. Vorbis inherently uses variable bit rate encoding; the bit rate can vary from one sample to the next as needed during the compression process.

Generally, Vorbis is more efficient in terms of size and bit rate than MP3 at similar quality levels. This and its free and open license, make it a good choice for many kinds of audio data as long as its high latency isn't a problem.

Supported bit rates45 kbps - 500 kbps
Variable Bit Rate (VBR) supportYes
Supported sample formats16-bit integer
Supported sample rates8 kHz - 192 kHz
Recommended minimum bit rate for stereo sound192 kbps at 48 kHz; this is typically accomplished by setting the quality level to 6 to 8.
CompressionLossy
Maximum audio channels255
Audio frequency bandwidth
LatencyAt least 100 ms
Browser compatibility
FeatureChromeEdgeFirefoxInternet ExplorerOperaSafari
Vorbis support4173.5No11.5No

This information refers to support for Vorbis in HTML <audio> and <video> elements, and not to WebRTC.

Container supportOgg, WebM
RTP / WebRTC compatibleYes
LicensingFully open and free of any licensing requirements

Choosing an audio codec

Typically, regardless of what codec you use, it will generally get the job done, even if it's not the ideal choice, as long as you choose a codec not specifically designed for a totally different kind of source audio. Choosing a voice-only codec and trying to use it for music will not give you usable results, for instance.

Some codecs may limit compatibility, however, and others may be more optimal for your needs than others. Here we'll provide guidance to help you select an appropriate codec for your use case.

When choosing a codec to use for your audio, you should first consider the following questions:

  • Will the encoded audio be getting remixed or recompressed? If so, avoid lossy compression, which would be compounded by recompressing the audio; or at least use as little compression as possible.
  • If the audio needs to go into a specific file type, keep that in mind, since media containers typically support a specific subset of the available codecs.
  • What kind of audio content will the codec be handling? Certain codecs are specifically designed for voice-only content (they take advantage of the reduced frequency range needed for human speech). Others may have an algorithmic tendency to perform worse when encoding specific genres of music.
  • What bit rates and other configurable properties does each codec have that may make it a good (or poor) choice?
  • To what extent, if at all, does latency matter for your needs? If you need sound that is very precisely timed, the lower the latency the better.
  • How much compression do you need to achieve?

Let's consider a few common scenarios to get a feel for the decision-making process.

Example: Music for streaming

For streaming music, you want to select a codec that minimizes bandwidth usage as much as possible while introducing as few artifacts as possible into the audio through compression. This is necessary because the rate at which the music is downloaded needs to be no greater than the amount of bandwidth available on the network, and ideally there should be room left for network speed fluctuations and use of the network by other applications.

Unless there's a specific need for lossless compression, or the network bandwidth is guaranteed to be high enough to support it, a lossy compression scheme is a good choice. Which one you choose depends on browser compatibility and the availability of any special features you may need the codec to support.

Usually the latency is not particularly important when streaming music. Possible exceptions include looped music, where you need music to be able to play back uninterrupted over and over, or when you need to be able to play songs back to back with no gap between them. This can be particularly important for classical music, theatrical soundtracks, and for background music during gameplay.

For general music playback, the three most likely candidates are MP3, AAC, and Vorbis.

  • AAC in an MP4 container is supported by all major browsers, which makes this a great choice.
  • Vorbis is almost always used in Ogg files, but Ogg containers are not universally supported. Even Microsoft Edge, which supports both Vorbis, does not yet support Ogg containers.
  • MP3 (MPEG-1 Audio Layer III) is supported by all major browsers. These files are MPEG-1 files that contain an Audio Layer III track.

If you need to minimize latency during music playback, you should strongly consider Opus, which has the lowest range of latencies of the general-purpose codecs (5 ms to 66.5 ms, compared to at least 100 ms for the others).

Note: Compatibility information described here is generally correct as of the time this article was written; however, there may be caveats and exceptions. Be sure to refer to the compatibility tables before committing to a given media format.

Based on this, AAC is likely your best choice if you can only support one audio format. Of course, if you can provide multiple formats (for example, by using the <source> element within your <audio> and <video> elements), you can avoid many or all of those exceptions.

Example: Music for download

Music being downloaded by the user can be compressed to a larger overall file size than streamed music, since (unlike when streaming) it doesn't matter if the download speed is slower than the playback speed of the media. That means you can consider using lossy compression at higher bit rates, which results in larger files but with less loss of fidelity. Or you can choose a lossless format. The choice depends largely on the union of your application's requirements and the preferences of your users.

For an actual music download service, you might offer songs for download as 128 Kbps MP3 files, 256 kbps AAC files (in MP4 containers), or FLAC files, depending on a preference selected by the user. If you need to choose just one format, select one that makes sense given your requirements and the type of audio content being downloaded.

Generally, of course, MP3 is the most common format used for music; choose a bit rate of at least 192 kbps if possible. The iTunes store, on the other hand, distributes music in 256 kbps AAC format.

Example: Voice recording and playback

The specific characteristics of human speech allow speech-specific codecs to compress the audio far more than most general-purpose codecs can. That's because although humans hear frequencies ranging from around 20 Hz to 20,000 Hz, and human speech sounds range from about 300 Hz to 18,000 Hz, the majority of the speech sounds that we need in order to understand what's being said lie in the frequency range of 500 Hz to 3,000 Hz or so. That means voice-only codecs can throw away everything else.

The voice-specific codecs are all inherently very lossy, however, and any sound with significant information in the frequency bands outside the vocal range captured will be entirely lost. This makes these codecs completely unsuitable for anything but spoken words. Even audio that contains only voices, but singing rather than speaking, will likely not be of acceptable quality in one of these formats.

Voice recording and playback usually needs to be low-latency in order to synchronize with video tracks, or in order to avoid cross-talk or other problems. Fortunately, the characteristics that lead to speech codecs being so efficient storage space-wise also make them tend to be very low latency. If you're working with WebRTC, G.722, for example, has 4 ms latency (compared with over 100 ms for MP3), and AMR's latency is around 25 ms.

Note: For more about WebRTC and the codecs it can use, see Codecs used by WebRTC.

The codecs generally used on the web that are used for voice-only encoding are G.722 and AMR. AMR is a narrowband codec, encoding only the frequencies between 200 Hz and 3,400 Hz at bit rates typically around 7.4 kbps, while G.722 is a wideband codec that expands the audio bandwidth to 50 Hz to 7,000 Hz at much higher bit rates—usually 64 kbps.

If you have plenty of network bandwidth to work with, and are reasonably certain your users will as well, G.722 is the better choice. To maximize the storage and network efficiency in a constrained environment, choose AMR.

Example: Audio clips for professional mixing

When compressing audio that will be getting mixed or remixed, you typically want zero or close to zero loss of fidelity, which suggests a lossless codec may be in order. However, because lossless encoding naturally has a much lower compression level than lossy encoding, you may find that if your source audio is large enough, you might have to choose a lossy encoder anyway, especially in a web environment in which you can't control the download rate of the media.

Assuming that lossless compression is our best option here (which it usually is, as long as the audio files are small), the three strongest candidates from a codec perspective are FLAC, Apple Lossless (ALAC), and MPEG-4 ALS. Which we choose will depend on browser support and which media container formats support them.

For the purposes of this example, we'll assume that all browsers have the same codec and container support as Firefox (although this is far from true). Consider the breadth of actual support for the codecs when making your decisions.

  • Firefox supports FLAC in FLAC's native containers as well as in Ogg and MPEG-4 (MP4) files.
  • Firefox only supports Apple Lossless through its platform-specific QuickTime support.
  • Firefox doesn't support MP4 ALS.

In this case, it appears that FLAC is most likely the best codec to go with; ALAC has little to no direct browser support.

Audio encoding software

There are many tools available for encoding audio. The simplest are those which are meant for ripping CDs or pulling in audio files and quickly and automatically converting them to MP3 or AAC format to store in a library, such as iTunes. But when developing web apps that use audio as a component of the app, such as games, you will need more control over the encoding process and more options surrounding the format used while encoding the audio.

A few popular options:

FFmpeg
Arguably the best-known and most widely regarded open source codec packages available, FFmpeg supports the majority of the most popular audio formats and provides command-line tools and libraries for encoding, decoding, and performing format conversions of both audio and video. Binaries are available for macOS, Linux, and Windows.
Handbrake
A highly popular open source front-end to FFmpeg which adds a graphical user interface that makes it much easier to control the wide variety of options FFmpeg offers while encoding audio and/or video. Binaries are available for macOS, Linux, and Windows.
Audacity
An open source audio editor that supports loading audio from many different formats, editing, filtering, and adjusting the audio, and saving it back out in either the original format or a new format.  Available for macOS, Linux, and Windows.
LAME
A high quality open source MP3 encoder with support for CBR, ABR, and VBR encoding as well as a variety of other options. Distributed only in source form by the LAME project, but can be installed using Homebrew or similar tools.

See also

如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。

扫码二维码加入Web技术交流群

发布评论

需要 登录 才能够评论, 你可以免费 注册 一个本站的账号。
列表为空,暂无数据

词条统计

浏览:62 次

字数:65634

最后编辑:7年前

编辑次数:0 次

    我们使用 Cookies 和其他技术来定制您的体验包括您的登录状态等。通过阅读我们的 隐私政策 了解更多相关信息。 单击 接受 或继续使用网站,即表示您同意使用 Cookies 和您的相关数据。
    原文