Audio mixing is the process by which multiple sounds are combined into one or more audio channels. In the process, a source's volume level, frequency content, dynamics, and panoramic position are manipulated or enhanced. This practical, aesthetic, or otherwise creative treatment is done in order to produce a finished version that is appealing to listeners.
Before the introduction of multitrack recording, all the sounds and effects that were to be part of a recording were mixed together at one time during a live performance. If the sound blend was not satisfactory, or if one musician made a mistake, the selection had to be performed over until the desired balance and performance was obtained. However, with the introduction of multitrack recording, the production phase of a modern recording has radically changed into one that generally involves three stages: recording, overdubbing, and mixdown.[1]
Film and television
During production dialogue recording of actors is done by a person variously known as location sound mixer, production sound or some similar designation. That person is a department head with a crew consisting of a boom operator and sometimes a cable person.
Audio mixing for film and television is a process during the post-production stage of a moving image program by which a multitude of recorded sounds are combined. In the editing process, the source's signal level, frequency content, dynamics, and panoramic position are commonly manipulated and effects added. In video production, this is called sweetening.
The process takes place on a mixing stage, typically in a studio or purpose-built theater, once the picture elements are edited into a final version. Normally the engineers will mix four main audio elements called stems: speech (dialogue, ADR, voice-overs, etc.), ambience (or atmosphere), sound effects, and music. As multi machine synchronization became available, filmmakers were able to split elements into multiple reels. With the advent of digital workstations and growing complexity, track counts in excess of 100 became common.
Dialogue intelligibility
Since the 2010s, critics and members of the audience have reported that dialogue in films tends to be increasingly more difficult to understand than in older films, to the point where viewers need to rely on subtitles to understand what is being said. Ben Pearson of SlashFilm attributed this to a combination of factors, only some of which can be addressed through audio mixing:[2]
Unintelligibility as a stylistic choice by filmmakers, particularly by Christopher Nolan and those influenced by him
Soft, under one's breath delivery of lines by actors, a practice particularly popular among younger actors, as opposed to the theatrical clarity of delivery previously used
Low priority of sound recording on set, with priority given to the visual aspects of production
Increased technological possibilities, including in post-production, no longer compel filmmakers to obtain an optimal recording on set
The film crew's familiarity with the dialogue can lead them to overestimate its intelligibility
Theaters play films at a lower than recommended volume to avoid excessive loudness complaints from the audience
Different standards of compression and volume balance applied by the various streaming platforms
Inadequate audio remixing for films played in a home theater setting or on mobile devices, where the audio playback capabilities of the various setups strongly differ from each other and from cinema settings
Live sound mixing is the process of electrically blending together multiple sound sources at a live event using a mixing console. Sounds used include those from instruments, voices, and pre-recorded material. Individual sources may be equalised and routed to effect processors to ultimately be amplified and reproduced via loudspeakers.[3] The live sound engineer balances the various audio sources in a way that best suits the needs of the event.[4]