Audio Mixing Question

jak.willis wrote on 11/24/2021, 11:33 AM

Hi,

I have a question to ask concerning audio mixing. When adding in my ambient sound effects, I use one set of headphones. But I use another set of headphones for watching/listening TV. So once I’ve finished editing my videos and start watching them back, the effects don’t sound quite the same as how they sound on my mixing headphones. Is there something that can be done about this to make the mix sound as close as possible to how I intended them to sound across a wide range of different headphones? Or is the only answer to simply use the same headphones for both mixing and listening?

Comments

CubeAce wrote on 11/24/2021, 12:23 PM

@jak.willis

Hi Jak.

The best bet is to use the same headphones but even then there is no guarantee the sound will be exactly the same due to different power deliveries and frequency responses from different amplifiers. Some amplifiers may even have frequency response curves built in. Also before you use an editing package make sure any windows settings are neutral. That the computers internal mixer is not connected to any effects or tone controls of any kind.

You don't want anything in the computers audio chain that could influence what you hear through your headphones on the computer.

Just as with video monitors and video, sound can be coloured by the equipment itself and can change dependent on the equipment used.

Ray.

 

Windows 10 Enterprise. Version 22H2 OS build 19045.5737

Direct X 12.1 latest hardware updates for Western Digital hard drives.

Asus ROG STRIX Z390-F Gaming motherboard Rev 1.xx with Supreme FX inboard audio using the S1220A code. Driver No 6.0.8960.1 Bios version 1401

Intel i9900K Coffee Lake 3.6 to 5.1GHz CPU with Intel UHD 630 Graphics .Driver version Graphics Driver 31.0.101.2135 for 7th-10th Gen Intel® with 64GB of 3200MHz Corsair DDR4 ram.

1000 watt EVGA modular power supply.

1 x 250GB Evo 970 NVMe: drive for C: drive backup 1 x 1TB Sabrent NVMe drive for Operating System / Programs only. 1X WD BLACK 1TB internal SATA 7,200rpm hard drives.1 for internal projects, 1 for Library clips/sounds/music/stills./backup of working projects. 1x500GB SSD current project only drive, 2x WD RED 2TB drives for latest footage storage. Total 31TB of 10 external WD drives for backup.

ASUS NVIDIA GeForce RTX 3060 12GB. nVidia Studio driver version 572.60 - 3584xCUDA cores Direct X 12.1. Memory interface 192bit Memory bandwidth 360.05GB/s 12GB of dedicated GDDR6 video memory, shared system memory 16307MB PCi Express x8 Gen3. Two Samsung 27" LED SA350 monitors with 5000000:1 contrast ratios at 60Hz.

Running MMS 2024 Suite v 23.0.1.182 (UDP3) and VPX 14 - v20.0.3.180 (UDP3)

M Audio Axiom AIR Mini MIDI keyboard Ver 5.10.0.3507

VXP 14, MMS 2024 Suite, Vegas Studio 16, Vegas Pro 18, Vegas Pro 21,Cubase 4. CS6, NX Studio, Mixcraft 9 Recording Studio. Mixcraft Pro 10 Studio. CS6 and DXO Photolab 8, OBS Studio.

Audio System 5 x matched bi-wired 150 watt Tannoy Reveal speakers plus one Tannoy 15" 250 watt sub with 5.1 class A amplifier. Tuned to room with Tannoy audio application.

Ram Acoustic Studio speakers amplified by NAD amplifier.

Rogers LS7 speakers run from Cambridge Audio P50 amplifier

Schrodinger's Backup. "The condition of any backup is unknown until a restore is attempted."

jak.willis wrote on 11/24/2021, 6:09 PM

@jak.willis

Hi Jak.

The best bet is to use the same headphones but even then there is no guarantee the sound will be exactly the same due to different power deliveries and frequency responses from different amplifiers. Some amplifiers may even have frequency response curves built in. Also before you use an editing package make sure any windows settings are neutral. That the computers internal mixer is not connected to any effects or tone controls of any kind.

You don't want anything in the computers audio chain that could influence what you hear through your headphones on the computer.

Just as with video monitors and video, sound can be coloured by the equipment itself and can change dependent on the equipment used.

Ray.

Hello Ray, thank you for your response.

So what do you think happens in the case of professional productions? For instance, in the Soap Operas you can usually hear all of the different ambient sounds in the background such as traffic, birds, trains passing, etc. How can the person doing the mixing be sure that those effects will sound the way they want them to sound on the viewer’s speakers and/or headphones?

browj2 wrote on 11/24/2021, 6:37 PM

@jak.willis

Hi,

Well, audio is a very complex subject. If you take a look at what the pros do, they use several types of speakers and headphones knowing that the listener may be using a cell phone, cell phone with headphones - cheap to expensive, computer audio with built in speakers or computer sound systems from cheap to expensive, TV, TV with sound bar, with sound system, etc. They also use different studio speaker systems in a sound conditioned room, with speakers set up for optimal listening. Then, they try to do the best that they can for most users.

Most of us don't have that luxury or the talent and experience, so we just try to do the best that we can with what we have. Try using both of your sets of headphones plus output to a stereo system of some kind.

As for hearing the background or any particular audio, that is what mixing is all about.

I suggest that you search in this forum for my long thread on using YouLean Loudness Meter2. It doesn't do anything, just gives you output. You use the tools in MEP to adjust audio - gain, volume curve, compression, noise reduction, EQ, etc. I use YouLean to help me balance the audio tracks and the overall sound so that I keep within the limits required for YouTube (about -13 or -14 LUFS), no distortion and to keep an even perceived volume throughout. Etc.

John CB

John C.B.

VideoPro X(16); Movie Studio 2025 Platinum; Music Maker 2025 Premium Edition; Samplitude Pro X8 Suite; see About me for more.

Desktop System - Windows 10 Pro 22H2; MB ROG STRIX B560-A Gaming WiFi; Graphics Card Zotac Gaming NVIDIA GeForce RTX-3060, PS; Power supply EVGA 750W; Intel Core i7-10700K @ 3.80GHz (UHD Graphics 630); RAM 32 GB; OS on Kingston SSD 1TB; secondary WD 2TB; others 1.5TB, 3TB, 500GB, 4TB, 5TB, 6TB, 8TB, 12TB, 14TB; three monitors - HP 25" main, LG 4K 27" second, HP 27" third; Casio WK-225 piano keyboard; M-Audio M-Track USB mixer.

Notebook - Microsoft Surface Pro 4, i5-6300U, 8 GB RAM, 256 SSD, W10 Pro 20H2.

YouTube Channel: @JCBrownVideos

CubeAce wrote on 11/24/2021, 7:09 PM

@jak.willis

Hi Jack.

They can't, and believe me some sounds are not reproduced on some sets but the sound is mixed on Studio monitors not headphones in as neutral a sound environment as possible.

Studio control rooms tend not to have reflective walls, floors or ceilings. They are made as acoustically 'dead' as possible so as to monitor only the recordings.

How it is mixed or even recorded depends on the type of production. If a sound that is needed doesn't come out in the recording at the time it is added in post production. Ambience can be matched using impulse reverb units that will listen to the ambience of the original recording and replicate it. If a sound ruins a take on location and not noticed at the time of the take (such as sirens or aircraft sounds in a historical drama) it will be completely re-recorded in the studio bit by bit. Sometimes the sound track is re-recorded anyway for foreign language sales. Sometimes a sound is re-recorded because the source doesn't make the sound the viewer expects to hear, doesn't actually exist in nature, or the producer wants a different sound.

If I watch something on a TV set with headphones I often do not hear as much as sitting in my listening room over studio monitors playing the same program via the internet, dependent on the source of the files being watched.

If you go via the optical out of a TV into a good DAC you may get closer to that experience.

As far as I'm aware TV companies use SRS 3D (stereo source transferred into an expanded sound stage) and SRS TruSurround (up to 7.1 true speaker outputs.) at present to expand the stereo / surround image for broadcast which may not be reproduced that accurately over headphones as there is no cross talk from one transceiver to the other. In other words the right ear never receives audio from the left speaker and vice versa.

In the days of analog broadcast we used to have several speaker types from very cheap transistor radio speakers all the way upwards to test the mixes for clarity, reproduction of problem frequencies, and most importantly phase testing, but this can be done electronically now and with the disappearance of tape heads the need for phase testing beyond polarity of inputs is no longer required.

In my experience whatever you use to master on will never be exactly reproduced on another system unless the system reaches a required minimum standard such as the LucasFilm THX standard for film theaters. I'm not sure if broadcasters or TV manufacturers have got that far yet.

Ray.

Last changed by CubeAce on 11/24/2021, 7:11 PM, changed a total of 1 times.

 

Windows 10 Enterprise. Version 22H2 OS build 19045.5737

Direct X 12.1 latest hardware updates for Western Digital hard drives.

Asus ROG STRIX Z390-F Gaming motherboard Rev 1.xx with Supreme FX inboard audio using the S1220A code. Driver No 6.0.8960.1 Bios version 1401

Intel i9900K Coffee Lake 3.6 to 5.1GHz CPU with Intel UHD 630 Graphics .Driver version Graphics Driver 31.0.101.2135 for 7th-10th Gen Intel® with 64GB of 3200MHz Corsair DDR4 ram.

1000 watt EVGA modular power supply.

1 x 250GB Evo 970 NVMe: drive for C: drive backup 1 x 1TB Sabrent NVMe drive for Operating System / Programs only. 1X WD BLACK 1TB internal SATA 7,200rpm hard drives.1 for internal projects, 1 for Library clips/sounds/music/stills./backup of working projects. 1x500GB SSD current project only drive, 2x WD RED 2TB drives for latest footage storage. Total 31TB of 10 external WD drives for backup.

ASUS NVIDIA GeForce RTX 3060 12GB. nVidia Studio driver version 572.60 - 3584xCUDA cores Direct X 12.1. Memory interface 192bit Memory bandwidth 360.05GB/s 12GB of dedicated GDDR6 video memory, shared system memory 16307MB PCi Express x8 Gen3. Two Samsung 27" LED SA350 monitors with 5000000:1 contrast ratios at 60Hz.

Running MMS 2024 Suite v 23.0.1.182 (UDP3) and VPX 14 - v20.0.3.180 (UDP3)

M Audio Axiom AIR Mini MIDI keyboard Ver 5.10.0.3507

VXP 14, MMS 2024 Suite, Vegas Studio 16, Vegas Pro 18, Vegas Pro 21,Cubase 4. CS6, NX Studio, Mixcraft 9 Recording Studio. Mixcraft Pro 10 Studio. CS6 and DXO Photolab 8, OBS Studio.

Audio System 5 x matched bi-wired 150 watt Tannoy Reveal speakers plus one Tannoy 15" 250 watt sub with 5.1 class A amplifier. Tuned to room with Tannoy audio application.

Ram Acoustic Studio speakers amplified by NAD amplifier.

Rogers LS7 speakers run from Cambridge Audio P50 amplifier

Schrodinger's Backup. "The condition of any backup is unknown until a restore is attempted."

jak.willis wrote on 11/25/2021, 7:04 AM

@jak.willis

Hi,

Well, audio is a very complex subject. If you take a look at what the pros do, they use several types of speakers and headphones knowing that the listener may be using a cell phone, cell phone with headphones - cheap to expensive, computer audio with built in speakers or computer sound systems from cheap to expensive, TV, TV with sound bar, with sound system, etc. They also use different studio speaker systems in a sound conditioned room, with speakers set up for optimal listening. Then, they try to do the best that they can for most users.

Most of us don't have that luxury or the talent and experience, so we just try to do the best that we can with what we have. Try using both of your sets of headphones plus output to a stereo system of some kind.

As for hearing the background or any particular audio, that is what mixing is all about.

I suggest that you search in this forum for my long thread on using YouLean Loudness Meter2. It doesn't do anything, just gives you output. You use the tools in MEP to adjust audio - gain, volume curve, compression, noise reduction, EQ, etc. I use YouLean to help me balance the audio tracks and the overall sound so that I keep within the limits required for YouTube (about -13 or -14 LUFS), no distortion and to keep an even perceived volume throughout. Etc.

John CB

Hello,

Yes, I see what you're saying. So really the best thing to do would be to test how the audio mix sounds on a range of different audio equipment? Obviously I don't have access to lots of different equipment but I'm talking about an ideal world.

jak.willis wrote on 11/25/2021, 7:06 AM

@jak.willis

Hi Jack.

They can't, and believe me some sounds are not reproduced on some sets but the sound is mixed on Studio monitors not headphones in as neutral a sound environment as possible.

Studio control rooms tend not to have reflective walls, floors or ceilings. They are made as acoustically 'dead' as possible so as to monitor only the recordings.

How it is mixed or even recorded depends on the type of production. If a sound that is needed doesn't come out in the recording at the time it is added in post production. Ambience can be matched using impulse reverb units that will listen to the ambience of the original recording and replicate it. If a sound ruins a take on location and not noticed at the time of the take (such as sirens or aircraft sounds in a historical drama) it will be completely re-recorded in the studio bit by bit. Sometimes the sound track is re-recorded anyway for foreign language sales. Sometimes a sound is re-recorded because the source doesn't make the sound the viewer expects to hear, doesn't actually exist in nature, or the producer wants a different sound.

If I watch something on a TV set with headphones I often do not hear as much as sitting in my listening room over studio monitors playing the same program via the internet, dependent on the source of the files being watched.

If you go via the optical out of a TV into a good DAC you may get closer to that experience.

As far as I'm aware TV companies use SRS 3D (stereo source transferred into an expanded sound stage) and SRS TruSurround (up to 7.1 true speaker outputs.) at present to expand the stereo / surround image for broadcast which may not be reproduced that accurately over headphones as there is no cross talk from one transceiver to the other. In other words the right ear never receives audio from the left speaker and vice versa.

In the days of analog broadcast we used to have several speaker types from very cheap transistor radio speakers all the way upwards to test the mixes for clarity, reproduction of problem frequencies, and most importantly phase testing, but this can be done electronically now and with the disappearance of tape heads the need for phase testing beyond polarity of inputs is no longer required.

In my experience whatever you use to master on will never be exactly reproduced on another system unless the system reaches a required minimum standard such as the LucasFilm THX standard for film theaters. I'm not sure if broadcasters or TV manufacturers have got that far yet.

Ray.

Hi, yes, because the other thing I've noticed is that hard panned sounds can sound quite distracting and unrealistic when listening with headphones. Would normalizing the whole mix at the mastering stage help at all?

browj2 wrote on 11/25/2021, 9:29 AM

@jak.willis

Hi,

Would normalizing the whole mix at the mastering stage help at all?

How would you do that? Mixdown the audio and normalize?

You first have to understand what normalizing does. See the definition in the manual. Normalizing the project file would change the gain of the whole audio to the loudest sample of the entire project. That does not resolve parts that are too soft. Normalizing each object would give you parts that are too loud. You have to mix first to get the volume of each part to be what you want.

I have 3 video clips of me walking in the woods and you can hear the traffic in the background. In the first clip, I call the dog, thus the loudest part is my voice. In another, there is a slightly louder part than in the third clip. If I normalize each clip, the volume of 2 of them are greatly increased and the one with my voice slightly increased. Playing back, the background in the first clip is normal, but in the other 2 it is far too loud. I went back and adjusted the gain of the 2 objects that were too loud to be the same as the first clip. So, normalizing must be used intelligently.

Also, read through this thread to understand a bit more about analysing and adjusting the volume. And, I didn't touch on other things, like compression and EQ.

John CB

Last changed by browj2 on 11/25/2021, 9:32 AM, changed a total of 1 times.

John C.B.

VideoPro X(16); Movie Studio 2025 Platinum; Music Maker 2025 Premium Edition; Samplitude Pro X8 Suite; see About me for more.

Desktop System - Windows 10 Pro 22H2; MB ROG STRIX B560-A Gaming WiFi; Graphics Card Zotac Gaming NVIDIA GeForce RTX-3060, PS; Power supply EVGA 750W; Intel Core i7-10700K @ 3.80GHz (UHD Graphics 630); RAM 32 GB; OS on Kingston SSD 1TB; secondary WD 2TB; others 1.5TB, 3TB, 500GB, 4TB, 5TB, 6TB, 8TB, 12TB, 14TB; three monitors - HP 25" main, LG 4K 27" second, HP 27" third; Casio WK-225 piano keyboard; M-Audio M-Track USB mixer.

Notebook - Microsoft Surface Pro 4, i5-6300U, 8 GB RAM, 256 SSD, W10 Pro 20H2.

YouTube Channel: @JCBrownVideos

CubeAce wrote on 11/25/2021, 9:34 AM

@jak.willis

Hi Jak.

Normalizing only brings the loudest peak in a recording up to a preset arbitrary level. It has no effect on sound positioning.

Audio post production has now got quite clever and complex. I can get to hear things slightly behind me now to the left and right from a stereo transmission and correct speaker placement. I'm not sure that is the intent but more likely to widen a stereo image from a TV setup perhaps with just a soundbar in front of it.

When we hear things around us our ears pick up the delay between our left and right ears. This is how we get our positioning of an object around us. TV and sometimes Film normally only have one stereo mic above and slightly to the front placed central to where the camera angle is set up to shoot. Nearly everything else is added in post.

To do that, the mix must be performed on speakers so anything added can be placed accurately to match the ambience and placement. That means the sounds from the left source must be heard by the right ear etc. I can't say for certain all production is done that way in all countries but major production companies would do this.

That was the method used by the studio I worked at. Looking at a couple of post production studio websites in the UK it doesn't seem to have changed much judging by the images they post.

The equipment is digital but the layouts remain basically the same.

Ray.

 

Windows 10 Enterprise. Version 22H2 OS build 19045.5737

Direct X 12.1 latest hardware updates for Western Digital hard drives.

Asus ROG STRIX Z390-F Gaming motherboard Rev 1.xx with Supreme FX inboard audio using the S1220A code. Driver No 6.0.8960.1 Bios version 1401

Intel i9900K Coffee Lake 3.6 to 5.1GHz CPU with Intel UHD 630 Graphics .Driver version Graphics Driver 31.0.101.2135 for 7th-10th Gen Intel® with 64GB of 3200MHz Corsair DDR4 ram.

1000 watt EVGA modular power supply.

1 x 250GB Evo 970 NVMe: drive for C: drive backup 1 x 1TB Sabrent NVMe drive for Operating System / Programs only. 1X WD BLACK 1TB internal SATA 7,200rpm hard drives.1 for internal projects, 1 for Library clips/sounds/music/stills./backup of working projects. 1x500GB SSD current project only drive, 2x WD RED 2TB drives for latest footage storage. Total 31TB of 10 external WD drives for backup.

ASUS NVIDIA GeForce RTX 3060 12GB. nVidia Studio driver version 572.60 - 3584xCUDA cores Direct X 12.1. Memory interface 192bit Memory bandwidth 360.05GB/s 12GB of dedicated GDDR6 video memory, shared system memory 16307MB PCi Express x8 Gen3. Two Samsung 27" LED SA350 monitors with 5000000:1 contrast ratios at 60Hz.

Running MMS 2024 Suite v 23.0.1.182 (UDP3) and VPX 14 - v20.0.3.180 (UDP3)

M Audio Axiom AIR Mini MIDI keyboard Ver 5.10.0.3507

VXP 14, MMS 2024 Suite, Vegas Studio 16, Vegas Pro 18, Vegas Pro 21,Cubase 4. CS6, NX Studio, Mixcraft 9 Recording Studio. Mixcraft Pro 10 Studio. CS6 and DXO Photolab 8, OBS Studio.

Audio System 5 x matched bi-wired 150 watt Tannoy Reveal speakers plus one Tannoy 15" 250 watt sub with 5.1 class A amplifier. Tuned to room with Tannoy audio application.

Ram Acoustic Studio speakers amplified by NAD amplifier.

Rogers LS7 speakers run from Cambridge Audio P50 amplifier

Schrodinger's Backup. "The condition of any backup is unknown until a restore is attempted."