diff options
author | Sacha Chua <sacha@sachachua.com> | 2022-11-03 15:36:38 -0400 |
---|---|---|
committer | Sacha Chua <sacha@sachachua.com> | 2022-11-03 15:36:38 -0400 |
commit | 01ba6a4cc24862299f48288014acfa12010012be (patch) | |
tree | 56e973e8943d3a069f3e3d6b6655c04080278c2a /2022/organizers-notebook | |
parent | 29687c30adba8cf7955fa597d9f791921c09d89f (diff) | |
parent | 734a769475bd24f8dec0af446f85c95669a45107 (diff) | |
download | emacsconf-wiki-01ba6a4cc24862299f48288014acfa12010012be.tar.xz emacsconf-wiki-01ba6a4cc24862299f48288014acfa12010012be.zip |
Merge branch 'master' of git://git.emacsconf.org/emacsconf-wiki
Diffstat (limited to '')
-rw-r--r-- | 2022/organizers-notebook.md | 123 | ||||
-rw-r--r-- | 2022/organizers-notebook/index.org | 120 |
2 files changed, 242 insertions, 1 deletions
diff --git a/2022/organizers-notebook.md b/2022/organizers-notebook.md index ef5303fe..7b432a22 100644 --- a/2022/organizers-notebook.md +++ b/2022/organizers-notebook.md @@ -2327,6 +2327,127 @@ before the conference! Sacha Chua +### Mastering the prerec’s audio-track + +Mastering is the process of preparing an audio-track for a purpose. For +us, the purpose is quite simple: maximize the intelligibility of the +speaker and minimize the noise. + +We can get great results with Audacity for the vast majority of +audio-tracks. Sometimes, however, some audio-tracks have intractable +noise-profile that require the use of model-based denoising filters that +can applied with ffmpeg. + +We’ll start with the average Audacity workflow, and we’ll move on to the +model-based filters after. + + +#### Audacity workflow + +When we process a prerec, we extract the audio of the original upload +and add it to the backstage. You should be able to find it under the +name –original.$audio\_format. If it’s not there, it’s easy to extract +the audio from the original video, but we’d prefer if you warned +core-organizers about it because it’s not normal. + +We’ve simplified the process down to these steps: + +1. Open the audio file in Audacity. + + You might want to increase the size of the waveform by pulling on the + bottom of the bottom of the track. + + [audacity-demo-noise-reduction.webm](https://media.emacsconf.org/misc/audacity-demo-noise-reduction.webm) + +2. Find a moment of quiet in the video, and select it. + + We ask our speakers to include 5 seconds of quiet at the beginning or + end of their prerecs, but even if they don’t, it’s relatively. + +3. Effects → Noise Reduction → Get Noise Profile + +4. Select → All + +5. Effects → Noise Reduction → OK + + You can select a spoken portion of the track before applying the + effect and preview it to test your settings. The default are usually + enough (Noise reduction (dB): 12, Sensitivity: 6.00, Frequency smoothing + (bands): 3). + + [audacity-demo-noise-reduction.webm](https://media.emacsconf.org/misc/audacity-demo-noise-reduction.webm) + +6. Tools → Apply Macro → Alpha + + Before you can apply the Alpha macro, you need to save its content to + disk and import it via Tools → Macro Manager → Import. + + Reverb:Delay="20" DryGain="5" HfDamping="99" Reverberance="15" RoomSize="70" StereoWidth="25" ToneHigh="0" ToneLow="100" WetGain="-13" WetOnly="0" + Amplify:Ratio="1" + FilterCurve:f0="79.621641" f1="101.02321" FilterLength="8191" InterpolateLin="0" InterpolationMethod="B-spline" v0="5.9148936" v1="0.042552948" + Normalize:ApplyGain="1" PeakLevel="-3" RemoveDcOffset="1" StereoIndependent="1" + Compressor:AttackTime="0.1" NoiseFloor="-50" Normalize="1" Ratio="2" ReleaseTime="1" Threshold="-30" UsePeak="0" + +1. Export → Export Audio… → Opus Files (.opus format) + + Use the following settings: + + [audacity-export-settings.png](https://media.emacsconf.org/misc/audacity-export-settings.png) + + > Bit Rate: 64 kbps + > VBR Mode: On + > Compression: 10 + > Application: Audio + > Frame Duration: 20 ms + > Cutoff: Disabled + + +#### Model-based denoising filter + +If you can’t manage to get a good result with Audacity, chances are it’s +because there’s too much noise in the video, even after profile-based +denoising. This usually happens when the noise-pattern of an +audio-track evolves over the video, or if has an aperiodic quality. For +those, we’re going to need a bigger boat. + +Model-based denoising means using an AI-generated model to remove the +audio frequencies that are usually associated to noise and preserve +those that aren’t. A different context (e.g. noisy room with statics, +noisy room with people chatting, etc.) means a different model; for us, +this means a model that minimizes background noise and maximizes clear +voices (the speakers’). + +This is the model we’ve been using: + +[audio-denoiser-model-mp.rnnn](https://media.emacsconf.org/misc/audio-denoiser-model-mp.rnnn) (download link) + +Source: [rnnoise-models](https://github.com/GregorR/rnnoise-models), Model: [marathon-prescription](https://raw.githubusercontent.com/GregorR/rnnoise-models/master/marathon-prescription-2018-08-29/mp.rnnn) + +You should always apply the filter on the original’s audio, as opposed +to an Audacity-processed audio. This is to ensure that we have the most +information about the signal, which means we can have gather the most +information about the noise-profile. + +Following is the ffmpeg incantation to use to apply the filter-model. +Make sure to modify the `DENOISER` variable and adapt input/output. + + DENOISER="/path/to/audio-denoiser-model-mp.rnnn" + input="original.opus" + output="denoised.opus" + ffmpeg -i "$input" -af "arnndn=m=$DENOISER" "$output" + +There’s no need to customize the libopus export information; the default +is more than enough for human-speech. + +When you’re done with this step, you can then process the outputted +audio-track with Audacity, skipping the denoising steps (1 to 5). + + +#### Questions? + +If you’ve got any question on the process, you canget in touch with me (zaeph)! + + <a id="when-captioned"></a> ## When a talk is captioned @@ -3799,7 +3920,7 @@ Where: Nice if there’s an Ansible playbook sachac’s notes: - <file:///home/sacha/code/docker/emacsconf-publish/> + <file:///home/zaeph/code/docker/emacsconf-publish/> - probably good to set it up on front It’s now on front. diff --git a/2022/organizers-notebook/index.org b/2022/organizers-notebook/index.org index 8e3b284f..f7edc169 100644 --- a/2022/organizers-notebook/index.org +++ b/2022/organizers-notebook/index.org @@ -1658,6 +1658,126 @@ EmacsConf ${year}, and thank you for submitting the prerecorded video before the conference! Sacha Chua +*** Mastering the prerec’s audio-track +Mastering is the process of preparing an audio-track for a purpose. For +us, the purpose is quite simple: maximize the intelligibility of the +speaker and minimize the noise. + +We can get great results with Audacity for the vast majority of +audio-tracks. Sometimes, however, some audio-tracks have intractable +noise-profile that require the use of model-based denoising filters that +can applied with ffmpeg. + +We’ll start with the average Audacity workflow, and we’ll move on to the +model-based filters after. + +**** Audacity workflow +When we process a prerec, we extract the audio of the original upload +and add it to the backstage. You should be able to find it under the +name --original.$audio_format. If it’s not there, it’s easy to extract +the audio from the original video, but we’d prefer if you warned +core-organizers about it because it’s not normal. + +We’ve simplified the process down to these steps: + +1. Open the audio file in Audacity. + + You might want to increase the size of the waveform by pulling on the + bottom of the bottom of the track. + + [[https://media.emacsconf.org/misc/audacity-demo-noise-reduction.webm][audacity-demo-noise-reduction.webm]] + +2. Find a moment of quiet in the video, and select it. + + We ask our speakers to include 5 seconds of quiet at the beginning or + end of their prerecs, but even if they don’t, it’s relatively. + +3. Effects → Noise Reduction → Get Noise Profile + +4. Select → All + +5. Effects → Noise Reduction → OK + + You can select a spoken portion of the track before applying the + effect and preview it to test your settings. The default are usually + enough (Noise reduction (dB): 12, Sensitivity: 6.00, Frequency smoothing + (bands): 3). + + [[https://media.emacsconf.org/misc/audacity-demo-noise-reduction.webm][audacity-demo-noise-reduction.webm]] + +6. Tools → Apply Macro → Alpha + + Before you can apply the Alpha macro, you need to save its content to + disk and import it via Tools → Macro Manager → Import. + +#+begin_src txt :eval no :tangle audacity-macro-alpha.txt +Reverb:Delay="20" DryGain="5" HfDamping="99" Reverberance="15" RoomSize="70" StereoWidth="25" ToneHigh="0" ToneLow="100" WetGain="-13" WetOnly="0" +Amplify:Ratio="1" +FilterCurve:f0="79.621641" f1="101.02321" FilterLength="8191" InterpolateLin="0" InterpolationMethod="B-spline" v0="5.9148936" v1="0.042552948" +Normalize:ApplyGain="1" PeakLevel="-3" RemoveDcOffset="1" StereoIndependent="1" +Compressor:AttackTime="0.1" NoiseFloor="-50" Normalize="1" Ratio="2" ReleaseTime="1" Threshold="-30" UsePeak="0" +#+end_src + +7. Export → Export Audio… → Opus Files (.opus format) + + Use the following settings: + + [[https://media.emacsconf.org/misc/audacity-export-settings.png][audacity-export-settings.png]] + + #+begin_quote + Bit Rate: 64 kbps + VBR Mode: On + Compression: 10 + Application: Audio + Frame Duration: 20 ms + Cutoff: Disabled + #+end_quote + + +**** Model-based denoising filter +If you can’t manage to get a good result with Audacity, chances are it’s +because there’s too much noise in the video, even after profile-based +denoising. This usually happens when the noise-pattern of an +audio-track evolves over the video, or if has an aperiodic quality. For +those, we’re going to need a bigger boat. + +Model-based denoising means using an AI-generated model to remove the +audio frequencies that are usually associated to noise and preserve +those that aren’t. A different context (e.g. noisy room with statics, +noisy room with people chatting, etc.) means a different model; for us, +this means a model that minimizes background noise and maximizes clear +voices (the speakers’). + +This is the model we’ve been using: + +[[https://media.emacsconf.org/misc/audio-denoiser-model-mp.rnnn][audio-denoiser-model-mp.rnnn]] (download link) + +Source: [[https://github.com/GregorR/rnnoise-models][rnnoise-models]], Model: [[https://raw.githubusercontent.com/GregorR/rnnoise-models/master/marathon-prescription-2018-08-29/mp.rnnn][marathon-prescription]] + +You should always apply the filter on the original’s audio, as opposed +to an Audacity-processed audio. This is to ensure that we have the most +information about the signal, which means we can have gather the most +information about the noise-profile. + +Following is the ffmpeg incantation to use to apply the filter-model. +Make sure to modify the ~DENOISER~ variable and adapt input/output. + +#+begin_src sh :tangle audio-denoiser.sh +DENOISER="/path/to/audio-denoiser-model-mp.rnnn" +input="original.opus" +output="denoised.opus" +ffmpeg -i "$input" -af "arnndn=m=$DENOISER" "$output" +#+end_src + +There’s no need to customize the libopus export information; the default +is more than enough for human-speech. + +When you’re done with this step, you can then process the outputted +audio-track with Audacity, skipping the denoising steps (1 to 5). + +**** Questions? +If you’ve got any question on the process, you canget in touch with me (zaeph)! + ** When a talk is captioned :PROPERTIES: :CUSTOM_ID: when-captioned |