Why is D unpopular?

Ola Fosheim Grøstad ola.fosheim.grostad at gmail.com
Mon May 2 11:19:18 UTC 2022


On Monday, 2 May 2022 at 09:23:10 UTC, Guillaume Piolat wrote:
> Sometimes artifacts sound "good", be it for cultural or 
> "objective" reason.

Yes, this is true. Like, the loudness-competition that lead to 
excessive use of compression (multiband?) and ducking (to let 
bass drum through) lead to a sound image that was pumping in and 
out. I personally find that annoying, but when you see kids 
driving in the streets playing loud music they seem to favour 
this "musically bad" sound. I guess they find excitement in it, 
where I think of it as poor mastering. And I guess in some genres 
it is now considered bad mastering if you don't use excessive 
compression.

I believe this loudness-competition and "overproduction" also has 
affected non-pop genres. If you get the ability to tweak, it is 
difficult to stop in time... I frequently find the live 
performances of talented singers on youtube more interesting than 
their studio recordings, actually.

The french music scene might be different? French "electro" 
seemed more refined/sophisticated in the sound than many other 
"similar" genres, but this is only my impression, which could be 
wrong.

> Many small delays can help a voice "fit in the mix", and 
> spectral leakage in a phase vocoder do just that. So some may 
> want to come through a STFT process just for the sound of 
> leakage, that makes a voice sound "processed" (even without 
> pitch change). Why? Because in a live performance, you would 
> have those delays because of mic leakage.

I hadn't thought of that. Interesting perspective about mics, but 
a phase vocoder have other challenges related to changing the 
frequency content. How would you create a glissando from scratch 
just using inverse FFT, it is not so obvious? How do you tell the 
difference between a click and a "shhhhhhh" sound? The only 
difference is in the phase… so not so intuitive in the frequency 
domain, but very intuitive in the time domain. You don't only get 
spectral leakage from windowing, you also can get some 
phasing-artifacts when you manipulate the frequency content. And 
so on…

But, the audience today is very much accustomed to electronic 
soundscapes in mainstream music, so sounding "artificial" is not 
a negative. In the 80s you could see people argue seriously and 
with a fair amount of contempt that electronic music wasn't real 
music… That is a big difference!

Maybe similar things are happening in programming. Maybe very 
young programmers have a completely different view of what 
programming should be like? I don't know, but I've got a feeling 
that they would view C as a relic of the past. If we were teens, 
would we then focus on the GPU and forget about the CPU, or just 
patching together libraries in Javascript? Javascript is actually 
quite capable today, so…

> The phase-shift in oversampling? It can make drums sound more 
> processed by delaying the basses, again. To the point people 
> use oversampling for processors that only add minimal aliasing.

I didn't understand this one, do you mean that musicians 
misunderstand what is causing the effect so that they think that 
it is caused by the main effect, but instead it caused by the 
internal delay of the unit? Or did you mean something else?

> Plus in the 2020s, anything with the sound of a popular codec 
> is going to sound "good" because it's the sound of streaming.

I hadn't though of that. I'm not sure if I hear the difference 
between the original or mp3 when playing other people's music 
(maybe the hi-hats). I do hear a difference when listening to my 
own mix (maybe because I've spent so many hours analysing it).



More information about the Digitalmars-d mailing list