Sorry to be so slow to respond, Kev.

I love the idea of identifying locations in a song via a frequency table mapped to some window.

Though, I think that the endpoint of this line of dev is to search for an arbitrary acoustic fingerprint within a song and call the location wherever it’s found the timecode.

The concept would be: here’s an acoustic fingerprint for a five second snippet, including patterns in frequency and amplitude and anything else you can find.

One way to identify frequency and amplitude patterns is to treat the acoustic data as if it were a face and use face recognition algorithms, e.g. principle components analysis.

A simpler method is to measure variance in some dimension over a fixed time window. So say that the identifier for a time segment is the difference between the highest and lowest amplitude.

It’s fun hacking. Makes me wish, yet again, that I did comp sci grad school.