I fired up one of the system management apps and as i remembered, that one has a "cache misses per second" parameter. Looked at two mildly cluttered forts (yes, that's totally unrepresentative):
first one - 84 FPS: cache misses/s typically ranged between 150 and 300, with rare spikes to 1000+. About 250 on average, i'd say
second one - 33 FPS: between 200 and 400 misses/s, spikes of just over 2000. Estimated average just above 300.
Naively, this looks to me like at this range, it's not cache misses that matter: assuming one cache miss costs 200ns, even 2000 in one second would amount to just 400µs, while the average even in the more-cluttered fort should sit comfortably below 100µs/s, i.e. less than 1% of 1% of operation time. I'd assume that purely cache-miss driven operation would be extreme lag spikes of under one FPS.
I'm almost certainly missing something fundamental, but i think it'd help if knowledgeable people tried to run some analytics on what influences FPS. I get the feeling we're still mostly dealing with assumptions that have little research behind them.
***
I've also looked at the parametrics on the old Core2Duo under Linux, and there, the manager shows a processor load of 100% on both cores, with no intensive process other than DF running. Curious...
Under Windows 10, with a 2C/4T processor, all threads show similar workloads, 25-40% mostly. The process gets moved between threads so often that load averages out over the measurement times. Thread switching is part of thermal management, i guess. The Core2 heats up from ~45 °C to ~75 when playing DF, and it's certainly better if it heats the cores evenly.