Welcome to Overclockers Forums! Join us to reply in threads, receive reduced ads, and to customize your site experience!
32M is memory dependent... big time.. trash it.
Not sure on 1M..Cache amount and speed is different, etc. but, unless you have the systems next to each other and the settings, comparing like this is anecdotal at best.
Again, IPC wasn't tauted here a couple % at best between them.
Subtimings matter.........
The cache is full in CL and again, its less.
I never expected IPC increases. And due to differing cache, etc, I can imagine these differences are about right.
I'm thinking of the generic case. If you have N threads available, can you scale work to optimally use all N threads? The short answer is no. Until it gets closer to a universal yes, throwing more cores at things aren't going to help as much as you'd like.
Do you run an operating system? If so, there will be other processes to use those N available threads. Windows itself runs 40+ processes. Would you rather have context switching on 4 fast and hot cores, or everything running just fine on 16 slow and cool cores that are still more than enough for the tasks assigned to each?
Do you run an operating system? If so, there will be other processes to use those N available threads. Windows itself runs 40+ processes. Would you rather have context switching on 4 fast and hot cores, or everything running just fine on 16 slow and cool cores that are still more than enough for the tasks assigned to each?
For general consumer level tasks, I'd argue 4 not-slow cores are generally plenty for a good experience. It would be likely that 16 slow cores will give a worse general experience and most of those extra cores will sit around mostly idle while fewer cores are maxed out for longer. All this complicated by variable active core turbo... I have a 14 core Xeon system, and it definitely doesn't feel that as great on a desktop as consumer quads. The OS background tasks are generally minimal impact, outside of a Windows Update which cripples anything no matter how fast.
Let's take the ideal scenario. If you had perfect core scaling, the total throughput is proportional to cores * clock (* architecture adjustment * HT/SMT benefit). When we buy more cores, we aren't just buying more cores. We're buying a bigger number of core * clock, otherwise more cores wont be worth it. I've done a lot of testing in my specific compute areas of interest (comparable workloads to Prime95, linpack etc.). Multi-threading helps but the cases it helps are not universal. It can shift the bottleneck to other areas like disk, cache or ram, or simply showing up weaknesses in the software. For a given total potential, fewer faster cores will generally give more real world throughput than an equivalent case of more slower cores, all else being equal. Yes, faster cores tend to be less power efficient. But more cores = more silicon area = more cost too. It is a tough balance and we can only buy what is right for us at a given time.
That's what "turbo" is for already But really, x86 needs to somehow implement the "big.little" thing ARM has. I'd love to have 4 fast cores for games and 12 low-power cores to deal with all the background processes.
Intel Will No Longer Provide Per-Core Turbo Frequencies, Making Motherboard Tuning Impossible - https://www.extremetech.com/computi...quencies-making-motherboard-tuning-impossible
Intel Will No Longer Provide Per-Core Turbo Frequencies, Making Motherboard Tuning Impossible - https://www.extremetech.com/computi...quencies-making-motherboard-tuning-impossible
Point is that they did barely anything. Most performance gains are from faster internal connections and cache. Their marketing is working great and it's not hard to convince most users that new is better. If you were on any Intel event then you know how much BS they can say in one hour I will only say one thing what I heard ... Intel IGP is good graphics for gamers ... list is long.
As it was already said, larger cache = slower cache. They have to improve it in other way to make it at the same time faster and larger but it's not so easy. It's also one of the reasons why AMD FX was so slow ... while Intel had 2-4MB L2 then AMD was pushing 8MB. Also difference in popular software like games is visible because of cache. Many people are complaining because of Skylake-X slow L3 cache.
Most changes are in chipsets, pcie lanes, additional features/controllers, lower power usage, more cores etc. Performance on single cores is not much higher than couple of years ago but there is higher turbo clock, faster memory controllers and some other things.
Finally 8th gen of mobile Intel CPUs will have more cores. Since first Core gen almost all i3 and i5 and most i7 were 2 cores. Now i5 are 2/4 or 4+HT so something is moving.
In desktops there are less changes than in other series. Now you can get 16 core Atom CPU which has 32W TDP but has full functionality of higher series. Couple of years ago at this TDP were 2-4 cores. For home/office users it doesn't matter but in business in global scale it's huge difference. In laptops in the same TDP you can get 2x more cores what makes possible to build ultrabooks with more than 2 cores. Even though performance per core isn't much better then overall it's a big difference.
So the real question is what will be better for gaming and general Windows stuff, a 7600k at 4.8GHz or a 8700k at whatever I can OC it to?
The second part is partly true. Many games are dependent on user input to control flow, and you can't multithread user input. You CAN multithread the effects of user input, though. If you've designed your AI to actually play well rather than just cheating to implement difficulty levels, in a simultaneous turn-based game where you're not supposed to know what actions everybody else took until the next turn, every AI player should be running in its own thread taking their turns at the same time. If you're writing a GUI app in any language, you're probably already using multithreading even if you don't know it; there's an "Event Dispatch Thread" that deals with diplaying the GUI and user input, and then there's the main thread that you're actually doing whatever work you have in.
The first part is not so true. If all schools from primary up did a better job of teaching logic (obviously not possible, because that would deplete various parties' gullible voter pools), and CS/IT schools did a better job of pointing out how various workloads can be separated into steps... Almost anything can be split at least to two threads if only people were taught what to look for.