Generally speaking, the faster the processor, the higher the system-wide bandwidth and the lower the latency. As is always the case, faster is better when it comes to processors, as we’ll see below. But with Core i7, the game changes up a bit.
Whereas previous memory controllers utilized a dual-channel operation, Intel threw that out the window to introduce triple-channel, which we talked a lot about at August’s IDF. Further, since Intel integrates the IMC onto the die of the new CPUs, benefits are going to be seen all-around.
Before jumping into the results, we already had an idea of what to expect, and just as we did, the results seen are nothing short of staggering.
It goes without saying that i7 dominates this area, thanks to its triple-channel memory controller, but the overall FSB can play a huge role as well. Because we were forced to 800MHz, we took a huge hit on the bandwidth and latency. However, the problem pretty-much vanishes with light overclocking.
How fast can one core swap data with another? It might not seem that important, but it definitely is if you are dealing with a true multi-threaded application. The faster data can be swapped around, the faster it’s going to be finished, so overall, inter-core speeds are important in every regard.
Even without looking at the data, we know that Core i7 is going to excel here, for a few different reasons. The main is the fact that this is Intel’s first native Quad-Core. Rather than have two Dual-Core dies placed beside each other, i7 was built to place four cores together, so that in itself improves things. Past that, the ultra-fast QPI bus likely also has something to do with speed increases.
This Sandra test gives pretty expected results. The faster the CPU, the lower the latency and faster the core negotiations. This is one area where i7 excels, thanks in part to its ultra-fast QPI bus.