When Intel first announced the shift from the dual-stage ‘Tick-Tock’ methodology that had driven their core CPU design teams for over a decade into the three-way ‘Process, Architecture, Optimization’ paradigm, there were questions as to how much of the final stage, the Optimization, would actually change the way CPUs were presented or offer upgrades in performance. At the time, ‘Kaby Lake’ as the name for Intel’s third crack at their 14nm process was well known, but users wondered if it was just another Devil’s Canyon (better overclocking) or a full on CPU launch. Well the answer is the latter, and the launch is staggered between today and January. Roll on Kaby Lake. Is that Kah-bee Lake, or Kay-bee Lake?

Tick-Tock to Process-Architecture-Optimization (PAO)

Intel’s two year processor cadence of Tick-Tock allowed the company to alternate reducing the lithography node (a tick) with an upgraded microarchitecture (a tock) with modest gains of performance. Each stage of the Tick-Tock would take 12-15 months, with new lithography nodes taking longer and longer to reach maturity. Eventually it has gotten to a point where 14nm took too long to become pervasive in Intel’s product stack. For a manufacturing company and a CPU company, slowly rolling out Broadwell notebook and low power parts and then gutting the mid-range desktop was the result of the troubles of bringing 14nm to the prime time. Now in 2016 we have the launch of full-sized Broadwell-EP server parts running up to 22 cores in a 145W envelope, but the low power parts are a couple of microarchitectures ahead. We have this strange situation of a stretched landscape, partially driven by markets but also for manufacturing.

The move from Intel’s 14nm process to 10nm is a long, slow burn, taking much longer to develop than any process previous. We first saw 14nm in Broadwell mobile CPUs in Q3 of 2014, almost two years ago, and 10nm still isn’t on any public roadmap for early 2017. With the tick-tock strategy, and the launch of Skylake in mid-2015, this would leave Intel without a new CPU launch for almost two years, which is unheard of from Intel. Also, Intel’s partners rely on product cycles to announce and launch new products to generate revenue. To fill the gap, plus with a few other techniques, Intel moved to Process-Architecture-Optimization, or PAO for short.

This means three bites of the cherry for 14nm before we see 10nm in prime time. First it was Broadwell, then Skylake, and today Intel is announcing Kaby Lake. As part of the release, Intel has mentioned that a number of key benefits for Kaby Lake will be based on an optimized 14nm process, called 14PLUS (or 14nm+, 14FF+). This process as a quick summary has a higher fin height and larger pitch, essentially giving a less-dense set of transistors that have more room to breathe. Normally a larger pitch means more voltage required, but this is offset by the fin height and Intel says is good for another few hundred MHz for performance. The less-dense design, in theory, may also help in overclocking, however we will have to wait until January to see those results.

Today’s Announcement

The reveal today covers several aspects of Kaby Lake. First is the actual SKUs that will be launched, which consist of three Kaby Lake-Y parts around 4.5W aimed at high-end tablets and 2-in-1 devices and three Kaby Lake-U parts at 15W for notebooks. Both sets of KBL-Y and KBL-U CPUs will feature in mini-PCs as well, so we expect to see the usual array of Zotac and ECS announcements in due course. We have details on all six CPUs to give you, including the new Core M branding regime for the 4.5W family of parts.

The pre-briefings we have had go into some detail regarding changes in the processor, particularly regarding the new enhanced media blocks inside the GPU to support new encode/decode features. The graphics power, in terms of EUs or the microarchitecture, hasn’t changed but the fixed function hardware has some nice upgrades for an updated version of Gen 9 graphics aimed at the upcoming era of 4K support. Ganesh has gone into detail for us over what that means, especially where power and battery life is concerned for anyone creating/consuming 4K content.

Kaby Lake also has an updated Speed Shift package, to accelerate the work already done with Skylake in boosting the performance of the CPU quicker to save power. Intel call it a refinement in the mechanism of handing frequency control back from the OS to the CPU, however ‘Speed Shift v2’ is an adequate moniker to show the upgraded difference.

Intel hasn’t gone into much detail regarding the new 14nm+ process itself in terms of specifics, but has listed a number of performance gains that come out of the new CPU. The fundamental microarchitecture between Skylake and the new Kaby Lake parts is practically unchanged (DMI 3.0 now allows PCIe 3.0 x4 NVMe drives from the integrated PCH), but the updated fin profile and reduced ‘strain’ by the larger fin pitch is being quoted as giving a 12% performance increase due to process alone, typically through additional frequency for the same power. The main benefits to KBL will be in that frequency due to the 14nm+ process as well as the new media capabilities.

As always, Intel consistently emphasizes the difference between a new notebook/2-in-1 based on a 15W Kaby Lake processor compared to a 5-year old device, such one based on a mobile Sandy Bridge ULV part. Intel continually sees its market revolving around new experiences created by new form factors, and to generate revenue requires people to upgrade and decide how these new experiences can influence an upgrade. Intel expects to have over 100 Kaby Lake system designs in the consumer channel by Q4, including 120+ using Thunderbolt 3, 100+ using Windows Hello (either via a Real Sense camera or a fingerprint sensor), 50+ designs with 4K UHD as a primary feature and 25+ designs with contact enabled pens. Kaby Lake should be pin compatible with Skylake based designs, meaning that for a number of cases we will simply see a drop in, but others will have new design IDs on show.

The New CPUs, Updates to Core M Branding
Comments Locked

129 Comments

View All Comments

  • hansmuff - Tuesday, August 30, 2016 - link

    Does any of the new fixed-function logic that is part of the GPU get to work when I use a discrete GPU instead of the integrated?

    I remember that on my old SB chip, the GPU just was turned off because I use discrete. How have things changed, if at all?
  • Ryan Smith - Tuesday, August 30, 2016 - link

    Typically you'll be using the dGPU for video decoding since it's closer to the display pipeline. However you can totally use QuickSync for video encoding, even with a dGPU.
  • hansmuff - Tuesday, August 30, 2016 - link

    Ah yes, QuickSync in particular was a question for me. While NVENC certainly does do a fine job, if I have a hardware encoder laying dormant in the CPU, it might as well do stream encoding for me :)
  • fabarati - Tuesday, August 30, 2016 - link

    I just messed about with NVENC, QSVEncC and x265 when ripping som DVDs. X265 still gives the best quality and size. With a i5-6500, the encoding speed wasn't all that, at around 65 fps. Of course, QSVEncC was closer to 200 fps and NVENC (GTX 1070) clocked in at 1300-2000 FPS.

    Quality and size of the file are of course the opposite, with x265 looking the best and being the smallest, then QSVEncC and finally NVENC.
  • Guspaz - Tuesday, August 30, 2016 - link

    Can you? Last I looked, that required enabling both the dGPU and iGPU simultaneously (and simply not plugging a monitor into the iGPU). Attempts to enable the iGPU while having a dGPU plugged in on my Ivy Bridge resulted in Windows not booting.
  • nathanddrews - Tuesday, August 30, 2016 - link

    I can't speak for your system, but my Z77 motherboard features Virtu multi-GPU support that allows me to use Quick Sync while having my monitor plugged into my dGPU. You have to activate both IGP and dGPU in BIOS, then load both drivers. It worked for me under W7 and W10.
  • Guspaz - Tuesday, August 30, 2016 - link

    Errm, you've got dedicated hardware specifically for the purpose of supporting multiple GPUs (the Lucid Virtu), so that's not really a typical example.
  • extide - Tuesday, August 30, 2016 - link

    Lucid Virtu is all software
  • Gigaplex - Tuesday, August 30, 2016 - link

    Last I checked, it requires motherboard support. You can't just install some software and expect it to work. That's what they meant by dedicated hardware.
  • CaedenV - Tuesday, August 30, 2016 - link

    yep, there is a chip that enable the virtu stuff. It is little more than a soft-switch to route traffic to the right chip, but still required for the software to work.

Log in

Don't have an account? Sign up now