AMD Fusion - A New Era of Computing Coming Soon!

At the AMD Technical Forum & Exhibition 2010, we were treated to more AMD Fusion demos as well as more information leading up to the availability of the Fusion ecosystem. If you're looking for the definite AMD Fusion guide and the latest updates, look no further. Plus, peer into the future of the Fusion.

New Experiences at AMD Technology Forum and Exhibit (TFE 2010)

In the rain-drenched city of Taipei this time of the year in Taiwan, AMD held it's sixth annual Technology Forum and Exhibit (TFE 2010) today and it goes without saying that the HardwareZone team was present to soak it up (literally speaking). While most tech enthusiasts are more than aware of the imminent launch of the Radeon HD 6000 series, we'll save those details for a forthcoming article as it's still under embargo. Well, apart from a senior AMD executive showing off the Radeon HD 6870 that's coming in a few days:-

Matt Skynner, Corporate VP and GM for AMD Graphics Division showing off the new Radeon HD 6870. Stay tuned for more info on HardwareZone on the 22nd October when it goes official.

Matt Skynner, Corporate VP and GM for AMD Graphics Division showing off the new Radeon HD 6870. Stay tuned for more info on HardwareZone on the 22nd October when it goes official.

Here's a quick sneak peek look at a Radeon HD 6800 series card used in one of the demo systems at the show. It certainly isn't any larger than the existing Evergreen series; thankfully.

Here's a quick sneak peek look at a Radeon HD 6800 series card used in one of the demo systems at the show. It certainly isn't any larger than the existing Evergreen series; thankfully.

So what's left to share today? AMD Fusion of course! , it was to have launched in 2009, but according to the latest update, it is now expected to be available early 2011. During the recent Computex 2010, AMD proudly showed off the world's first silicon wafer of the Fusion APU wafer, some basic details of what the APU is about and even gave a sneak peek of gaming performance with Alien vs. Predator game. You can catch all of these in our previous video here:-

As the days draw closer to 2011, you can expect more Fusion related publicity from AMD. Not a whole lot more was revealed today than we already are aware, but let us give you a quick rundown of the codenames and other aspects of Fusion is an easy to digest manner.

Chris Cloran, Corporate VP and GM of AMD's Client Division shared with the audience on the kind of experience people are enveloped in these days from the display type used, to content creation and consumption. All of these are essential aspects that AMD's Fusion strategy is expected to deliver to any consumer.

Chris Cloran, Corporate VP and GM of AMD's Client Division shared with the audience on the kind of experience people are enveloped in these days from the display type used, to content creation and consumption. All of these are essential aspects that AMD's Fusion strategy is expected to deliver to any consumer.

 

Fusion and APUs Explained

So, what is the AMD Fusion concept about? At the simplest level, the basic concept is the merging of the CPU and GPU processing capabilities into a single processing die. In other words, the incorporation of the GPU processing core into the CPU die to leverage on their combined processing advantage, reduce the CPU to GPU bottlenecks, and finally bring about reduced power savings from the platform level. The CPU is great at serial workloads whereas the GPU has more than proven itself for massively parallel workloads, thus the combination of both types of processing engines in one die struck AMD as a time to redefine this new class of chips as the Accelerated Processing Unit or APU for short (because each engine compliments each other to propel performance further than any one unit could have done alone). Furthermore, AMD has a few more goals for the APU and the slide below captures that nicely:-

AMD's design goals for their APU and the reason for embarking on this major project.

AMD's design goals for their APU and the reason for embarking on this major project.

Now that we've re-established what the Fusion hoo-ha is all about, let's get into some codenames that currently establish the various variants of the AMD Fusion that the company has planned in their pipeline and what's it made of. To keep it simple, we've arranged the information in this table:-

Platform code name
Brazos
Sabine
APU code name
Ontario
Zacate
Llano
Manufacturing Process
40nm (bulk)
40nm (bulk)
32nm SOI
TDP Rating
9W
18W
25W
CPU Core Architecture
Bobcat
Bobcat
K10 (Propus) derivate
No. of CPU Cores
1 to 2
2
2 to 4
Built-in GPU
Yes
Yes
Yes
GPU Class
DX11 (ATI Cedar derivate)
DX11 (ATI Cedar derivate)
DX11 (ATI Redwood derivate)
Target Segment
Low Power platforms
Mainstream
Performance-Mainstream
Target Products
  • Netbooks
  • Ultrathin notebooks
  • Compact Computing
  • Mainstream Notebooks / Desktops
  • AIO Desktops
  • Mainstream Notebooks / Desktops
  • AIO Desktops
Availability
Q1 - 2011 onwards
Q1 - 2011 onwards
Q2 - 2011 onwards

 During this event, AMD further showcased a few more demos to whet our appetite of Fusion's capabilities. Whether they remain competitive upon actual availability remains to be seen as Intel has their new Sandy Bridge based processors due early in 2011 too and vie for a similar segment. For now take a look at the following video we stringed together to that captures AMD's Brazos platform (with a Zacate APU) facing off Intel's Core i5 mobile platform, as well as the Llano APU in heavy multitasking. The video ends off highlighting Fusion's role in the near future:-




The Brazos Platform (with Ontario, Zacate APUs) and the Bobcat Core

The Brazos platform is AMD's ticket to the netbook platform and other ultrathin notebooks. AMD has always maintained that they didn't see a point to enter these segments previously as they couldn't deliver their balanced platform perspective and nor is the competition doing such. Sure, Intel was ruling the roost with low-voltage Core 2 Duo and Atom processors, but its accompanying graphics was abysmal and often needed third-party intervention from AMD or NVIDIA to better equip this class of systems for multimedia needs. AMD on the other hand, was busy attacking the desktop front and didn't have an ideal architecture or resources to counter Intel's move into the compact notebooks space.

Brazos will finally plug this gap quickly and will do so with a more optimized CPU architecture and a decent DX11 GPU for a balanced solution - thus the Ontario APU design was born. This 9W TDP part targeted at netbooks and ultrathin notebooks will be complemented by a more powerful Zacate 18W APU variant for more mainstream notebooks and desktops.

A quick look at the Brazos platform which will span two segments using an Ontario APU and Zacate APU.

A quick look at the Brazos platform which will span two segments using an Ontario APU and Zacate APU.

The basic building blocks for both the Ontario and Zacate APU come in the form of the Bobcat CPU core architecture. This purpose designed core attacks the low power computing needs, as seen in the below diagram. It's also modular in nature allowing further cores to be fused together to create more options for various price and TDP points. For example, the Ontario is rumored to come in either single or dual-core variants (basically two bobcat cores) while the Zacate will be a faster dual-core processor. AMD can create further APU models with more cores if it requires, but for now, these two models are official.

The bobcat core architecture is the building block for Ontario and Zacate APUs in the Brazos platform aimed at the netbooks, ultrathin notebooks and other compact form factor devices/systems.

The bobcat core architecture is the building block for Ontario and Zacate APUs in the Brazos platform aimed at the netbooks, ultrathin notebooks and other compact form factor devices/systems.

Of interest is the out-of-order execution pipeline of the Bobcat architecture which places it ahead of the Intel Atom processor. All things being equal, the out-of-order pipeline (used by most mainstream processors) immediately brings about better processing performance than an . Plus, the sub one-watt operation of the core is again ideal for very compact form factors and devices. In fact, this core is likely the ideal candidate for the tablet market when it matures somewhat with more developmental prospects. Currently the tablet market is just too fragmented for AMD to make any significant push on its own.

Having talked about the CPU portion in these APUs, the GPU portion would be most people's main interest - just what sort of graphics engine is AMD packing into them? The good thing is that for all the APU variants, AMD will stick to an Evergreen class (Radeon HD 5000 series) DX11 compliant GPU engine. Remember, AMD wants to deliver discrete-class graphics but it also has to maintain low power and price in this sensitive segment of the Ontatio and Zacate APUs. As such, both of them will be equipped with Ceder-derivate graphics engines (essentially class). That doesn't sound very exciting, but it fulfills the design specs of having a discrete-class graphics built-in to the chip (with a low TDP to boot). Again, even GPU engines are rather modular in nature and like the CPU portion, they can be scaled if required.

So now that we're better acquainted with the basic design and specs, here's a slide from AMD that depicts what sort of needs the Brazos platform is designed to address and to better manage expectations of end-users:-

Multimedia content consumption needs, light productivity and basic fluid gaming are the scope of the Brazos platform - a perfect fit for current day compact and portable systems.

Multimedia content consumption needs, light productivity and basic fluid gaming are the scope of the Brazos platform - a perfect fit for current day compact and portable systems.



 

The Sabine Platform (with Llano APU)

Moving up the ladder for more performance, the Sabine platform is targeted at the performance-mainstream crowd who dabbles in more content creation activities and would like to have close to mid-tier discrete GPU performance. Answered by the Llano APU that will come in Q2 2011 (a quarter after the Zacate and Ontario), its CPU core is actually a derivative of the existing Athlon II processors and will be available in both dual and quad-core variants. The GPU portion is better equipped, possibly a Redwood-variant graphics engine (essentially a / class). This would be really interesting to benchmark when it does hit the market.

Key touch points of the Sabine platform with the Llano APU. "Lynx" is just the desktop equivalent codename for the mobile-oriented Sabine platform.

Key touch points of the Sabine platform with the Llano APU. "Lynx" is just the desktop equivalent codename for the mobile-oriented Sabine platform.

Here's a closer look at the Llano APU. Not much detail revealed at this level unfortunately. Note that the Fusion APUs still require some sort of hub controller on the motherboard, aka, chipset.

Here's a closer look at the Llano APU. Not much detail revealed at this level unfortunately. Note that the Fusion APUs still require some sort of hub controller on the motherboard, aka, chipset.

The requirements that make up the 32nm Llano APU, which is also AMD's first power-gated CPU. Hopefully this would mean a more power efficient APU that's able to shut down unused CPU cores completely like the modern Nehalem derivative cores in Intel's lineup currently.

The requirements that make up the 32nm Llano APU, which is also AMD's first power-gated CPU. Hopefully this would mean a more power efficient APU that's able to shut down unused CPU cores completely like the modern Nehalem derivative cores in Intel's lineup currently.

The first Llano APU 32nm SOI wafer held up by Chris Cloran, Corporate VP and GM of AMD's Client Division.

The first Llano APU 32nm SOI wafer held up by Chris Cloran, Corporate VP and GM of AMD's Client Division.



 

Looking Ahead on the Fusion Platform

In the near-term, the APU strategy is really going to benefit notebook platform the most with the higher integration and power savings aspect from the platform level. However, the biggest gains from the APU have yet to be untapped as it highly depends on the extent of collaboration and software design tools to take advantage of both the CPU and the GPU concurrently to execute the tasks - hence the combined acceleration of the APU. The APUs that AMD are launching have yet to be optimized for better programmability and when that happens, that's when the second wave of advantage awaits the AMD Fusion platform.

This slide depicts the development path for the APU in the near future - clearly, there's room to improve both the hardware and software aspects to get the best out of the APU with more optimized programming models for starters.

This slide depicts the development path for the APU in the near future - clearly, there's room to improve both the hardware and software aspects to get the best out of the APU with more optimized programming models for starters.

Before the conference ended, AMD also teased the audience with a die shot of their next generation APU, codenamed Orochi. It's also going to be a 32nm part like the Llano, but the CPU cores are going to be based on Bulldozer - not an existing K10 derivate nor the lower-end Bobcat. Bulldozer is the next generation core module (featuring two full integer execution cores) for upcoming Opeterons and high speed consumer processors sometime in 2011. The Orochi using this will have four Bulldozer core modules capable of addressing 8 threads with a total of '8 processing cores'.

The next generation APU revealed - Orochi. It will use four Bulldozer modules for a total of 8 processing cores to execute 8 threads.

The next generation APU revealed - Orochi. It will use four Bulldozer modules for a total of 8 processing cores to execute 8 threads.

 

Fusion Insights from Joe Macri - AMD's Corporate VP and CTO

Before we wrapped up our time in Taipei, we were privileged to have some time with Joe Macri, AMD's Corporate VP and CTO for Client Division, to get more insights of AMD's Fusion plans.

 

HWZ: Difference Between AMD's Fusion concept and the competition like Intel's current Core i3/i5 and upcoming Sandy Bridge processors?

Joe: There are some big differences between what we're doing and what our competitors are doing. When we say our tagline, the future is fusion, it actually embodies a lot more than just a marketing term. We really see the fusion architecture moving from this initial step out of time and evolving in a way where it's always backwards compatible. So we've a very cohesive vision as we look forward. When we compare our initial version of fusion that we're launching, to what Intel has done on the Sandy Bridge, we see differences. So Sandy Bridge is literally, I look at it and say a graphics unit put on the same piece of silicon as the CPU, but they are not cohesive in any way, shape or form. When you compare with Ontario, our graphics unit can, firstly, do compute; so you can actually have the CPU and GPU can work on the same problem. When you need floating point capabilities, the GPU can work on it, and when u need a lot of scalar capabilities to do a lot of integer processing quickly with low latency, the CPU can work on it. So right off the back, we have huge performance and power advantage over what they've done with Sandy Bridge.

The second thing we've done is that we want to have a cohesive architecture on the graphics side, which means we need to have the latest graphics architecture. We want our external GPUs and our internal GPUs to be able to work with each other. When putting in an external GPU, you don't want to turn off the internal one but rather, you would want it to work together on the same problem together. And if it's a visual problem, you've got to do it in ways like we do with CrossFire. And CrossFire means you've got to have a compatible architecture. We put in the latest graphics engine into the APU unlike Sandy Bridge and that's another differentiator. So the latest graphics programs are going to work; and if you look back at DX9 and DX10, we've got a software investment at AMD with the unified driver model that all your old stuff works well. That's why it took us a long time to give you something so extensible; to give you something that it's the latest and provide full compute capabilities on both sides. We could have taken short cuts to deliver some aspects of these earlier, but we couldn't figure out why it would be compelling for the end-users. On the Sandy Bridge, if you want to run any of the latest stuff, you need to plug in a discrete card and when you plug that in, your internal graphics just gets turned off because it's just incompatible with it. At least it would have been nice if the Sandy Bridge graphics could do compute, but instead it just gets turned off. It has a very good x86 core, but the rest of it, is just a 'hack'. The L3 cache is required if you're going to share data between two different compute units, but the GPU in Sandy Bridge doesn't do compute, so it's just there to hide latencies. GPUs don't need low latency, they need lots of bandwidth to a lot of memory. So their use of a cache is likely a design shortcut. The memory subsystem on the Fusion is very complicated as it needs to manage both the low latencies required by the CPU and the high bandwidth needs of the GPU.

The other thing is that we balance our GPU to the CPU on the APU. As we put in bigger CPUs, you get bigger GPUs, so that for most of the users in the world, the APU would be fine. But when you need to have an imbalanced system, such as gamers who need a lot of 3D performance, and when they stick in a discrete GPU, you don't turn off the internal GPU. You either use it in conjunction depending on the size of the external graphics or you use it for compute, and so you get a cohesive system.

As we look forward, we're going to enhance it in a way that will make the APU easier to program, easier to do compute, easier to share data between the CPU and GPU. But all these enhancements we're doing, all the old stuff is still going to work (such as the old DX games, x86 programs), code optimized for Ontario and Llano will still work; we never want to have anything be thrown away. As we go forward, we got a good vision that make the CPU and GPU be equally programmable, equally usable. It will allow the software writer to be an artist and not be an engineer. When the software writer thinks engineering, he's not thinking of the end vision he wants to give the consumers. What he's focused on his the data disparity, two different memory systems, how to manipulate them and ensure the latest data sets are obtained, etc. - we want to get rid of all that. We want to make it so simple for programmers to tackle the problem, make the memory models simpler, etc. So as we move forward, after Ontarion and Llano, you'll see this continuous enhancement, but it's all backwards compatible. And that's what I don't see out of our competitors.

HWZ: How would you get vendors to optimize programs for the APU specifically when the rest of the market is still on a conventional programming model?

Joe: We do it because our biggest metric is performance per watt. If you write code that runs on for the system yesterday, it will still run very well on an APU; actually in many cases it will only run on an APU if you write for DX11 while on Intel's Sandy Bridge, it just wouldn't work. We want to offer to the programmers a better path, but it is also compatible with the model of today and that's a really old model. Intel talks about the future of programming, though not in a cohesive vision but they talk about going to Larrabee, but it's a totally funny way of connecting x86 units to vector units; very wasteful of the hardware if you need a lot vector units instead of x86 units, all those x86 cores are going to be dead and if you need a lot of x86 units, those vector units will be dead. That's radically different from what they've got on Sandy Bridge, nothing in common in that programming model at all, maybe in some cases it won't even run because it's such a radical change. So I think architecturally, we're heading in the right direction, even if we've multiple models, our model is designed to work with them all. You can use it in a way to give you the best, but it will still run everything else just as good as the others.

 

HWZ: Will Fusion eventually replace low-end and mid-range discreet GFX?

Joe: Well, I think it's going to get redefined. As shown in the slides, today you have low-end discrete positioned above integrated graphics. And APUs aren't really integrated graphics since they bring about a whole new level of programming, in fact a whole new way of really looking at the platform. We're going to want our discrete GPUs to be above it. There's just a limit of how much graphics we can put into an APU; you've got to keep the chips cost effective. When you're buying an APU, you're basically buying into a balanced system. We'll want you to add more discrete graphics on to the system to tune it to your particular needs. So you'll find AMD products are always very cohesive but for the discrete graphics, we might still offer some low-end SKUs to match up to Intel's platform, because we make a lot of money selling them as Intel doesn't do such a good job at graphics. You can see that Sandy Bridge, they can't even run some of the latest stuff. So I think our low-end discrete will still ship on Intel (platforms) when it makes sense. We've no problems making money where they're not doing so good.

 

HWZ: Do you think Super Phones would bring the demise of mainstream computing?

HWM: No, not at all actually. I think ultraportable devices are very critical to all of our lifestyle. They don't replace what we see a PC can do today. The ability to manipulate content, create content, the ability to get the ultimate multimedia experience, these things will come down in form factor, but the user interface has to adapt with it. Otherwise you just won't be able to manipulate it. So I really believe in a cohesive set of devices. Desktop systems aren't going away, all the way down to the smallest form factors. My belief is that Fusion will span them, and span that with a common architecture, so that applications can move from one device to the next. So I think that's the real forward vision. I don't think cell phones and thin clients with limited compute capabilities are our future. Actually I think we need powerful computers everywhere; you want to move the computer to where the data is. Stuff like advanced user interfaces will have huge amounts of data that will be generated and will be thrown away, hence you'll want those computes locally. And when you advance the UI, you'll need more compute and be able to shrink form factors. I think the future is very bright, for both the super phones as well as larger form factors.

HWZ: So x86 will have a long shelf life?

Joe: ARM has no advantage. Don't let anyone tell you that ARM is superior to x86 in any way. I also won't say the other direction. What matters is how you design; it's not the ISA that truly differentiates a processing unit. What's beautiful of the x86 ISA is compatibility. That's the most important thing we've offered and you're going to continue to see that extend. I don't ever think twice about being able to outdo an ISA, it's just that there's no real advantage or disadvantage to it. The only disadvantage is that there isn't a lot out there; and what is out there, isn't very cohesive because the ecosystem doesn't want it. The guys at Blackberry don't want to work on an Apple, and vice-versa. So the ARM ecosystem is more about a business model, it's not anything, as an engineer, make me want to jump up and down.

Our articles may contain affiliate links. If you buy through these links, we may earn a small commission.

Share this article