“‘We believe the Android operating system will be the most important console operating system in the future,’ Nvidia CEO and co-founder Jen-Hsun Huang said at the company’s CES 2014 press conference tonight. Then he effectively declared war on next-generation console gaming by announcing the Tegra K1, the company’s next mobile chip with a whopping 192-cores,” Jason Evangelho reports for Forbes.
“Prior to the announcement, Jen-Hsun caught up the audience with the company’s Tegra milestones like the Tegra 3 being the first mobile quad-core processor. ”So what can we do next? We could do 8 cores, but that seems a little pedestrian,” Jen-Hsun joked. Then he unveiled the 192-core Tegra K1, rooted in Kepler architecture,” Evangelho reports. “According to Nvidia, Tegra K1 offers 3x the performance of Apple’s A7 processor.”
“Multiple presentations showed off photo-realistic environments, dynamic lighting, and next-generation level graphics. From what I’ve seen, they have bridged the gap between next-generation console gaming and mobile gaming,” Evangelho reports. “Tegra K1 will be offered in two pin-to-pin compatible versions: a 32-bit quad-core (4-Plus-1 ARM Cortex-A15 CPU) and a custom, Nvidia-designed 64-bit dual ‘Super Core’ Denver CPU.”
Read more in the full article here.
MacDailyNews Note: According to NVIDIA (whitepaper, “The architecture of the Kepler GPU in Tegra K1 is virtually identical to the Kepler GPU architecture used in high-end systems, but also includes a number of optimizations for mobile system usage to conserve power and deliver industry-leading mobile GPU performance. While the highest-end Kepler GPUs in desktop, workstation, and supercomputers include up to 2880 single-precision floating point CUDA cores and consume a few hundred watts of power, the Kepler GPU in Tegra K1 consists of 192 CUDA cores and consumes less than two watts (average power measured on GPU power rail while playing a collection of popular mobile games).
[Thanks to MacDailyNews Reader “Take To Task” for the heads up.]
And what’s the battery life?
It’s great. Might be a bit cumbersome to carry a car battery in your backpack, but it’s great. 🙂
With 6 double A batteries, you’ll get 2 ¼ hours of pure gaming joy.
At 2 watts I expect a hell of lot better than the piss poor iPhone battery performance.
Freaking posts – this was supposed to be a reply to Bob, who for some reason has 5 star votes on that comment by people who obviously don’t realize the K1 has significantly less power consumption than A7. Talk about blind faith. I would think everyone here has an iPhone and is sick of the poor battery performance.
Nah. We’re just sick of trolls that don’t belong here.
But I do belong. I own apple products and have every right to speak the truth as much as anyone else here, including you ! Everyone who up voted that comment is obviously unaware the K1 uses much less power then A7.
And based on reactions to the newer story about iPhone 6 thickness, it is quite clear that many iPhone owners are fed up of important battery life.
Argh Siri ! My bad for clicking too fast. Last sentence should read “poor battery life”, because it is…
Nothing poor about my phone’s performance, considering the stuff it’s expected to do. The only times my phone’s battery suffers is when I’ve been taking lots of photos, using mapping software,mand forgotten to turn the screen off when I put it back in my pocket; I don’t enable auto-screenlock, so occasionally suffer.
I do, however, carry a pack with a 10,000mAh battery in to charge my phone, pad and camera with, and any other accessories, something any intelligent person should be doing who uses and relies on modern consumer electronics.
Instead of snivelling like a little girl, ivid, perhaps you ought to invest twenty or thirty bucks, and then shut up your whining.
That’s your best ? You have to resort to insulting me? Give me a break. No one should have to haul around gear to extend battery life, that’s ridiculous. In fact that just validates what I am saying.
Having to charge our phones daily or more is lame. Everyone knows this and sorry if you can’t handle that truth but don’t insult me for speaking the same truth that many many iPhone owners complain about on this very site all the time! Extending battery life is one of the top feature requests every year. I wonder why?
I speak truth and get insulted, I certainly didn’t “whine” in any way. What a wonderful example you are.
You just select the post in question and click reply it’s not rocket science. Hope your information about battery consumption is more reliable mind.
Being born stupid doesn’t mean you have to be obnoxious also.
I apologize if I came off obnoxious by speaking the truth. We all know iPhone battery is terrible (as competition’s may too but I wouldn’t know). What’s obnoxious to me is people jumping on the K1 assuming it will have terrible battery life when MDN clearly added a note to explain it takes a mere 2 watts ! WTF is so obnoxious about my statements?
And calling me stupid really makes you sound immature.
I agree with you – the iPhone’s battery performance is nothing short of embarrassing. Mine hardly lasts half a day under intense use.
My battery lasts 4 hours when idle. This is because the placement of my desk in my office causes constant switching from 3G to GPRS.
I have resorted to switching to airplane mode when at my desk.
Oh well. I hope one day we can find a more power friendly way to activate radios.
I agree airplane mode is the savior. When I’m overseas like I am right now my iPhone is on airplane mode and it lasts several days even with frequent photo & video camera use.
Thank you! And thank God someone’s talking sense around here instead of hurling immature insults at people for speaking the truth.
Also, it takes 192 cores to ONLY be 3x faster (in theory) than the dual core A7, which can run an iPhone or iPad for 10 hours of typical “real world” usage. WOW! 😉
How many GPU core’s does the A7 have? One? Two? (Too lazy to remember or look up)
But agree… Seems like 192 is a lot. But did it say “single purpose core”?? Does this mean 1 core per GPU instruction? Or just a core per screen grid section?
192 cores sounds cool… Next up the iPhone 7 with dedicated NVIDIA GPU’s… !!
It’s the 1990s all over again — the clock speed wars. Did faster processors ever make Windows “better”?
And what is the price? And how does it perform outside of graphics applications?
Kepler is all about GPU accelerated computing, not just graphics computing. So I’d venture that it’s kick ass.
The number of times that’s been claimed when reality shows something rather different. There’s more than sounding impressive there’s real life experience. I’m sure this will be pretty good but talk and bluster is easy.
And what is the difference between “graphics computing” and “GPU [graphics processing unit] accelerated computing”, pray tell?
It refers to using the GPU to assist the main CPU and use the GPU to perform standard CPU calculations rather than graphics.
Here is nvidia’s explanation:
GPU-accelerated computing is the use of a graphics processing unit (GPU) together with a CPU to accelerate scientific, engineering, and enterprise applications. Pioneered in 2007 by NVIDIA, GPUs now power energy-efficient datacenters in government labs, universities, enterprises, and small-and-medium businesses around the world.
The 64-bit Denver core based chip has been announced to be shipping “in the second half of 2014”.
Well, guess what? Apple’s A8 chip will be shipping then too.
Will this then shipping Nvidia chip be 3x the speed of the A8? No one knows for sure, but it is extremely unlikely. Apple’s A8 could even be faster!
Yes you always have to suspect the motives let alone intelligence of those who claim superiority of a future chip against one gat is already approaching its half life. Intel did the same thing with its new Atom that proved inferior to the A7 prior to its actual release in the end. Apple could easily boast about future chips but it isn’t its style.
it would be great if INTEL and Apple work together… but alas, whatever deal they have with Samsung is what it is.
Funny how everyone is bashing Apple yet uses it as a point of reference.
Bashing is a classic reaction to jealousy.
Excellent *DING* Factor!
Never underestimate the insecurity inherent in:
A whole new take on “going green”
You clever girl. 😉
It would be nice if this announcement freezes the Android market for the rest of the year, but that is probably too much to ask.
Most Android users just end up owning one somehow. There is not a lot of thought going into the purchase. There are definately a small minority of Android users who research and want the latest greatest specs. Most, however, end up with Android becuase the failed to do any research.
You’re assuming that Android users are informed and not just buying whatever the sales person tells them to buy.
A claim is just a claim.
A rose by any other name smells just sweetly but this claim smells like a troll
“We believe the Android operating system will be the most important console operating system in the future.”
Spoken as true CEO for company that does not count Apple as a customer.
Good for them for pushing the edge, we all benefit from that. But will the pedestrian – not worth charging money for – Android OS be able to take full advantage of what chip makers are producing?
Yes, who’s going to modify Android to utilize and optimize 192 cores?
And what software changes will allow for the addressing of this many cores?
First the OS has to enable it. Second, the app has to access it.
And what about backwards compatibility?
People posting reactions here seem to be ignorant to the fact that Nvidia has been making excellent processors for a very long time. Far longer than Apple and their A chips. I think Nvidia is pretty damn good at it.
Yes but has generally been out performed by the A line or haven’t you noticed. Or no doubt you would have been on here before prattling on about more than raw and potential performance that like Samsung specs never relates to the real world. So let’s see shall we when it’s actually measurable rather than board table in nature.
And by that time it will be competing with Apple’s A8 chip, not the A7.
Yes. It is infinitely more helpful to make comparisons using a common time line.
Perhaps you didn’t read the full story, they are showing it in action at CES including a demo of something than ran previously on the GTX Titan board with same performance. Then there are Epic’s comments. This is the first Kepler based mobile chip and I do know what Kepler architecture is capable of so I stand by by comment. Nvidia is damn good at it. I am sure the A8 will leapfrog the K1 but then the K2 will be announced soon after.
Whatever the case, the weakness of the K1 won’t be Nvidia, it will be android.
Apple has always led in integration, which is something that “spec sheets” can’t compete with.
That said, I hope this is another leap forward. As an Apple device user, it doesn’t concern me for several reasons.
1) It’s a game of leap frog in technology, particularly with hardware. If the A7 was in the lead — and perhaps now this — then watch what the A8 does . . . and so on.
2) Apple gets the most bang for whatever the specs are because the hardware and software are designed in conjunction.
3) I’ve always thought Apple’s CPU plan for mobile was genius because it allows for differentiation — but it doesn’t depend on it. In other words, if Apple’s own designs aren’t competing they can always go back to commodity parts from the suppliers they used to get chips from. Nothing really lost there.
Yes a good reflection of reality. Personally I will be interested to know which spec of this chip gives what performance and which appear in standard phones once launched too. Easy to make unspecific claims based on untypical versions but truth is its late to the ball so will have to be very good to be front runner for long.
But I thought 64-bit was a gimmick? I guess that’s only if Apple does it?
But…but….it’s not a gimmick because Android phones will do it properly.
Yes and its 64bit version is going to hit a barrier with Android for at least most of this year and one wonders how the 32 bit will perform when saddled with Android for that matter.
No 64-bit OS, no 64-bit apps, no multi-core enabled apps and non proliferating products. New products are unsupported the following year therefore disposable. Why is the android ecosystem a plus?
Android is a risk for everyone from the manufacturer to app developers and on down to the users.
Manucturer – Will I be be put out of business by patents or per device fees to companies like MS.
App Devs – will I make a living or go broke supporting 1000 variations of Android.
User – Will my carrier ever update this or did I just buy a paperweight on contract?
Fact is they have to talk big as they tried when faced with the A7 because talk is all that they have to offer in reality. Time will tell if this chip will live up to the hype.
We believe the Android operating system will be the most important console operating system in the future.
Have fun with that. It’s your blunder to make.
As for the chip competition with Apple:
Excellent! Bring it on. 😀
I only care what Apple does, not what their competitors do or don’t do. Apple will only adopt new technology if it can provide a great user experience at a reasonable price point.
Battery power is poorly misunderstood in iPhones and smartphones in general. If you think you can play intense, high demand, games etc, non-stop for hours, and expect a day of battery life then you are f***in stupid! Its a blimmin’ phone and if it does just that you get heaps of time even with general mail and web surfing. If you go and buy a dedicated portable gaming console with similar graphics grunt then do you get significantly more life from a similar size battery? No. Put it in context and if you are away from an adapter for that long and seriously need to Candy Crush for longer then buy a sodding back up battery. Geez!
This chip will find a huge market in the auto industry, but whether the features found in the K1 translate to revenue for an Android OS is still a ways off.
This latest chip will debut in an Audi and not a dumb phone.
If rumors are true the Apple A8 will be 64bit quad core 4-8 FPUs, 2.6GHz, Quad core GPU, SoC all on a 20nm die burning 1,100 ma in high gear GPU mode.
This will tear into Intel and potentially destroy all other ARM makers.
Samsung Austin Employees might start planning for some serious changes to their current standard of living and beautiful location!!
You guys are fucking retarded, the tegra k1 is a more like a raspberry pi than a android. It is plugged into a power outlet then out to a hd tv… the os is ubuntu too and not fucking android!!!!!!!!!!!!!!