The NVIDIA Blog |
- NVIDIA Propels Deep Learning with TITAN X, New DIGITS Training System and DevBox
- Ready for Battle: Battlefield Hardline’s Release Highlights Our Game Readiness
- Live: Jen-Hsun Huang Kicks Off NVIDIA’s 2015 GPU Technology Conference
- NVIDIA’s Next-Gen Pascal GPU Architecture to Provide 10X Speedup for Deep Learning Apps
- How NVIDIA DRIVE PX Will Help Automakers Slim Down Self-Driving Cars
|
Posted: 17 Mar 2015 10:40 AM PDT
Addressing one of today's most sophisticated technology challenges, we just unveiled new hardware and software that bring unprecedented speed, ease and power to deep learning research.
Deep learning – a rapidly growing segment of artificial intelligence – is an engine of computing innovation for areas as diverse as advanced medical and pharmaceutical research to fully autonomous, self-driving cars. NVIDIA CEO and co-founder Jen-Hsun Huang showcased three new technologies that will fuel deep learning during his opening keynote address to the 4,000 attendees of the GPU Technology Conference:
Another Side of TITAN XTITAN X is our new flagship GeForce gaming GPU, but it's also uniquely suited for deep learning.We gave a sneak peek of TITAN X two weeks ago at the Game Developers Conference, in San Francisco, where it drove a stunning virtual reality experience called "Thief in the Shadows," based on the dragon Smaug, from "The Hobbit." The latest AAA titles are breathtaking on TITAN X in 4K. Middle-earth: Shadow of Mordor, for example, runs at 40 frames per second on high settings with FXAA enabled, compared with 30fps on the GeForce GTX 980, released in September. Built on the NVIDIA Maxwell GPU architecture, TITAN X delivers twice the performance and double the power efficiency of its predecessor by combining 3,072 processing cores for 7 teraflops of peak single-precision performance with 12GB of onboard memory. With that processing power and 336.5GB/s of memory bandwidth, it can rip through the millions of pieces of data used to train deep neural networks. On AlexNet, an industry-standard model, for example, TITAN X took less than three days to train the model using the 1.2 million image ImageNet dataset, compared with over 40 days for a 16-core CPU. Available today, TITAN X is selling for just $999. DIGITS: Quick, Easy Path to the Best Deep Neural NetUsing deep neural networks to train computers to teach themselves how to classify and recognize objects can be an onerous, time-consuming task.The DIGITS Deep Learning GPU Training System software changes that by giving users what they need from start to finish to build the best possible deep neural nets. Available for download at http://developer.nvidia.com/digits, it's the first all-in-one graphical system for designing, training and validating deep neural networks for image classification. DIGITS guides users through the process of setting up, configuring and training deep neural networks – handling the heavy lifting so that scientists can focus on the research and results. Preparing and loading training data sets with DIGITS – whether on a local system or from the web – is simple, thanks to its intuitive user interface and workflow management capabilities. It's the first system of its kind to provide real-time monitoring and visualization, so users can fine tune their work. And it supports the GPU-accelerated version of Caffe, the popular framework used by many data scientists and researchers today to build neural nets. DIGITS DevBox: World's Fastest Deskside Deep Learning MachineBuilt by the NVIDIA deep learning engineering team for its own R&D work, the DIGITS DevBox is an all-in-one powerhouse of a platform for speeding up deep learning research.Starting with its four TITAN X GPUs, every component of the DevBox – from memory to I/O to power – has been optimized to deliver highly efficient performance for the toughest deep learning research. It comes preinstalled with all the software data scientists and researchers require to develop their own deep neural networks. This includes the DIGITS software package, the most popular deep learning frameworks – Caffe, Theano and Torch – and cuDNN 2.0, NVIDIA's robust GPU-accelerated deep learning library. And it's all wrapped up in an energy-efficient, quiet, cool-running — and great-looking — package that fits under a desk and plugs into an ordinary wall socket. Very early results of multi-GPU training show the DIGITS DevBox delivers almost 4X higher performance than a single TITAN X on key deep learning benchmarks. Training AlexNet can be completed in only 13 hours with the DIGITS DevBox, compared to over 2 days with the best single GPU PC, or over a month with a CPU-only system. Learn more about the NVIDIA DIGITS DevBox. The post NVIDIA Propels Deep Learning with TITAN X, New DIGITS Training System and DevBox appeared first on The Official NVIDIA Blog. |
|
Posted: 16 Mar 2015 05:25 PM PDT
The release of Battlefield Hardline didn't sneak up on me. I've been looking forward to it for months. My anticipation started long before I began playing the open Beta.
So, ahead of the game's launch, I'm working through my pregame checklist:
NVIDIA invests more than anyone else in delivering a superior gaming experience, from install to play, through a combination of engineering efforts and support for developers. One of those engineering efforts is called out in the second step of my routine: Game Ready drivers. Your graphics driver is the software that links the operating system to a particular GPU. It controls how your graphic components work with the rest of your computer. Your GPU is only as good as your drivers. Nothing's worse than launching a new game, only to find out it's borked and you have to wait on a driver to fix it. That's why we work hard to ensure that our Game Ready drivers are available the day a game launches (or earlier) to provide the best possible gaming experience. These Game Ready drivers make sure that your GeForce GTX GPU delivers the best possible experience with each game. This means optimal performance, bug-free game play, and smooth frame rates. They sometimes are an instant upgrade, delivering double-digit performance increases. They also deliver essentials like SLI and Optimus profiles or new features like Battery Boost or Dynamic Support Resolution (DSR). And since Game Ready drivers are nearly always WHQL-certified at launch by Microsoft, you know you're getting the best quality. In the last year, we've had a steady cadence of Game Ready drivers, all timed with important game launches. If you were excited about Thief, TitanFall, Watch Dogs, Metro: Redux, Middle-earth:Shadow of Mordor, F1 2014, Aliens: Isolation, Borderlands: The Pre-Sequel!, Civilization:Beyond Earth, Lords of the Fallen, Assassin's Creed: Unity, World of Warcraft: Warlords of Dranear, Call of Duty: Advanced Warfare, Far Cry 4, Dragon Age Inquisition, The Crew, Dying Light, Evolve or Total War: Attalla, we had you covered with a Game Ready driver on or before launch day. Our goal is to set the benchmark in the industry with our driver program. And our hard work is paying off. In a recent review, hardware website Vortez.com said that our driver program sets the standard when it comes to being ready for every relevant game launch and offers excellent post-launch support. NVIDIA works hard to deliver a superior experience through customer-focused initiatives such as day-of-launch driver support and the way GeForce Experience simplifies PC optimization and maintenance. If you have a GeForce GTX GPU, you can get comfortable in your pregame routine. So, get ready. Download our Game Ready driver for Battlefield Hardline and enjoy the fun. The post Ready for Battle: Battlefield Hardline's Release Highlights Our Game Readiness appeared first on The Official NVIDIA Blog. |
|
Posted: 16 Mar 2015 01:00 PM PDT
11:12 – JHH thanks Elon and talks about how much work there is to do, and how committed NVIDIA is to do that. Elon strolls off stage as JHH calls him "the engineer's engineer." Things are winding down now.
There's a recap of the day's news:
11:10 – JHH asks about protecting cars from security attacks, from hacks, like computer networks have been hacked. Elon notes that there's great importance for a car to still have a steering wheel and brake, with additional levels of security. Thus, while an infotainment message might be able to be hacked, there'd be much tighter security that would prevent a car's engine or self-steering system from being hacked. Tesla is the leader in electric cars and we'll be the leader in autonomous cars, Elon says. We're going to put a lot of car into autonomous driving and we're going to save a lot of lives. When it comes to AI, I'm not worried about autonomous cars or autonomous air conditioning units, it's the deep intelligence, Elon says. Live streaming video by Ustream 11:06 – "We’ll take autonomous cars for granted in a very short period of time," Elon says. JHH asks about government policies, saying he'd love to work on email while driving to work, quickly adding that he'd like to do so without breaking the law. What does the government need to do? Elon notes that it will be several years after self-driving becomes possible that government will allow it. They want a car to be not as safe as a person driving but significantly safer. "When it comes to public safety, there’s an argument for being quite cautious before there's a change. I don't think it's the case right now that there's a full autonomous system that regulators aren’t approving. But there could be next year." "I don't think it's the case right now that there's a full autonomous system that regulators aren’t approving,” Tesla Motors CEO Elon Musk. “But there could be next year."Elon adds that the first thing we try to do is establish a hardware platform so we can do continuous updates to the software. A lot of that will happen later this year. I have an announcement on Thursday, and I don't want to get ahead of that.He deflects Jen-Hsun’s kidding effort to ask if he want to share that news, "There's going to be a call on Thursday on what's going to be in our next version for anyone that's interested, that's all" 11:01 – the conversation between NVDIA CEO Jen-Hsun Huang and Tesla Motors CEO Elon Musk continues… JHH: what's your roadmap about how to get to self-driving cars? Elon: First, you can keep uploading new software, so the car you have will steadily improve with new capabilities. The car will get smarter and smarter with the current hardware suite. Even with just what we have, we'll make huge progress in autonomy. We can make car steer itself on freeway, do lane changes. Autonomy is about what level of reliability and safety we want. But with the current hardware suite, it can't drive safely at 30MPH, with children playing. To solve that, we need a bigger sensor suite and more computing power. What NVIDIA's is doing with Tegra in really interesting and really important for self-driving in the future. Asked by JHH, Elon begins discussing technological hurdles. It gets tricky around 30-40MPH open driving environment. At 5-10MPH it's relatively easy because you can stop within the range of ultra-sonics. And then from 10-50MPH in a complex environment, that's where you get a lot of unexpected things happening. Over 50MPH in freeway environment, it gets easier because possibilities get narrower. So, it's the midrange that's challenging. "But we know exactly what to do and we'll get there in a few years." 10:56 – We can talk about cars forever, JHH says. But JHH introduces Elon Musk. He says he's bought all three of Tesla's cars – the original sports car and two versions of the Model S Sedan. Elon comes on stage, dressed in black, identical to JHH, though lacking the black leather jacket. JHH: We made it a point not to rehearse anything. So, let's get directly to the juicy stuff. You were quoted as saying artificial intelligence is more dangerous than nuclear weapons, like a demon. How do you consolidate and rationalize that with deep learning’s potential? Elon: We don't have to worry about autonomous cars. Doing self-driving is easier than people think. There used to be elevator operators, but now we've developed circuitry so they go where you want to go. In the distant future, they may outlaw driven cars because they're too dangerous. “In the distant future, they may outlaw driven cars because they're too dangerous,” Tesla Motors CEO Elon MuskJHH: If we have the right intelligence, we don't have to make cars that heavy because they don't need to survive the same collisions, can we relax some of those laws?Elon: If you can count on not having an accident , you can get rid of a lot things, though we're still some time away from that. Capacity of cars and trucks is 100M a year and there are 2B on the roads now, so it could take 20 years for the whole base to be transformed. Similarly, it would take 20 years to replace the world's fleet of cars with electric. 10: 50 – JHH shows a DARPA self-driving car initiative called Dave, which was trained on 225K images. Dave was trained using deep learning on these images to calculate what behavior to do. He sums it up nicely: "The input was images, the output was drive commands. It's a deep neural network in action." DRIVE PX is based on two NVIDIA Tegra X1 processors, and takes input from 12 cameras. Instead of Dave's 3.1 million brain-like connections, AlexNet on DRIVE PX has 630 million connections. While Dave processes at 12 frames a second, DRIVE PX processes AlexNet at 184 million frames a second. Dave's connections was firing off 38 million times a second, DRIVE PX can fire off its neural capacity at a rate of 116 billion times a second. To augment today's ADAS systems with deep learning has enormous potential. This can be done with DRIVE PX, which will be available in May 2015 as a developer kit, for $10,000 – to qualified buyers. 10:44 – JHH now talks about NVIDA's DRIVE PX Self-Driving Car Computer, which will bring deep learning to ADAS. It's going to be very difficult to code if-then scenarios of an infinite number of possibilities to lead to a self-driving cars. How do you teach a baby to play ping-pong? With conventional analysis, you'd have to teach a baby when a ball leaves a hand, when the paddle should strike the ball, the impact of the sponge on the paddle on the ball. You'd have to teach the baby Newtonian physics. It requires simulating trajectory, among other things. JHH shows a wonderful video that shows a dad teaching a one-year-old to play pingpong. The baby's sitting on the ping pong table and hits back the ball with kind of amazing consistency. He does from time to time get distracted. "It turns out, that teaching a car to drive itself isn't that different from teaching a baby to play ping pong, " NVIDIA CEO Jen-Hsun Huang.10:38 – JHH now shifts gears, so to speak, to the subject of self-driving cars.ADAS – advanced driver assistance systems – is used for endeavors like adaptive cruise control, self-braking in case of emergencies. He asksi, "But stopping a car when you get to an object is far off from self-driving cars. How do we get from here to self-driving cars?" Our vision is to augment today's ADAS systems with deep learning. Stopping when an object is ahead is available today. We want to augment this and learn behavior over time to get smarter and smarter. The big bang of self-driving cars is about to come. The availability of massively parallel supercomputers, the availability of training from cameras like GoPro, we'll be able to train cars with smarter behavior. This is going to require more than if-then training. We can't program every possibility. We'll never get there. Deep learning is what will get us there. 10:34 - To get a closer description of this 10X speedup, you might want to check out our latest blog post, which goes into some detail of how this is derived. There will be a deeper drill down later, as well, available on our Parallel for All technical blog. 10:31 – We're shifting gears again and going to the promised Roadmap Reveal. JHH promises that this will address the concern of how much faster can we go. He cites Pascal, the next generation of GPUs following the current gen, Maxwell, which launched just a year ago. Pascal was initially mentioned at last year's GTC but without much details. Now it's being provided. Pascal has three great features. One is mixed precision, which is provides at 3X the level of Maxwell. Another is 3D memory, which provides more bandwidth and capacity simultaneously. A third is NVLink, which is an ability for multiple GPUs to be connected at very high speeds. It also has 2.7X more capacity than Maxwell. As a result, Pascal will be 10X faster than Maxwell, although JHH mentions that this is just a very rough, high level estimate. Live streaming video by Ustream 10:24 – JHH rolls out a software offering by NVIDIA called DIGITS, which enables deep learning to be carried out. It can process data, configure a deep neural network, monitor progress along the way and visualize layers. To package this stuff up, he says that NVIDIA has built a single appliance called the DIGITS DevBox, which plugs into the wall. It's the maximum GPU performance from a wall socket. It's packed with TITAN X GPUs, boots up Linux and software that runs deep learning. Early results have been fantastic, he notes. There are rave testimonials from execs at Flickr and Facebook. DIGITS DevBox isn't intended to be sold wildly – it's for researchers, not the mass market. "Researchers can now very quickly get fully up to speed, so you can get back to work and not worry about being a plumber or electrician," JHH says. Starting price is $15,000, available in May 2015. "Researchers can now very quickly get fully up to speed, so you can get back to work and not worry about being a plumber or electrician,” NVIDIA CEO Jen-Hsun Huang on the DIGITS DevBox"It's priced at a level all researchers can afford."10:19 – JHH now talks about some amazing work that's being done in what's called "Automated Image Captioning with ConvNets and Recurrent Nets" by researchers at Stanford University. He introduces Andrej Karpathy, a Stanford researcher behind this work, which is involved in the network "seeing" images and describing them in a simple sentence. The science behind this is unfathomably complicated, but the end result is easy to understand. He flashes pictures on a screen and shows how the network identifies them based on 500,000 images. He shows an image of bird on a branch of a tree, which the network identifies as well, "a bird perched on a branch of a tree." This may lose something in the writing.
10:11 - Mike shows how a computer network is trained by pumping thousands, even millions of images through it. The network gets smarter and smarter until it's able to identify pictures. After some pretty deep technical description, Mike finishes up. "This is a sense of what a deep neural network is about," JHH says. There's lots of mathematics with iteration upon reiteration. Anybody who's a data scientist, anyone who's trying to predict the future, based on non-computationally-extracted patterns – anyone doing this is jumping into deep learning. The number of companies doing this has exploded. The list of companies doing this is massive. There's Facebook, IBM, Baidu and Google, among the biggest. And there is a sea of start ups engaging in this. As a result, applications will seem smarter than ever. Deep leraning is also sweeping through science, he notes. It's involved in detecting mitosis in breast cancer cells, being done by Switzerland's IDSIA. It's used to predict the toxicity of new drugs by Johannes Kepler University in Austria, and understanding gene mutation to prevent disease by researchers at the University of Toronto. "This is all groundbreaking work, and it's just the beginning," JHH says. "This is all groundbreaking work, and it's just the beginning," NVIDIA CEO Jen-Hsun Huang.10:02 – JHH now shows the way a deep neural network works, depicting a multitude of filters that work the way the brain does. Each layer, each filter identifies many different aspects of an image, which ultimately gets recognized by the computer.There's a bit of a kerfuffle with the slides that he's showing, and it turns out he's a slide ahead of himself. He asks the tech team to back up a slide and then falls back into stride. "One of the best things about GTC keynotes is we don't rehearse very much," he says, laughing. JHH now introduces a colleague, Mike Houston, a PhD researcher at NVIDIA, who gives a description of how a finely tuned version of AlexNet works. Take a picture of a cute little bunny, which gets read by 96 different filters. Dominant features get identified and then put through weight networks. There are some pretty stunning graphics that make it feel like you're flying through computer network. Take a picture of a koala or a housecat Mike says. You can put it through the network and see how they're identified not as a squirrel, not as a stuffed animal, but as what they are. Live streaming video by Ustream 9:54 – In a more recent image-recognition challenge, called ImageNet, the error rate based on using NVIDIA GPUs in a deep network has fallen from just 5.98 percent by Baidu, to 4.94 percent by Microsoft, to 4.82 percent by Google. Any of these is a lower error rate than a human could produce. There are three reasons for this. Call it the big bang of deep learning, which occurred in 2012. First, is the rise of algorithms in deep neural networks, developed by some of the finest minds in computing. Second, is the rise of huge data sets, which has been possible only with the rise of the Internet, which produces sets with millions of constituents. Third, is the rise of GPU-powered supercomputers, which are able to exert the processing grunt. 9:49 – Now, JHH is on to his second announcement that he promises will be about "a very fast box." But first, he provides background about the history of deep learning. Although its origins go back 50-60 years, deep learning started in earnest in 1995 at Bell Labs by Yann Lecun. He did the initial work in the field, recognizing edges using hand-engineered codes built up over those previous decades. Work done in 1998 developed a ground-breaking method for hand-writing and number recognition, which has replaced humans at banks and other organizations. More recently, the basis for deep learning moved from computer vision to deep neural networks, and accuracy began improving markedly, thanks to work by Jeff Hinton and Alex Krizhevsky and others in developing AlexNet. In international competition, AlexNet derived accuracy that rose from a previous record of 74 percent accuracy in 2011 to 84 percent in 2012. Now, it is pushing towards 90 percent and beyond. 9:40 – But, JHH notes, TITAN X is about more than graphics. It's also about deep learning. TITAN X can train AlexNet in under three days, compared with 43 days for a 16-core Xeon CPU, and six days on NVIDIA's original TITAN GPU from a year and a half ago. JHH, says, "When you speed something up from a month and a half and you reduce it to a week, it's the difference between your willingness to do the work and maybe not at all. Down to three days is utterly life changing." JHH announces its prize: $999. "It will pay for itself in just one afternoon." “It will pay for itself in just one afternoon,” NVIDIA CEO Jen-Hsun Huang on our new GeForce TITAN X9:37 - Man, it's an incredibly beautiful animation that looks close to being filmed, but with far more detail than a filmed version could capture. A young boy’s runaway red kite soars through a mountainous valley to the edge of a cave. The boy recovers the kite at the cave's edge and then enters it. In an opening, he lets the kite go and it joins dozens of others, swirling in the sky. The tagline comes up, "If you love something, set it free." Hearty applause comes up. "That story was impossible to tell without the advancement of computer graphics. When you see it, it takes your breath away," JHH says. 9:34 – He talks about the importance of acceleration. "Most of the applications we serve are about speed. Without speed, you can't do your work. We need to find a delicate balance between programmability and speed." He says that by putting CUDA in every single GPU, we make it as easy as possible for you to develop and apply your software. I'd like to show you today how we're living up to this promise. So, he introduces NVIDIA's latest GPU, TITAN X with a very elegant, highly animated video that's part Blade Runner, part film noir. TITAN X was first mentioned two weeks ago at the Game Developers Conference, with few details. Now, he's specific. It has 8 billion transistors, 3,072 CUDA cores, 7 teraflops of single-precision throughput, and a 12GB framebuffer. It's based on Maxwell. He shows what it can do by showing an animation that's a real-time animation called Red Kite by Epic Games. It's running on TITAN X. It captures 100 square miles of 3D graphics, depicting a yawning valley with 13 million plants. Any given frame is 20-30 million polygons. It shows physically based rendering – rocks look like rocks, shadows like shadows. "This is the most beautiful rendering I've ever seen." 9:27 – Now, JHH talks about breakthrough with GPU Computing, thanks to the kind of people who attend GTC. GTC's first year was 2009. The prior year, there were 150,000 downloads of CUDA, NVIDIA's language that allows GPUs to crunch data. There were 27 CUDA apps, 60 universities teaching CUDA and 4,000 academic papers on CUDA, and 6,000 Tesla GPUS. BY contrast, last year, there 3 million CUDA downloads,319 CUDA apps, 800 universities teaching CUDA, 60,000 papers have cited GPUs and CUDA, and 450,000 Tesla GPUs are in use. It's a pretty impressive chart that he shows. "It's amazing progress in just a few years, and I want to thank all of you for that." 9:23 – JHH recaps the past year in visual computing – highlighted by breakthroughs in gaming, including game-streaming to the television. Think of it as a Netflix for gaming. He also cites breakthroughs in cars, which he says are increasingly software and hardware on wheels. The type of experiences we'll see in near future are inspiring. Nobs, buttons and dials are being replaced with touch, nodding and gestures. Cars will be just one delightful experience rolling down the street. He also describes breakthroughs in enterprise – with GRID-accelerated VDI, virtual desktop infrastructure, that allows graphics-intensive applications to stream to your connected screen, no matter where you are. NVIDIA technology is being used to make the new “Star Wars,” movie. "I'm out-of-my-body excited about this. I'll create any technology I can to allow them to make “Star Wars” faster," he says to laughter. "I'm out-of-my-body excited about this. I'll create any technology I can to allow them to make “Star Wars” faster," – NVIDIA CEO Jen-Hsun HuangAnd he describes breakthroughs in deep learning – how for the first time recently a computer was able to beat humans at recognizing images.It's been an amazing year in visual computing 9:17 – There's a rousing cheer. JHH comes out in his trademark black — leather jacket, black pants and shirt. He says he's going to cover four things: a new GPU and deep learning; a very fast box and deep learning; a roadmap reveal and deep learning; and self driving cars and deep learning. Is there a pattern here? Deep learning is arguably as exciting athe invention of the internet, he says. The potential of this technology and this work, and its implications for all industries and science are amazing, he says. The possibilities are endless. He says he will dedicate his entire talk to deep learning, as much of this show will be dedicated, as much of the work of the next decade will be. 9:12 - The room grows black, and the video comes up. NVIDIA puts massive time – and resources – into creating so-called anthem videos for its signature shows, put together by a richly talented in-house creative studio. This one's built on the theme "Do you remember the Future," a reference to the thousands of researchers in the audience doing work in fields ranging from nano-particles to astronomy, cancer detection to materials-science. Against a bouncy, slightly nostalgic-tinged soundtrack, it presents a flowing collage of videos with a voice over by the mellifluous Peter Coyote. The vids depict the way the future has been seen in the past – the way adults today remember it being seen in our youth – compared with how technology is being applied today. There's the 1960s smile-inducing Jetsons, of course, in flying cars and grainy footage, and the flying DeLorean from the 1980s-era Back to the Future – contrasted with self-driving capabilities being developed by Audi and Tesla. There's a scene from 1966's Fantastic Voyage of a crew navigating through the human body – contrasted with an animation of a video-equipped pill being developed by the University of Barcelona that sends images as it wends its way through the digestive tract. There's a scene of the Star Trek tricorder medical scanning device – contrasted against work being done today with machine learning-based efforts to detect malignancies in cancer diagnostics. The narrator intones at the end, "Do you remember the future, look around you. It's here." It's enough to make an aging Trekkie well-up a bit. 9:11 – A hopeful sign. The so-called Voice of God comes up and asks people to be seated. The keynote, she says, will get started momentarily. The music is starting to shift gears, another promising sign. 8:55 – Folks are still streaming in, packing this airplane-hanger sized space. Well, it's not quite like an elevator car in the morning rush, but there aren't empty seats. Jen-Hsun – we'll refer to him as JHH in this live blog – usually comes on within a minute or two of the official start. So, it shouldn't be long. 8:50 – We're back in the San Jose Convention Center for the sixth annual GPU Technology Conference. It's a big one – 4,000 attendees from 50 countries. There will be 500 sessions this week on how GPUs are pushing research forward on a jillion fronts. And there will be three high-profile keynotes. But the highlight of GTC each year is invariably Jen-Hsun Huang's keynote. He's a great showman with exacting standards and a taste for wildly ambitious demos. Plus, he's being joined today by Elon Musk, the man behind Tesla Motors and SpaceX. The room is largely black, with a few poison-green illuminated GPU Technogy Conference signs. The stage itself stretches across the back wall of the massive room. It's 120×20 feet, the size of a third of a tennis court. Against it is a moving graphic of what appears to be brain neurons with synapses moving through them, like slow-mo lighting. The post Live: Jen-Hsun Huang Kicks Off NVIDIA's 2015 GPU Technology Conference appeared first on The Official NVIDIA Blog. |
|
Posted: 16 Mar 2015 10:44 AM PDT
NVIDIA's Pascal GPU architecture, set to debut next year, will accelerate deep learning applications 10X beyond the speed of its current-generation Maxwell processors.
NVIDIA CEO and co-founder Jen-Hsun Huang revealed details of Pascal and the company's updated processor roadmap in front of a crowd of 4,000 during his keynote address at the GPU Technology Conference, in Silicon Valley. "It will benefit from a billion dollars worth of refinement because of R&D done over the last three years," he told the audience. The rise of deep learning – the process by which computers use neural networks to teach themselves – led NVIDIA to evolve the design of Pascal, which was originally announced at last year's GTC. Pascal GPUs will have three key design features that will result in dramatically faster, more accurate training of richer deep neural networks – the human cortex-like data structures that serve as the foundation of deep learning research. Along with up to 32GB of memory — 2.7X more than the newly launched NVIDIA flagship, the GeForce GTX TITAN X — Pascal will feature mixed-precision computing. It will have 3D memory, resulting in up to 5X improvement in deep learning applications. And it will feature NVLink – NVIDIA's high-speed interconnect, which links together two or more GPUs — that will lead to a total 10X improvement in deep learning. Mixed-Precision Computing – for Greater Accuracy Mixed-precision computing enables Pascal architecture-based GPUs to compute at 16-bit floating point accuracy at twice the rate of 32-bit floating point accuracy. Increased floating point performance particularly benefits classification and convolution – two key activities in deep learning – while achieving needed accuracy. 3D Memory – for Faster Communication Speed and Power Efficiency Memory bandwidth constraints limit the speed at which data can be delivered to the GPU. The introduction of 3D memory will provide 3X the bandwidth and nearly 3X the frame buffer capacity of Maxwell. This will let developers build even larger neural networks and accelerate the bandwidth-intensive portions of deep learning training. Pascal will have its memory chips stacked on top of each other, and placed adjacent to the GPU, rather than further down the processor boards. This reduces from inches to millimeters the distance that bits need to travel as they traverse from memory to GPU and back. The result is dramatically accelerated communication and improved power efficiency. NVLink – for Faster Data Movement The addition of NVLink to Pascal will let data move between GPUs and CPUs five to 12 times faster than they can with today's current standard, PCI-Express. This is greatly benefits applications, such as deep learning, that have high inter-GPU communication needs. NVLink allows for double the number of GPUs in a system to work together in deep learning computations. In addition, CPUs and GPUs can connect in new ways to enable more flexibility and energy efficiency in server design compared to PCI-E. The post NVIDIA's Next-Gen Pascal GPU Architecture to Provide 10X Speedup for Deep Learning Apps appeared first on The Official NVIDIA Blog. |
|
Posted: 16 Mar 2015 10:40 AM PDT
The future, as the saying goes, is already here. It's just unevenly distributed. Maybe that's because it comes loaded with a trunk full of computers.
So today we're announcing pricing and availability for our NVIDIA PX development platform. Introduced at the Consumer Electronics Show in January, NVIDIA DRIVE PX is a powerful self-driving car computer designed to slip the power of deep neural networks into real-world cars. We're making our DRIVE PX development platform available in May to automakers, Tier 1 automotive suppliers and research institutions, to get cars on the road toward driving themselves. NVIDIA DRIVE PX is built for the growing number of automakers that have already put – or soon will put – self-driving cars on the roads. The common denominator: all of these projects rely on NVIDIA GPU technology to help process and analyze, in real time, the data streaming in from sensors and cameras mounted all over the car. One of our partners has even announced plans to send off its self-piloted cars across the United States. DRIVE PX's twin NVIDIA Tegra X1 processors deliver 2.3 teraflops of performance. Yet each individual superchip is no bigger than a thumbnail. That's enough to weave together data streaming in from 12 camera inputs and enable a wide range of advanced driver assistance features to run simultaneously – including surround view, collision avoidance, pedestrian detection, mirror-less operation, cross-traffic monitoring and driver-state monitoring. More Is More Yet DRIVE PX is built to tap into a new technology called "deep learning" to give it capabilities far beyond what you can stuff into any of today's passenger cars. That's because today's advanced driver assistance systems have evolved around the principle of classifying objects that the car’s sensors would detect. It works, but it's not enough. Imagine training such a system to be ready for any possible eventuality. It's just not possible. Our DRIVE PX development platform is built to crack that problem. It includes a new deep neural network software development kit we call DIGITS, as well as video capture and video processing libraries. DIGITS is a deep learning training system that can be run on systems powered by our GPUs – including our new DIGITS DevBox development platform – and that lets computers train themselves to understand objects in the world around them. Much like a human learns through experiences, so does the deep neural network. Now we can do more than just train systems to recognize objects – but also training behavior. (See "Here's How Deep Learning Will Accelerate Self-Driving Cars" for an overview of how this works.) Then the model created by the DIGITS DevBox can be loaded into a vehicle and run in real time on the DRIVE PX. It's a system that can be trained, and retrained, with more data. Every time your self-driving car gets an over-the-air update, it can get smarter. The result: a self-driving system that extends well beyond the hardware found in any car, without all that extra electronic junk in the trunk. The post How NVIDIA DRIVE PX Will Help Automakers Slim Down Self-Driving Cars appeared first on The Official NVIDIA Blog. |