GTC 2017 in San Jose

 

GTC17

 

When: May 8 - 11, 2017

Where: San Jose, California

Booth: #520

What: NVIDIA’s GPU Technology Conference (GTC) is the place to learn, connect, and discover what’s next for the future of high performance computing. GTC 2017 will feature 500+ sessions, tutorials, and hands-on programming labs covering a diverse range of topics from Deep Learning to Web Acceleration.

Explore the exhibit hall and connect with technology experts in a one stop shop for the hottest information on GPU-enabled applications, developer tools, and hardware systems.

Discover the latest advances in GPU technology, and see how GPU technologies are creating amazing breakthroughs in important fields and how scientists, developers, engineers, and IT managers are using it to tackle their day-to-day computational and graphics challenges.

Improve your programming skills and hear about exciting innovations in the wide selection of tutorials and programming labs which are led by industry experts and NVIDIA engineers.

Learn more about this event here

 


Booth Giveaway - LEGO Star Wars Ultimate Collector Series TIE Fighter™!

 

Win Me!

This year at our booth, we have are having a giveaway for one (1) LEGO Star Wars Ultimate Collector Series TIE Fighter

 

Visit booth #520 during exhibit hours for your chance to win the iconic Imperial starfighter!
Just scan your badge, receive your number, and be back at our booth on Thursday, May 11 at 1:00pm for the live draw! 

 

   

Acceleware Tutorials at GTC

Session 1

Title: Session 1 of 4 – An Introduction to CUDA Programming

Session ID: S7699

When: Monday, May 8; 09:00 - 10:20

Where: Marriott Ballroom 3

Presenter: Chris Mason

Audience Level (all/intermediate/advanced): Beginner

Intended audience: This introductory tutorial is intended for those new to CUDA and is the foundation for our following three tutorials.  Those with no previous CUDA experience will leave with essential knowledge to start programming in CUDA.  For those with previous CUDA experience, this tutorial will refresh key concepts required for subsequent tutorials on CUDA optimization. 

Description: Join us for an informative introductory tutorial intended for those new to CUDA and is the foundation for our following three tutorials.  Those with no previous CUDA experience will leave with essential knowledge to start programming in CUDA.  For those with previous CUDA experience, this tutorial will refresh key concepts required for subsequent tutorials on CUDA optimization.

The tutorial will begin with a brief overview of CUDA and data-parallelism before focusing on the GPU programming model. We will explore the fundamentals of GPU kernels, host and device responsibilities, CUDA syntax and thread hierarchy. A programming demonstration of a simple CUDA kernel will be delivered. Printed copies of the material will be provided to all attendees for each session – collect all four!

 

Session 2

Title: Session 2 of 4 – An Introduction to the GPU Memory Model

Session ID: S7700

When: Monday, May 8; 10:30 - 11:50

Where: Marriott Ballroom 3

Presenter: Chris Mason

Audience Level (all/intermediate/advanced): Beginner

Intended audience: This tutorial is for those with a basic understanding of CUDA who want to learn about the GPU memory model and optimal storage locations. New to CUDA?  Join us for our first tutorial, An Introduction to GPU Programming – Session 1, to learn the basics of CUDA programming required for this tutorial. 

Description: This tutorial is for those with a basic understanding of CUDA who want to learn about the GPU memory model and optimal storage locations. To learn the basics of CUDA programming required for Session 2, attend Session 1 - An Introduction to GPU Programming.

This session begins with an essential overview of the GPU architecture and thread cooperation before focusing on different memory types available on the GPU. We will define shared, constant and global memory and discuss the best locations to store your application data for optimized performance. A programming demonstration of shared and constant memory will be delivered.  Printed copies of the material will be provided to all attendees for each session – collect all four!

 

Session 3

Title: Session 3 of 4 – Asynchronous Operations & Dynamic Parallelism in CUDA

Session ID: S7705

When: Monday, May 8; 13:00 - 14:20

Where: Marriott Ballroom 3

Presenter: Chris Mason

Audience Level (all/intermediate/advanced): All

Intended audience: This tutorial builds on the two previous sessions (An Introduction to GPU Programming and the Introduction to GPU Memory Model) and is intended for those with a basic understanding of CUDA programming.

Description: This tutorial builds on the two previous sessions (An Introduction to GPU Programming and the Introduction to GPU Memory Model) and is intended for those with a basic understanding of CUDA programming.

This tutorial dives deep into asynchronous operations and how to maximize throughput on both the CPU and GPU with streams. We will demonstrate how to build a CPU/GPU pipeline and how to design your algorithm to take advantage of asynchronous operations. The second part of the session will focus on dynamic parallelism. A programming demo involving asynchronous operations will be delivered. Printed copies of the material will be provided to all attendees for each session – collect all four!

 

Session 4

Title: Session 4 of 4 – Essential CUDA Optimization Techniques

Session ID: S7706

When: Monday, May 8; 14:30 - 15:50

Where: Marriott Ballroom 3

Presenter: Chris Mason

Audience Level (all/intermediate/advanced): All

Intended audience: This tutorial is for those with some background in CUDA including an understanding of the CUDA memory model and streaming multiprocessor. Our earlier tutorials (An Introduction to GPU Programming, an Introduction to the GPU Memory Model, and Asynchronous Operations & Dynamic Parallelism) provide the background information necessary for this session.

Description: This tutorial is for those with some background in CUDA including an understanding of the CUDA memory model and streaming multiprocessor. Our earlier tutorials will provide the background information necessary for this session.

This informative tutorial will provide an overview of the analysis performance tools and key optimization strategies for compute, latency and memory bound problems. The session will include techniques for ensuring peak utilization of CUDA cores by choosing the optimal block size. This session will include code examples and a programming demonstration highlighting the optimal global memory access pattern applicable to all GPU architectures. Printed copies of the material will be provided to all attendees for each session – collect all four!

 


Presenters

Acceleware training instructor Chris Mason

Chris Mason

Director of Product Management, Acceleware Ltd.
Chris is the Product Manager for Acceleware's GPU accelerated electromagnetic product line. He is responsible for the successful development and launch of Acceleware products used by companies world-wide. Chris has 11 years of experience in developing commercial applications for the GPU and has delivered over 20 CUDA courses to students in a diverse range of industries. His previous experience also includes parallelization of algorithms on digital signal processors (DSPs) for cellular phones and base stations. Chris has a Masters in Electrical Engineering from Stanford University.