The University of Texas at Austin: LAFF-On Programming for High Performance
Learn to squeeze high performance out of modern CPUs.
LAFF-On Programming for High Performance
About this courseSkip About this course
Is my code fast? Can it be faster? Scientific computing, machine learning, and data science are about solving problems that are compute intensive. Choosing the right algorithm, extracting parallelism at various levels, and amortizing the cost of data movement are vital to achieving scalable speedup and high performance.
In this course, the simple but important example of matrix-matrix multiplication is used to illustrate fundamental techniques for attaining high-performance on modern CPUs. A carefully designed and scaffolded sequence of exercises leads the learner from a naive implementation to one that effectively utilizes instruction level parallelism and culminates in a high-performance multithreaded implementation. Along the way, it is discovered that careful attention to data movement is key to efficient computing.
Prerequisites for this course are a basic understanding of matrix computations (roughly equivalent toWeeks 1-5 of Linear Algebra: Foundations to Frontiers on edX) and an exposure to programming. Hands-on exercises start with skeletal code in the C programming language that is progressively modified, so that extensive experience with C is not required. Access to a relatively recent x86 processor such as Intel Haswell or AMD Ryzen (or newer) running Linux is required.
MATLAB Online licenses will be made available to the participants free of charge for the duration of the course.
Join us to satisfy your need for speed!
At a glance
- Subject: Computer Science
- Level: Intermediate
Exposure to programming and Linux. Basic understanding of matrix-matrix multiplication.
- Language: English
- Video Transcript: English
- Associated skills: Extract Transform Load (ETL), Machine Learning, C (Programming Language), Matrix Multiplication, Linear Algebra, Scalability, X86 Architecture, Linux, Scientific Computing, Algorithms, Amortization, Data Science
What you'll learnSkip What you'll learn
Mapping algorithms to architectures
Extracting parallelism at multiple levels
Amortizing data movement over computation
Understanding performance data
Managing complexity through layering of software
0 Getting Started
1 Loops and More Loops
2 Start Your Engines
3 Pushing the Limits
4 Multithreaded Parallelism
Learner testimonialsSkip Learner testimonials
"We will include this course as one of the main onboarding materials for our team''
- Dr. Misha Smelyanskiy, Technical Lead and Manager of AI System Co-design Group at Facebook