XLA compilation on GPU can greatly boost the performance of your models,

Learn how to use @tf.function(jit_compile=True) in TensorFlow to control what exact scopes are being compiled, and how to debug the performance of the resulting program.

We'll cover writing compiled models, debugging them, and exploring the performance characteristics and optimizations the XLA compiler performs, and we'll do a detailed case study on XLA usage for Google’s GPU MLPerf submission. We'll also cover how automatic kernel fusion by XLA reduces memory bandwidth requirements and improves the performance of your models. You should have basic familiarity with TensorFlow and GPU computing in general.