Skip to content

Optimize TensorFlow (TF) Models For Deployment with NVIDIA TensorRT.

Notifications You must be signed in to change notification settings

afondiel/Intro-to-TensorFlow-TensorRT-Coursera

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

Introduction to TensorFlow-TensorRT (TF-TRT)

Overview

This is a project-based course on optimizing TensorFlow (TF) models for deployment using TensorRT.

  • Instructor: Snehan Kekre
  • Certificate: Awarded upon completion
  • Duration: ~<2 hours

Course Objectives

By the end of this course, you will achieve the following objectives:

  • Optimize TensorFlow models using TensorRT (TF-TRT).
  • Optimize deep learning models at FP32, FP16, and INT8 precision using TF-TRT.
  • Analyze how tuning TF-TRT parameters impacts performance and inference throughput.

Course Outline

This course is divided into three parts:

  1. Course Overview: Introductory reading material.
  2. Optimize TensorFlow Models for Deployment with TensorRT: A hands-on project.
  3. Graded Quiz: A final assignment required to successfully complete the course.

About this Project

This hands-on project guides you in optimizing TensorFlow (TF) models for inference with NVIDIA's TensorRT (TRT).

By the end of this project, you will:

  • Optimize TensorFlow models using TensorRT (TF-TRT).
  • Work with models at FP32, FP16, and INT8 precision, observing how TF-TRT parameters affect performance and inference throughput.

Prerequisites

To complete this project successfully, you should have:

  • Competency in Python programming.
  • An understanding of deep learning concepts and inference.
  • Experience building deep learning models using TensorFlow and its Keras API.

Project Structure

Task Description
Task 1 Introduction and Project Overview
Task 2 Set up TensorFlow and TensorRT Runtime
Task 3 Load Data and Pre-trained InceptionV3 Model
Task 4 Create Batched Input
Task 5 Load the TensorFlow SavedModel
Task 6 Benchmark Prediction Throughput and Accuracy
Task 7 Convert TensorFlow SavedModel to TF-TRT Float32 Graph
Task 8 Benchmark TF-TRT Float32
Task 9 Convert to TF-TRT Float16 and Benchmark
Task 10 Work with TF-TRT INT8 Models
Task 11 Convert to TF-TRT INT8

Lab: Notebook

Description Notebook Demo
Intro to TensorFlow-TensorRT Open notebook in Colab HF/Gradio Space

References

Courses

Videos

Documentation

Deep Learning Model Optimization

Additional Resources

"It's hardware that makes a machine fast. It's software that makes a fast machine slow." - Craig Bruce

About

Optimize TensorFlow (TF) Models For Deployment with NVIDIA TensorRT.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published