Skip to content

This repos provides demo, scripts and a GUI to enable Nvidia TAO on Renesas H/W, including AI accelerate based MPUs and MCUs.

License

Notifications You must be signed in to change notification settings

renesas/renesas-nvidia-tao-integration

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Renesas-NVIDIA TAO Integration

🚀 Introduction

This repository provides demos, scripts, and an intuitive GUI to enable the NVIDIA TAO (Train, Adapt, Optimize) Toolkit on Renesas hardware, including DRP-AI-based MPUs and high-performance MCUs.

  • For beginners, the Renesas AI Model Deployer abstracts the command-line interface, enabling local execution on workstations for rapid evaluation and prototyping from training to deployment.
  • For intermediate/advanced users, a set of Jupyter Notebooks are included, offering deeper levels of customization, integration, and optimization.

NVIDIA TAO Toolkit is a low-code AI framework built on top of Pytorch/Tensorflow that enables users to train, fine-tune, and optimize state-of-the-art deep learning models for vision, speech, and language tasks.

The overall flow of Renesas' integration with the NVIDIA TAO Toolkit is illustrated below:
Renesas NVIDIA TAO Integration Overview


📁 Repository Structure

  • board_bringup/ – Instructions to bring up Renesas boards and prepare the environment for deployment.
  • docs/ – Reference documents, assets, and collateral.
  • examples/ – End-to-end demo pipelines provided by Renesas. As of the current release, the following three pipelines are supported:
Model Pipeline Support Use Case Devices Supported Reference
MobileNet v2 GUI & Notebooks Image Classification RA8D1 NGC Pretrained Classification
SegFormer-FAN hybrid Vision Transformer GUI & Notebooks Image Classification RZ/V2H or RZ/V2L NGC Pretrained SegFormer
DetectNetv2 (ResNet18 backbone) GUI & Notebooks Object Detection RZ/V2H or RZ/V2L TAO Pretrained DetectNet V2
  • gui/ – Instructions and usage guide for Renesas AI Model Deployer, along with explanations of available functionalities.
  • setup/ – Setup guide to install NVIDIA TAO Toolkit, GPU drivers, Renesas SDKs, and other required software.
  • releases - Contains the ecncapsulated package with Renesas AI Model Deployer, Jupyter notebooks and one-click installation scripts allowing customers for quick install & play.

🖥️ System Requirements

Supported Operating Systems

The Toolkit is tested and verified on Ubuntu 20.04 LTS and Ubuntu 22.04 LTS.

Hardware Requirements

Minimum Configuration

  • 8 GB system RAM
  • 4 GB GPU RAM
  • 8-core CPU
  • 1 NVIDIA GPU
  • 100 GB SSD storage

Recommended Configuration

  • 32 GB system RAM
  • 32 GB GPU RAM
  • 8-core CPU
  • 1 NVIDIA GPU
  • 100 GB SSD storage

Note:
TAO Toolkit is supported on discrete GPUs such as H100, A100, A40, A30, A2, A16, A100x, A30x, V100, T4, Titan RTX, and Quadro RTX.
It is not supported on GPUs older than the Pascal generation.

Software Requirements

  • Python (=3.8)
  • docker-ce (>19.03.5)
  • docker-API (1.40)
  • Nvidia-container-toolkit (>1.3.0-1)
  • nvidia-container-runtime (3.4.0-1)
  • nvidia-driver (>535.xx)
  • python-pip (>21.06)

Note:
All necessary system installations are handled by the provided scripts

Getting started

To get started using Renesas AI Model Deployer and the Jupyter notebooks:

  1. Please download Renesas_AI_Model_Deployer_v_x.y.z.tar under assets in releases.

  2. Get your NVIDIA NGC API Key to access the NVIDIA TAO Toolkit:

    • Go to the NGC sign-in page and log in.
    • Click your username in the top-left corner.
    • Select SetupGenerate API Key.
    • Choose both services (NGC Catalog & Helm Chart Registry).
    • Click Generate Key, then copy and store it in a safe location to be inputted during next step.
  3. Run the following shell scripts within the untarred directory, its recommended to run them one by one:

    chmod ug+x *.sh
    chmod ug+x bin/*.sh
    ./docker_gpu_install.sh
    ./setup_tao_env.sh

    This should start the setup script and install the necessary dependancies to use the GUI and the Jupyter notebooks. Ensure to select TAO, TOOLs, Easy_GUI and Pre_image.
    For Reneasas RZ/V, please install and setup AI SDK that includes DRP-AI TVM v2.5, steps 3-5 from here.

    groups

    If '''docker''' is not in the reported groups then a reboot is required to finish the installation.

  4. To start Renesas AI Model Deployer:

    ./gui_start.sh 
  5. For the Jupyter Notebooks, execute in untarred project directory:

    ./jupyter-notebook

For further explanation and error handling, please review the setup/ directory.

Supported Hardware and Kits

Renesas AI Model Deployer currently supports the following hardware based on respective pipeline:

Device Evaluation Kit Use Case
RZ/V2L RZ/V2L-EVKIT SegFormer-FAN hybrid Vision Transformer, DetectNetv2
RZ/V2H RZ/V2H-EVKIT SegFormer-FAN hybrid Vision Transformer, DetectNetv2
RA8D1 EK-RA8D1 MobileNetV2

Known Issues

  • In the DetectNet_v2 demo, bounding boxes are drawn incorrectly and only for one class due to a post-processing issue during deployment. This will be fixed in the next release.

  • In the Segformer demo, the output is always "dog" due to a post-processing bug during deployment. This issue will be addressed in the upcoming release.

  • For the DetectNet_v2 demo on RZ/V2L, inference speed is currently ~1–3 seconds as the first convolution layer runs on the CPU. A patch in the next release will offload this to DRP-AI, reducing inference time to ~220 ms.

About

This repos provides demo, scripts and a GUI to enable Nvidia TAO on Renesas H/W, including AI accelerate based MPUs and MCUs.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors 3

  •  
  •  
  •