Welcome to Triton-Viz, a visualization and profiling toolkit designed for deep learning applications. Built with the intention of making GPU programming on Triton more intuitive.
Table of Contents
Triton-Viz is a visualization and analysis toolkit specifically designed to complement the development and optimization of applications written in OpenAI's Triton, an open-source programming language aimed at simplifying the task of coding for accelerators such as GPUs. Triton-Viz offers a suite of features to enhance the debugging, performance analysis, and understanding of Triton code.
Given that Triton allows developers to program at a higher level while still targeting low-level accelerator devices, managing and optimizing resources like memory becomes a crucial aspect of development. Triton-Viz addresses these challenges by providing real-time visualization of tensor operations and their memory usage. The best part about this tool is that while it does focus on visualizing GPU operations, users are not required to have GPU resources to run examples on their system.
- Python installed (preferably the latest available version), minimum supported version is 3.10.
- Triton installed. Follow the installation instructions in the linked repository.
- Note: the below commands must be run in order.
Upon successfully installing Triton, install Torch using the following command:
pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cu121Upon successful installation of Torch make sure to uninstall pytorch-triton using the following command:
pip uninstall pytorch-tritonClone the repository to your local machine:
git clone https://github.com/Deep-Learning-Profiling-Tools/triton-viz.git
cd triton-viz
pip install -e .You're all set!
If you want to exercise the Neuron Kernel Interface (NKI) interpreter or run the NKI-specific tests:
-
Follow the AWS Neuron Torch-NeuronX Ubuntu 22.04 setup guide to add the Neuron APT repository and install the required system packages (for example
aws-neuronx-tools,aws-neuronx-runtime-lib,aws-neuronx-collectives, and their dependencies). -
Instead of running
pip install -e .in the above section, install Triton-Viz with the optional NKI extras so the Neuron Python packages (neuronx-cc,libneuronxla,torch-neuronx) are available:pip install -e .[nki] # or pip install triton-viz[nki]
- To run core Triton-viz tests, run
pytest tests/. - (if NKI installed) To run NKI-specific tests, run
pytest tests/ -m nki. - To run all tests (Triton + NKI), run
pytest tests/ -m "".
cd examples
python <file_name>.pyWhen running with the Triton CPU interpreter (TRITON_INTERPRET=1), you can emulate concurrent SMs by setting how many blocks execute in parallel:
export TRITON_VIZ_NUM_SMS=4 # or set triton_viz.config.num_sms in PythonThis is useful for kernels that rely on cross-block synchronization (e.g., producer/consumer patterns) when testing without a GPU.
If you're interested in fun puzzles to work with in Triton, do check out: Triton Puzzles
Triton-Viz is licensed under the MIT License. See the LICENSE for details.
If you find this repo useful for your research, please cite our paper:
@inproceedings{ramesh2025tritonviz,
author={Ramesh, Tejas and Rush, Alexander and Liu, Xu and Yin, Binqian and Zhou, Keren and Jiao, Shuyin},
title={Triton-Viz: Visualizing GPU Programming in AI Courses},
booktitle = {Proceedings of the 56th ACM Technical Symposium on Computer Science Education (SIGCSE TS '25)},
numpages = {7},
location = {Pittsburgh, Pennsylvania, United States},
series = {SIGCSE TS '25}
}
