Skip to content

Tutorials

Explore the following tutorials for a better understanding on how to use the RBLN SDK.

RBLN Compiler

These tutorials demonstrate how to use RBLN Python API (Compile and Runtime APIs) for PyTorch and TensorFlow models.

RBLN C/C++ Runtime API

These tutorials demonstrate how to deploy precompiled models using RBLN C/C++ Runtime API.

Huggingface Model Support

These tutorials demonstrate how to compile and run inference on HuggingFace models using optimum-rbln.

Model Profiling

These tutorials demonstrate how to profile and analyze models during inference using RBLN Runtime.

Model Serving

These tutorials demonstrate how to serve precompiled AI models using Nvidia Triton Inference Server and TorchServe, both of which support vLLM.