Enable the edge ai products you dream of

Simply bringing the power of chips into the hands of innovators
Connect with us
{/*} //

Trusted and supported by

# products

Our full SOC deployment toolkit

SDK
The one-stop AI deployment SDK based on our next-gen AI compiler.
Runtime
The one-stop SoC-level inference engine across devices.
Performance Dashboard
The one-stop platform for evaluating and tracking real-world performance.

Software development kit (sdk)

runtime

performance dashboard

Benefits

Accelerated time-to-market
Full SoC enablement
Stability & tooling

for PRODUCT VENDORS

learn more

For hardware & IP vendors

learn more
# Use cases

Supported Applications

Google/
gemma-7b
cpu
gpu
npu
Meta-llama/
meta-Llama-3.2-1B
cpu
gpu
npu
OpenAI/
gpt2
cpu
gpu
npu
Deepseek-AI/
deepseek-R1-Distill-Qwen-1.5B
cpu
gpu
npu
Qwen/
qwen3-1.7B
cpu
gpu
npu
EleutherAI/
gpt-neo-2.7B
cpu
gpu
npu
FacebookAI/
xlm-roberta-large
cpu
gpu
npu
HuggingFaceTB/
smolLM2-1.7B
cpu
gpu
npu
HuggingFaceH4/
zephyr-7b-beta
cpu
gpu
npu
Qwen/
qwen3-0.6B
cpu
gpu
npu
Microsoft/
phi-2
cpu
gpu
npu
MistralAI/
mistral-7B-instruct-v0.2
cpu
gpu
npu
Meta-llama/
meta-llama-3-8B
cpu
gpu
npu
And many more ...
cpu
gpu
npu
Google/
inceptionv4
cpu
gpu
npu
Microsoft/
resnet50
cpu
gpu
npu
Hustvl/
yolos-tiny
cpu
gpu
npu
Microsoft/
resnet18
cpu
gpu
npu
Google/
mobilenetv2
cpu
gpu
npu
OpenAI/
clip-vit-base-patch32
cpu
gpu
npu
Apple/
mobilevit-xx-small
cpu
gpu
npu
Google/
vit-base-patch16-224
cpu
gpu
npu
Facebook/
detr-resnet-50
cpu
gpu
npu
PINTO/
CenterNet
cpu
gpu
npu
Yoloact-edge/
mobilenetv2
cpu
gpu
npu
Darknet/
YOLOv4
cpu
gpu
npu
Apple/
deeplabv3-mobilevit-small
cpu
gpu
npu
Nvidia/
segformer-b0-512
cpu
gpu
npu
Fast-SRGAN/
fsrgan
cpu
gpu
npu
MiDaS/
midas-v2.1-small
cpu
gpu
npu
Google/
deeplabv3
cpu
gpu
npu
Google/
mobile-deeplabv3-plus
cpu
gpu
npu
And many more ...
cpu
gpu
npu
Tensorflow/
mobilebert
cpu
gpu
npu
MLcommons_mobile/
mobile-bert
cpu
gpu
npu
Distilbert/
distilbert-base
cpu
gpu
npu
mlcommons_tiny/
keyword
cpu
gpu
npu
mlcommons_tiny/
visualwakewords
cpu
gpu
npu
Google/
bert-base-uncased
cpu
gpu
npu
Google/
mobilebert-uncased
cpu
gpu
npu
Amazon/
chronos-t5-tiny
cpu
gpu
npu
And many more ...
cpu
gpu
npu
FacebookAI/
wav2letter
cpu
gpu
npu
Microsoft/
speecht5
cpu
gpu
npu
Facebook/
mms-tts-eng
cpu
gpu
npu
Microsoft/
wavlm-base-plus
cpu
gpu
npu
OpenAI/
whisper-base
cpu
gpu
npu
usefulsensors/
moonshine-base
cpu
gpu
npu
MIT/
ast-finetuned-audioset
cpu
gpu
npu
And many more ...
cpu
gpu
npu
PointNet/
pointnet++
cpu
gpu
npu
RTCnet/
rtcnet
cpu
gpu
npu
TMVANet/
tmvanet
cpu
gpu
npu
FFTRadNet/
fftradnet
cpu
gpu
npu
And many more ...
cpu
gpu
npu

Supported hardware

ARM Cortex-M
No items found.
ARM Cortex-A
No items found.
ARM SVE
Apple Silicon
No items found.
Qualcomm Oryon
No items found.
Qualcomm Kryo
No items found.
Google Tensor
No items found.
AMD x86-64
No items found.
Intel x86-64
No items found.
RISC-V
No items found.
ARM Mali
No items found.
ARM Immortalis
No items found.
Qualcomm Adreno
No items found.
Apple Silicon GPU
No items found.
Nvidia RTX family
No items found.
ARM Ethos
No items found.
NXP Neutron
No items found.

How it works

Our next-gen AI compiler.

Roofline builds a next-generation AI compiler based on MLIR (Multi-Level Intermediate Representations).

Using layered abstraction, our compiler translates models from major AI frameworks into optimized intermediate representations and compiles them into efficient executables for diverse hardware backends.
The flexible architecture adapts easily to new hardware, especially NPUs, and enables heterogeneous execution across the system. With Roofline, the full potential of edge SoCs becomes accessible.

Join our future-focused AI team

At Roofline, we're advancing sustainable AI technology and seeking ambitious individuals for flexible on-site or remote roles.
Contribute to meaningful, disruptive innovations with us.
Open roles

FAQ

What is Roofline’s secret sauce?

Roofline’s secret sauce lies within the retargetable AI compiler. Building on MLIR and IREE, our SDK provides the first truely flexible, and thus scalable, deployment solution for edge AI.

Which AI frameworks does roofline support?

Roofline works seamlessly with PyTorch, TensorFlow Lite, TensorFlow, and ONNX. Our main focus is on PyTorch given its increasing importance for developers.

Which models does roofline support?

We test against a model zoo of 100s of models across frameworks on a daily basis, including many of the most popular architectures from Hugging Face. A selection of supported models can be found in the Supported Applications section. Reach out if you are interested in support for a specific model.

We have proprietary models, can we use them with Rooflines SDK?

Yes, we support proprietary models. They typically run out of the box, as our compiler covers the majority of common layers, operators and quantization techniques. Your models stay private at all times and can be deployed using the same simple workflow as any other model.

Do I need to make changes to my AI model to use roofline?

No, models should run without modification. We already support the majority of common layers, operators, and also quantization techniques.

Which hardware does roofline support?

Roofline supports any relevant CPU and mobile GPU. For NPUs, we are partnering with leading chip and IP vendors to integrate their hardware. A selection of hardware and links to case studies are available in the Supported Hardware section.