Welcome to the ExecuTorch Documentation (original) (raw)

ExecuTorch is PyTorch’s solution for efficient AI inference on edge devices β€” from mobile phones to embedded systems.

Key Value Propositions#


🎯 Wins & Success Stories#


Quick Navigation#

Get Started

New to ExecuTorch? Start here for installation and your first model deployment.

Quick Start

Deploy on Edge Platforms

Deploy on Android, iOS, Laptops / Desktops and embedded platforms with optimized backends.

Edge

Work with LLMs

Export, optimize, and deploy Large Language Models on edge devices.

LLMs

πŸ”§ Developer Tools

Profile, debug, and inspect your models with comprehensive tooling.

Tools


Explore Documentation#

Intro

Overview, architecture, and core concepts β€” Understand how ExecuTorch works and its benefits

Intro

Quick Start

Get started with ExecuTorch β€” Install, export your first model, and run inference

Quick Start

Edge

Android, iOS, Desktop, Embedded β€” Platform-specific deployment guides and examples

Edge

Backends

CPU, GPU, NPU/Accelerator backends β€” Hardware acceleration and backend selection

Backends

LLMs

LLM export, optimization, and deployment β€” Complete LLM workflow for edge devices

LLMs

Advanced

Quantization, memory planning, custom passes β€” Deep customization and optimization

Advanced

Tools

Developer tools, profiling, debugging β€” Comprehensive development and debugging suite

Tools

API

API Reference Usages & Examples β€” Detailed Python, C++, and Java API references

API

πŸ’¬ Support

FAQ, troubleshooting, contributing β€” Get help and contribute to the project

Support


What’s Supported#

Model Types

Platforms

Rich Acceleration