Previous Releases Notes (Neuron 2.x) — AWS Neuron Documentation (original) (raw)

Previous Releases Notes (Neuron 2.x)#

Table of contents

Neuron 2.22.1 (05/12/2025)#

Neuron 2.22.1 release includes a Neuron Driver update that resolves DMA abort errors on Trainium2 devices. These errors were previously occurring in the Neuron Runtime during specific workload executions.

Neuron 2.22.0 (04/03/2025)#

Table of contents

What’s New#

The Neuron 2.22 release includes performance optimizations, enhancements and new capabilities across the Neuron software stack.

For inference workloads, the NxD Inference library now supports Llama-3.2-11B model and supports multi-LoRA serving, allowing customers to load and serve multiple LoRA adapters. Flexible quantization features have been added, enabling users to specify which model layers or NxDI modules to quantize. Asynchronous inference mode has also been introduced, improving performance by overlapping Input preparation with model execution.

For training, we added LoRA supervised fine-tuning to NxD Training to enable additional model customization and adaptation.

Neuron Kernel Interface (NKI): This release adds new APIs in nki.isa, nki.language, and nki.profile. These enhancements provide customers with greater flexibility and control.

The updated Neuron Runtime includes optimizations for reduced latency and improved device memory footprint. On the tooling side, the Neuron Profiler 2.0 (beta) has added UI enhancements and new event type support.

Neuron DLCs: this release reduces DLC image size by up to 50% and enables faster build times with updated Dockerfiles structure. On the Neuron DLAMI side, new PyTorch 2.5 single framework DLAMIs have been added for Ubuntu 22.04 and Amazon Linux 2023, along with several new virtual environments within the Neuron Multi Framework DLAMIs.

More release content can be found in the table below and each component release notes.

What’s New Details Instances
NxD Core (neuronx-distributed) NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n,Trn2
NxD Inference (neuronx-distributed-inference) NxD Inference Release Notes (neuronx-distributed-inference) Inf2, Trn1/Trn1n,Trn2
NxD Training (neuronx-distributed-training) NxD Training Release Notes (neuronx-distributed-training) Trn1/Trn1n,Trn2
PyTorch NeuronX (torch-neuronx) PyTorch Neuron (torch-neuronx) release notes Trn1/Trn1n,Inf2,Trn2
NeuronX Nemo Megatron for Training neuronx-nemo-megatron github repo and AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) Neuron Compiler (neuronx-cc) release notes Trn1/Trn1n,Inf2,Trn2
Neuron Kernel Interface (NKI) Neuron Kernel Interface (NKI) release notes Trn1/Trn1n,Inf2
Neuron Tools Neuron System Tools Inf1,Inf2,Trn1/Trn1n,Trn2
Neuron Runtime Neuron Runtime Release Notes Inf1,Inf2,Trn1/Trn1n,Trn2
Transformers NeuronX (transformers-neuronx) for Inference Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
Neuron Deep Learning AMIs (DLAMIs) Neuron DLAMI User Guide Inf1,Inf2,Trn1/Trn1n
Neuron Deep Learning Containers (DLCs) Neuron DLC Release Notes Inf1,Inf2,Trn1/Trn1n
Release Annoucements Announcing end of support for Neuron DET tool starting next release Announcing migration of NxD Core examples from NxD Core repository to NxD Inference repository in next release Announcing end of support for Python 3.8 in future releases Announcing end of support for PyTorch 1.13 starting next release Announcing end of support for PyTorch 2.1 starting next release Neuron no longer includes support for Ubuntu20 DLCs and DLAMIs starting this release PyTorch Neuron versions 1.9 and 1.10 no longer supported See more at Announcements Inf1, Inf2, Trn1/Trn1n

For detailed release artificats, see Release Artifacts.

Neuron 2.21.1 (01/14/2025)#

Neuron 2.21.1 release pins Transformers NeuronX dependency to transformers<4.48 and fixes DMA abort errors on Trn2.

Additionally, this release addresses NxD Core and Training improvements, including fixes for sequence parallel support in quantized models and a new flag for dtype control in Llama3/3.1 70B configurations. See NxD Training Release Notes (neuronx-distributed-training) for details.

NxD Inference update includes minor bug fixes for sampling parameters. See NxD Inference Release Notes.

Neuron supported DLAMIs and DLCs have been updated to Neuron 2.21.1 SDK. Users should be aware of an incompatibility between Tensorflow-Neuron 2.10 (Inf1) and Neuron Runtime 2.21 in DLAMIs, which will be addressed in the next minor release. See Neuron DLAMI Release Notes.

The Neuron Compiler includes bug fixes and performance enhancements specifically targeting the Trn2 platform.

Neuron 2.21.0 (12/20/2024)#

Table of contents

What’s New#

Overview: Neuron 2.21.0 introduces support for AWS Trainium2 andTrn2 instances, including the trn2.48xlarge instance type and Trn2 UltraServer (Preview). The release adds new capabilities in both training and inference of large-scale models. It introduces NxD Inference (beta), a PyTorch-based library for deployment, Neuron Profiler 2.0 (beta), andPyTorch 2.5 support across the Neuron SDK, and Logical NeuronCore Configuration (LNC) for optimizing NeuronCore allocation. The release enables Llama 3.1 405B model inference on a single trn2.48xlarge instance.

NxD Inference: NxD Inference (beta) is a new PyTorch-based inference library for deploying large-scale models on AWS Inferentia and Trainium instances. It enables PyTorch model onboarding with minimal code changes and integrates with vLLM. NxDI supports various model architectures, including Llama versions for text processing (Llama 2, Llama 3, Llama 3.1, Llama 3.2, and Llama 3.3), Llama 3.2 multimodal for multimodal tasks, and Mixture-of-Experts (MoE) model architectures including Mixtral and DBRX. The library supports quantization methods, includes dynamic sampling, and is compatible with HuggingFace checkpoints and generate() API. NxDI also supports distributed strategies including tensor parallelism and incorporates speculative decoding techniques (Draft model and EAGLE). The release includes Llama 3.1 405B model sample, Llama 3.3 70B model sampleand Llama 3.1 405B model with speculative decoding for inference on a single trn2.48xlarge instance.

For more information, see NxD Inference documentation and check the NxD Inference Github repository: aws-neuron/neuronx-distributed-inference

Transformers NeuronX (TNx): This release introduces several new features, including flash decoding support for speculative decoding, and on-device generation in speculative decoding flows. It adds Eagle speculative decoding with greedy and lossless sampling, as well as support for CPU compilation and sharded model saving. Performance improvements include optimized MLP and QKV for Llama models with sequence parallel norm and control over concurrent compilation workers.

Training Highlights: NxD Training in this release adds support for HuggingFace Llama3/3.1 70B on trn2 instances, introduces DPO support for post-training model alignment, and adds support for Mixture-of-Experts (MoE) models including Mixtral 7B. The release includes improvedcheckpoint conversion capabilities and supports MoE with Tensor, Sequence, Pipeline, and Expert parallelism.

ML Frameworks: Neuron 2.21.0 adds support for PyTorch 2.5 and JAX 0.4.35.

Note

The CVEsCVE-2024-31583andCVE-2024-31580affect PyTorch versions 2.1 and earlier. Based on Amazon’s analysis, executing models on Trainium and Inferentia is not exposed to either of these vulnerabilities. We recommend upgrading to the new version of Torch-NeuronX by following the Neuron setup instructions.

Logical NeuronCore Configuration (LNC): This release introduces LNCfor Trainium2 instances, optimizing NeuronCore allocation for ML applications. LNC offers two configurations: default (LNC=2) combining two physical cores, and alternative (LNC=1) mapping each physical core individually. This feature allows users to efficiently manage resources for large-scale model training and deployment through runtime variables and compiler flags.

Neuron Profiler 2.0: The new profiler provides system and device-level profiling, timeline annotations, container integration, and support for distributed workloads. It includes trace export capabilities for Perfetto visualization and integration with JAX and PyTorch profilers, and support for Logical NeuronCore Configuration (LNC).

Neuron Kernel Interface (NKI): NKI now supports Trainium2 includingLogical NeuronCore Configuration (LNC), adds SPMD capabilities for multi-core operations, and includes new modules and APIs including support for float8_e5m2 datatype.

Deep Learning Containers (DLAMIs): This release expands support for JAX 0.4 within the Multi Framework DLAMI. It also introduces NxD Training, NxD Inference, and NxD Core withPyTorch 2.5 support. Additionally, a new Single Framework DLAMI for TensorFlow 2.10 on Ubuntu 22 is now available.

Deep Learning Containers (DLCs): This release introduces new DLCs for JAX 0.4 training and PyTorch 2.5.1 inference and training. All DLCs have been updated to Ubuntu 22, and the pytorch-inference-neuronx DLC now supports both NxD Inference and TNx libraries.

Documentation: Documentation updates include architectural details about Trainium2 and NeuronCore-v3, along with specifications and topology information for the trn2.48xlarge instance type and Trn2 UltraServer.

Software Maintenance: This release includes the following announcements:

Amazon Q: Use Q Developeras your Neuron Expert for general technical guidance and to jumpstart your NKI kernel development.

More release content can be found in the table below and each component release notes.

What’s New Details Instances
Known Issues and Limitations See 2.21.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Transformers NeuronX (transformers-neuronx) for Inference Flash decoding support for speculative decoding Added support for EAGLE speculative decoding with greedy and lossless sampling Enabled on-device generation support in speculative decoding flows See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n, Trn2
NxD Core (neuronx-distributed) Training: Added support for HuggingFace Llama3 70B with Trn2 instances Added DPO support for post-training model alignment See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n,Trn2
NxD Inference (neuronx-distributed-inference) Introduced new NxD Inference Library. See Introducing NeuronX Distributed (NxD) Inference Added Llama3.1 405B Inference Example on Trn2. See Tutorial: Deploying Llama3.1 405B (Trn2) Added Llama 3.2 Multimodal inference sample. See Tutorial: Deploying Llama3.2 Multimodal Models Added support for vLLM integration for NxD Inference. See vLLM User Guide for NxD Inference Introduced Open Source Github repository for NxD Inference. See aws-neuron/neuronx-distributed-inference See more at NxD Inference Release Notes (neuronx-distributed-inference) Inf2, Trn1/Trn1n,Trn2
NxD Training (neuronx-distributed-training) Added support for HuggingFace Llama3/3.1 70B with Trn2 instances Added support for Mixtral 8x7B Megatron and HuggingFace models Added support for custom pipeline parallel cuts in HuggingFace Llama3 Added support for DPO post-training model alignment See more at NxD Training Release Notes (neuronx-distributed-training) Trn1/Trn1n,Trn2
PyTorch NeuronX (torch-neuronx) Introduced PyTorch 2.5 support See more at PyTorch Neuron (torch-neuronx) release notes Trn1/Trn1n,Inf2,Trn2
NeuronX Nemo Megatron for Training Added support for HuggingFace to NeMo checkpoint conversion when virtual pipeline parallel is enabled. Added collective compute coalescing for ZeRO-1 optimizer See more at neuronx-nemo-megatron github repo and AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) Minor bug fixes and performance enhancements for the Trn2 platform. See more at Neuron Compiler (neuronx-cc) release notes Trn1/Trn1n,Inf2,Trn2
Neuron Kernel Interface (NKI) Added api/nki.compiler module with Allocation Control and Kernel decorators Added new nki.isa APIs. See api/nki.isa Added new nki.language APIs. See api/nki.language Added new kernels (allocated_fused_self_attn_for_SD_small_head_size, allocated_fused_rms_norm_qkv). See api/nki.kernels See more at Neuron Kernel Interface (NKI) release notes Trn1/Trn1n,Inf2
Neuron Deep Learning AMIs (DLAMIs) Added support for Trainium2 chips within the Neuron Multi Framework DLAMI. Added support for JAX 0.4 to Neuron Multi Framework DLAMI. Added NxD Training (NxDT), NxD Inference (NxDI) and NxD Core PyTorch 2.5 support within the Neuron Multi Framework DLAMI. See more at Neuron DLAMI User Guide Inf1,Inf2,Trn1/Trn1n
Neuron Deep Learning Containers (DLCs) Added new pytorch-inference-neuronx 2.5.1 and pytorch-training-neuronx 2.5.1 DLCs Added new jax-training-neuronx 0.4 Training DLC See more at Neuron DLC Release Notes Inf1,Inf2,Trn1/Trn1n
Neuron Tools Introduced Neuron Profiler 2.0. See Neuron Profiler 2.0 (Beta) User Guide See more at Neuron System Tools Inf1,Inf2,Trn1/Trn1n,Trn2
Neuron Runtime Added runtime support to fail in case of out-of-bound memory access when DGE is enabled. Added support for 4-rank replica group on adjacent Neuron cores on TRN1/TRN1N See more at Neuron Runtime Release Notes Inf1,Inf2,Trn1/Trn1n,Trn2
Release Annoucements Announcing end of support for Neuron DET tool starting next release Announcing migration of NxD Core examples from NxD Core repository to NxD Inference repository in next release Announcing end of support for Python 3.8 in future releases Announcing end of support for PyTorch 1.13 starting next release Announcing end of support for PyTorch 2.1 starting next release Neuron no longer includes support for Ubuntu20 DLCs and DLAMIs starting this release PyTorch Neuron versions 1.9 and 1.10 no longer supported See more at Announcements Inf1, Inf2, Trn1/Trn1n
Documentation Updates See Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n, Trn2
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1, Trn2
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1, Trn2

2.21.0 Known Issues and Limitations#

Neuron 2.21.0 Beta (12/03/2024)#

Note

This release (Neuron 2.21 Beta) was only tested with Trn2 instances. The next release (Neuron 2.21) will support all instances (Inf1, Inf2, Trn1, and Trn2).

For access to this release (Neuron 2.21 Beta), please contact your account manager.

This release (Neuron 2.21 beta) introduces support for AWS Trainium2 and Trn2 instances, including the trn2.48xlarge instance type and Trn2 UltraServer. The release showcases Llama 3.1 405B model inference using NxD Inference on a single trn2.48xlarge instance, and FUJI 70B model training using the AXLearn library across eight trn2.48xlarge instances.

NxD Inference, a new PyTorch-based library for deploying large language models and multi-modality models, is introduced in this release. It integrates with vLLM and enables PyTorch model onboarding with minimal code changes. The release also adds support for AXLearn training for JAX models.

The new Neuron Profiler 2.0 introduced in this release offers system and device-level profiling, timeline annotations, and container integration. The profiler supports distributed workloads and provides trace export capabilities for Perfetto visualization.

The documentation has been updated to include architectural details about Trainium2 and NeuronCore-v3, along with specifications and topology information for the trn2.48xlarge instance type and Trn2 UltraServer.

Use Q Developer as your Neuron Expert for general technical guidance and to jumpstart your NKI kernel development.

Note

For the latest release that supports Trn1, Inf2 and Inf1 instances, please see Neuron Release 2.20.2

Neuron 2.20.2 (11/20/2024)#

Neuron 2.20.2 release fixes a stability issue in Neuron Scheduler Extension that previously caused crashes in Kubernetes (K8) deployments. See Neuron K8 Release Notes.

This release also addresses a security patch update to Neuron Driver that fixes a kernel address leak issue. See more on Neuron Driver Release Notes and Neuron Runtime Release Notes.

Addtionally, Neuron 2.20.2 release updates torch-neuronx and libneuronxla packages to add support for torch-xla 2.1.5 package which fixes checkpoint loading issues with Zero Redundancy Optimizer (ZeRO-1). See PyTorch Neuron (torch-neuronx) release notes and Neuron XLA pluggable device (libneuronxla) release notes.

Neuron supported DLAMIs and DLCs are updated with this release (Neuron 2.20.2 SDK). The Training DLC is also updated to address the version dependency issues in NxD Training library. See Neuron DLC Release Notes.

NxD Training library in Neuron 2.20.2 release is updated to transformers 4.36.0 package. See NxD Training Release Notes (neuronx-distributed-training).

Neuron 2.20.1 (10/25/2024)#

Neuron 2.20.1 release addresses an issue with the Neuron Persistent Cache that was brought forth in 2.20 release. In the 2.20 release, the Neuron persistent cache issue resulted in a cache-miss scenario when attempting to load a previously compiled Neuron Executable File Format (NEFF) from a different path or Python environment than the one used for the initial Neuron SDK installation and NEFF compilation. This release resolves the cache-miss problem, ensuring that NEFFs can be loaded correctly regardless of the path or Python environment used to install the Neuron SDK, as long as they were compiled using the same Neuron SDK version.

This release also addresses the excessive lock wait time issue during neuron_parallel_compile graph extraction for large cluster training. See PyTorch Neuron (torch-neuronx) release notes and Neuron XLA pluggable device (libneuronxla) release notes.

Additionally, Neuron 2.20.1 introduces new Multi Framework DLAMI for Amazon Linux 2023 (AL2023) that customers can use to easily get started with latest Neuron SDK on multiple frameworks that Neuron supports. See Neuron DLAMI Release Notes.

Neuron 2.20.1 Training DLC is also updated to pre-install the necessary dependencies and support NxD Training library out of the box. See Neuron DLC Release Notes

Neuron 2.20.0 (09/16/2024)#

Table of contents

What’s New#

Overview: Neuron 2.20 release introduces usability improvements and new capabilities across training and inference workloads. A key highlight is the introduction of Neuron Kernel Interface (beta). NKI, pronounced ‘Nicky’, is enabling developers to build optimized custom compute kernels for Trainium and Inferentia. Additionally, this release introduces NxD Training (beta), a PyTorch-based library enabling efficient distributed training, with a user-friendly interface compatible with NeMo. This release also introduces the support for the JAX framework (beta).

Neuron 2.20 also adds inference support for Pixart-alpha and Pixart-sigma Diffusion-Transformers (DiT) models, and adds support for Llama 3.1 8B, 70B and 405B models inference supporting up to 128K context length.

Neuron Kernel Interface: NKI is a programming interface enabling developers to build optimized compute custom kernels on top of Trainium and Inferentia. NKI empowers developers to enhance deep learning models with new capabilities, performance optimizations, and scientific innovation. It natively integrates with PyTorch and JAX, providing a Python-based programming environment with Triton-like syntax and tile-level semantics, offering a familiar programming experience for developers. All of our NKI work is shared as open source, enabling the community developers to collaborate and use these kernels in their projects, improve existing kernels, and contribute new NKI kernels. The list of kernels we are introducing includes Optimized Flash Attention NKI kernel (flash_attention), a NKI kernel with an optimized implementation of Mamba model architecture (mamba_nki_kernels) and Optimized Stable Diffusion Attention kernel (fused_sd_attention_small_head). In addition to NKI kernel samples for average_pool2d, rmsnorm, tensor_addition, layernorm, transpose_2d, and matrix_multiplication.

For more information see NKI section and check the NKI samples Github repository: aws-neuron/nki-samples

NxD Training (NxDT): NxDT is a PyTorch-based library that adds support for user-friendly distributed training experience through a YAML configuration file compatible with NeMo,, allowing users to easily set up their training workflows. At the same time, NxDT maintains flexibility, enabling users to choose between using the YAML configuration file, PyTorch Lightning Trainer, or writing their own custom training script using the NxD Core. The library supports PyTorch model classes including Hugging Face and Megatron-LM. Additionally, it leverages NeMo’s data engineering and data science modules enabling end-to-end training workflows on NxDT, and providing compatability with NeMo through minimal changes to the YAML configuration file for models that are already supported in NxDT. Furthermore, the functionality of the Neuron NeMo Megatron (NNM) library is now part of NxDT, ensuring a smooth migration path from NNM to NxDT.

For more information see NxD Training (beta) and check the NxD Training Github repository: aws-neuron/neuronx-distributed-training

Training Highlights: This release adds support for Llama 3.1 8B and 70B model training up to 32K sequence length (beta). It also adds support for torch.autocast() for native PyTorch mixed precision support and PEFT LoRA model training.

Inference Highlights: Neuron 2.20 adds support for Llama 3.1 models (405b, 70b, and 8b variants) and introduces new features like on-device top-p sampling for improved performance, support for up to 128K context length through Flash Decoding, and multi-node inference for large models like Llama-3.1-405B. Furthermore, this release improves model loading in Transformers Neuronx for models like Llama-3 by loading the pre-sharded or pre-transformed weights and adds support to Diffusion-Transformers (DiT) models such as Pixart-alpha and Pixart-sigma.

Compiler: This release introduces Neuron Compiler support for RMSNorm and RMSNormDx operators, along with enhanced performance for the sort operator.

System Tools: As for the Neuron Tools, it enables NKI profiling support in the Neuron Profiler and introduces improvements to the Neuron Profiler UI.

Neuron Driver: This release adds support for the Rocky Linux 9.0 operating system.

Neuron Containers: This release introduces Neuron Helm Chart, which helps streamline the deployment of AWS Neuron components on Amazon EKS. See Neuron Helm Chart Github repository: aws-neuron/neuron-helm-charts. Additionaly, this release adds ECS support for the “Neuron Node Problem Detector and Recovery” artifact. See Neuron Problem Detector And Recovery.

Neuron DLAMIs and DLCs: This release includes the addition of the NxDT package to various Neuron DLAMIs (Multi-Framework Neuron DLAMI, PyTorch 1.13 Neuron DLAMI, and PyTorch 2.1 Neuron DLAMI) and the inclusion of NxDT in the PyTorch 1.13 Training Neuron DLC and PyTorch 2.1 Training Neuron DLC.

Software Maintenance Policy: This release also updates Neuron SDK software maintenance poclicy, For more information see Neuron Software Maintenance policy

More release content can be found in the table below and each component release notes.

What’s New Details Instances
Known Issues and Limitations See 2.20.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Transformers NeuronX (transformers-neuronx) for Inference Support for on-device sampling (Top P) and dynamic sampling (per request parameters) with Continuous batching. See developer guide Support for Flash Decoding to enable inference for higher sequence lengths of upto 128K. See developer guide and Llama-3.1-8B model sample. Support for multi-node inference for large models like Llama-3.1-405B. See developer guide and Llama-3.1-405B model sample. Support for bucketing, multi-node inference , on-device sampling and other improvements in Neuron vLLM integration. See developer guide Support for Llama 3.1 models (405B, 70B, and 8B variants). See samples for Llama-3.1-405B , Llama-3.1-70B and Llama-3.1-8B Support for improved model loading for models like Llama-3 by loading the pre-sharded or pre-transformed weights. See serialization support in developer guide. Support for ROPE scaling for Llama 3 and Llama 3.1 models. See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
NxD Core (neuronx-distributed) Training: Support for LoRA finetuning Support for Mixed precision enhancements Inference: Suppport for DBRX and Mixtral inference samples. See samples for DBRX and Mixtral Support for sequence length autobucketing to improve inference performance. Support for improved tracing in the inference samples. See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n
NxD Training (neuronx-distributed-training) First release of NxD Training (beta) See more at NxD Training Release Notes (neuronx-distributed-training) Trn1/Trn1n
PyTorch NeuronX (torch-neuronx) Support for inference of Diffusion-Transformers (DiT) models such as Pixart-alpha and Pixart-sigma. See samples for Pixart-alpha and Pixart-sigma. Support for inference of wav2vec2-conformer models. See samples for inference of wav2vec2-conformer with relative position embeddings and rotary position embeddings See more at PyTorch Neuron (torch-neuronx) release notes Trn1/Trn1n,Inf2
NeuronX Nemo Megatron for Training Fixed issue with linear warmup with cosine annealing Fixed indexing issues with MPI job checkpoint conversion. Fixed pipeline parallel bug for NeMo to HF checkpoint conversion See more at neuronx-nemo-megatron github repo and AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) Memory optimization that will reduce the generated compiler artifacts size (i.e., NEFFs) See more at Neuron Compiler (neuronx-cc) release notes Trn1/Trn1n,Inf2
Neuron Kernel Interface (NKI) First Release on Neuron Kernel Interface (NKI) See more at Neuron Kernel Interface (NKI) release notes Trn1/Trn1n,Inf2
Neuron Deep Learning AMIs (DLAMIs) Support for neuronx-distributed-training library in PyTorch Neuron DLAMI virtual enviornments. See Neuron DLAMI User Guide Updated existing Neuron supported DLAMIs with Neuron 2.20 SDK release. See more at :ref:`Neuron DLAMI Release Notes `_ Inf1,Inf2,Trn1/Trn1n
Neuron Deep Learning Containers (DLCs) Updated existing PyTorch Neuron DLCs with Neuron 2.20 SDK release. Support for neuronx-distributed-training library in pytorch-training-neuronx DLCs. See more at Neuron DLC Release Notes Inf1,Inf2,Trn1/Trn1n
Neuron Tools Improvements in Neuron Profile See more at Neuron System Tools Inf1,Inf2,Trn1/Trn1n
Neuron Runtime Introduced a sysfs memory usage counter for DMA rings (reference) See more at Neuron Runtime Release Notes Inf1,Inf2,Trn1/Trn1n
Release Annoucements Announcing Name Change for Neuron Component ‘neurondevice’ resource name in Neuron Device K8s plugin no longer supported ‘neuron-device-version’ field in neuron-monitor no longer supported Tensorflow-Neuron 1.x no longer supported Announcing maintenance mode for torch-neuron 1.9 and 1.10 versions Neuron Runtime no longer supports Amazon Linux 2 (AL2) See more at Announcements Inf1, Inf2, Trn1/Trn1n
Documentation Updates See Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

2.20.0 Known Issues and Limitations#

Neuron Components Release Notes#

Inf1, Trn1/Trn1n and Inf2 common packages#

Component Instance/s Package/s Details
Neuron Runtime Trn1/Trn1n, Inf1, Inf2 Trn1/Trn1n: aws-neuronx-runtime-lib (.deb, .rpm) Inf1: Runtime is linked into the ML frameworks packages Neuron Runtime Release Notes
Neuron Runtime Driver Trn1/Trn1n, Inf1, Inf2 aws-neuronx-dkms (.deb, .rpm) Neuron Driver Release Notes
Neuron System Tools Trn1/Trn1n, Inf1, Inf2 aws-neuronx-tools (.deb, .rpm) Neuron System Tools
Containers Trn1/Trn1n, Inf1, Inf2 aws-neuronx-k8-plugin (.deb, .rpm) aws-neuronx-k8-scheduler (.deb, .rpm) aws-neuronx-oci-hooks (.deb, .rpm) Neuron K8 Release Notes Neuron Containers Release Notes
NeuronPerf (Inference only) Trn1/Trn1n, Inf1, Inf2 neuronperf (.whl) NeuronPerf 1.x Release Notes
TensorFlow Model Server Neuron Trn1/Trn1n, Inf1, Inf2 tensorflow-model-server-neuronx (.deb, .rpm) TensorFlow-Model-Server-Neuron (tensorflow-modeslserver-neuronx) Release Notes

Trn1/Trn1n and Inf2 only packages#

Component Instance/s Package/s Details
PyTorch Neuron Trn1/Trn1n, Inf2 torch-neuronx (.whl) PyTorch Neuron (torch-neuronx) release notes PyTorch Neuron (torch-neuronx) - Supported Operators
TensorFlow Neuron Trn1/Trn1n, Inf2 tensorflow-neuronx (.whl) TensorFlow 2.x (tensorflow-neuronx) Release Notes
Neuron Compiler (Trn1/Trn1n, Inf2 only) Trn1/Trn1n, Inf2 neuronx-cc (.whl) Neuron Compiler (neuronx-cc) release notes
Neuron Kernel Interface (NKI) Compiler (Trn1/Trn1n, Inf2 only) Trn1/Trn1n, Inf2 Supported within neuronx-cc (.whl) Neuron Kernel Interface (NKI) release notes
Collective Communication library Trn1/Trn1n, Inf2 aws-neuronx-collective (.deb, .rpm) Neuron Collectives Release Notes
Neuron Custom C++ Operators Trn1/Trn1n, Inf2 aws-neuronx-gpsimd-customop (.deb, .rpm) aws-neuronx-gpsimd-tools (.deb, .rpm) Neuron Custom C++ Library Release Notes Neuron Custom C++ Tools Release Notes
Transformers Neuron Trn1/Trn1n, Inf2 transformers-neuronx (.whl) Transformers Neuron (transformers-neuronx) release notes
NxD Training Trn1/Trn1n, Inf2 neuronx-distributed-training (.whl) NxD Training Release Notes (neuronx-distributed-training)
NxD Core Trn1/Trn1n, Inf2 neuronx-distributed (.whl) NxD Core Release Notes (neuronx-distributed)
AWS Neuron Reference for NeMo Megatron Trn1/Trn1n neuronx-nemo-megatron github repo AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes

Inf1 only packages#

Component Instance/s Package/s Details
PyTorch Neuron Inf1 torch-neuron (.whl) PyTorch Neuron (torch-neuron) release notes PyTorch Neuron (torch-neuron) Supported operators
TensorFlow Neuron Inf1 tensorflow-neuron (.whl) TensorFlow Neuron (tensorflow-neuron (TF1.x)) Release Notes TensorFlow Neuron (tensorflow-neuron (TF1.x)) Supported operators TensorFlow 2.x (tensorflow-neuron) Release Notes
Apache MXNet Inf1 mx_neuron (.whl) Apache MXNet Neuron Release Notes Neuron Apache MXNet Supported operators
Neuron Compiler (Inf1 only) Inf1 neuron-cc (.whl) Neuron Compiler (neuron-cc) for Inf1 Release Notes Neuron Supported operators

Neuron 2.19.1 (07/19/2024)#

This release (Neuron 2.19.1) addresses an issue with the Neuron Persistent Cache that was introduced in the previous release, Neuron 2.19. The issue resulted in a cache-miss scenario when attempting to load a previously compiled Neuron Executable File Format (NEFF) from a different path or Python environment than the one used for the initial Neuron SDK installation and NEFF compilation. This release resolves the cache-miss problem, ensuring that NEFFs can be loaded correctly regardless of the path or Python environment used to install the Neuron SDK, as long as they were compiled using the same Neuron SDK version.

Neuron 2.19.0 (07/03/2024)#

Table of contents

What’s New#

Neuron 2.19 release adds Llama 3 training support and introduces Flash Attention kernel support to enable LLM training and inference for large sequence lengths. Neuron 2.19 also introduces new features and performance improvements to LLM training, improves LLM inference performance for Llama 3 model by upto 20%, and adds tools for monitoring, problem detection and recovery in Kubernetes (EKS) environments, improving efficiency and reliability.

Training highlights: LLM model training user experience using NeuronX Distributed (NxD) is improved by support for Flash Attention to enable training with longer sequence lengths >= 8K. Neuron 2.19 adds support for Llama 3 model training. This release also adds support for Interleaved pipeline parallelism to reduce idle time (bubble size) and enhance training efficiency and resource utilization for large cluster sizes.

Inference highlights: Flash Attention kernel support in the Transformers NeuronX library enables LLM inference for context lengths of up to 32k. This release also adds [Beta] support for continuous batching with mistralai/Mistral-7B-v0.2 in Transformers NeuronX.

Tools and Neuron DLAMI/DLC highlights: This release introduces the new Neuron Node Problem Detector and Recovery plugin in EKS supported Kubernetes environments:a tool to monitor the health of Neuron instances and triggers automatic node replacement upon detecting an unrecoverable error. Neuron 2.19 introduces the new Neuron Monitor container to enable easy monitoring of Neuron metrics in Kubernetes, and adds monitoring support with Prometheus and Grafana. This release also introduces new PyTorch 2.1 and PyTorch 1.13 single framework DLAMIs for Ubuntu 22. Neuron DLAMIs and Neuron DLCs are also updated to support this release (Neuron 2.19).

More release content can be found in the table below and each component release notes.

What’s New Details Instances
Known Issues and Limitations See 2.19.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Transformers NeuronX (transformers-neuronx) for Inference Support for Flash Attention kernel in Llama models to enable inference for higher sequence lengths. See developer guide and Llama-3-8B model sample. Support for running Top-K sampling on Neuron device for generation in Mixtral models. See Mixtral-8x7b sample. [Beta] Support for Continuous batching with mistralai/Mistral-7B-Instruct-v0.2 model inference. See developer guide. See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Training Support for Interleaved pipeline parallelism to reduce idle time (bubble size) and enhance training efficiency and resource utilization for large cluster sizes. See api guide , developer guide Support for Flash Attention kernel to enable training with longer sequence lengths. See Llama-3 sample with 8K sequence length training. See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Inference Support for Flash Attention kernel for longer sequence length inference. See [CodeLlama-13b Inference with 16k sequence length] [Beta] Support for speculative decoding. See developer guide. See more at NxD Core Release Notes (neuronx-distributed) Inf2,Trn1/Trn1n
PyTorch NeuronX (torch-neuronx) Support for FP32 master weights and BF16 all-gather during Zero1 training to enhance training efficiency. Support to add custom SILU activation functions by configuring NEURON_CUSTOM_SILU variable See more at PyTorch Neuron (torch-neuronx) release notes Trn1/Trn1n,Inf2
NeuronX Nemo Megatron for Training Support for FP32 gradient accumulation enhancing accuracy for large model training. Support for Zero1 training with master weights Support for Flash Attention kernel to train with longer sequence lengths (greater than 8K) See more at neuronx-nemo-megatron github repo and AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) Support for Flash Attention kernel to enable usage of long sequence lengths during training and inference. See more at Neuron Compiler (neuronx-cc) release notes Trn1/Trn1n,Inf2
Neuron DLAMI and DLC Neuron DLAMIs are updated with latest 2.19 Neuron SDK. See Neuron DLAMI User Guide New Neuron Single Framework DLAMIs with PyTorch-2.1 and PyTorch-1.13 for Ubuntu 22. See Neuron DLAMI User Guide New Base Deep Learning AMI (DLAMI) for Ubuntu 22. See Neuron DLAMI User Guide PyTorch 1.13 and PyTorch 2.1 Inference and Training DLCs are updated with latest 2.19 Neuron SDK. See Neuron Containers PyTorch 1.13 Inference and PyTorch 2.1 Inference DLCs are updated with TorchServe v0.11.0. See Neuron Containers Inf1,Inf2,Trn1/Trn1n
Neuron Tools Support for new Neuron Node Problem Detector and Recovery plugin in EKS supported kubernetes environments that monitors health of Neuron instances and triggers automatic node replacement upon detecting an unrecoverable error. See configuration and tutorial. Support for new Neuron Monitor container to enable easy monitoring of Neuron metrics in Kubernetes. Supports monitoring with Prometheus and Grafana. See tutorial Support for Neuron scheduler extension to enforce allocation of contiguous Neuron Devices for the pods based on the Neuron instance type. See tutorial Neuron Profiler bugfixes and UI updates, including improvements to visualizing collective operations and to the consistency of information being displayed Added memory usage metrics and device count information to neuron-monitor See more at Neuron System Tools Inf1,Inf2,Trn1/Trn1n
Neuron Runtime Support for dynamic Direct Memory Access (DMA) that reduces memory usage during runtime. Runtime Enhancements that improve collectives performance See more at Neuron Runtime Release Notes Inf1,Inf2,Trn1/Trn1n
Other Documentation Updates Announced maintenance mode of MxNet. See Neuron support for MxNet enters maintenance mode Announced End of support of Neuron TensorFlow 1.x (Inf1). See Announcing end of support for Tensorflow-Neuron 1.x Announce End of support for AL2. See Announcing end of support for Neuron Runtime support of Amazon Linux 2 (AL2) See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

2.19.0 Known Issues and Limitations#

Neuron Components Release Notes#

Inf1, Trn1/Trn1n and Inf2 common packages#

Component Instance/s Package/s Details
Neuron Runtime Trn1/Trn1n, Inf1, Inf2 Trn1/Trn1n: aws-neuronx-runtime-lib (.deb, .rpm) Inf1: Runtime is linked into the ML frameworks packages Neuron Runtime Release Notes
Neuron Runtime Driver Trn1/Trn1n, Inf1, Inf2 aws-neuronx-dkms (.deb, .rpm) Neuron Driver Release Notes
Neuron System Tools Trn1/Trn1n, Inf1, Inf2 aws-neuronx-tools (.deb, .rpm) Neuron System Tools
Neuron DLAMI Trn1/Trn1n, Inf1, Inf2 Neuron DLAMI Release Notes.
Neuron DLC Trn1/Trn1n, Inf1, Inf2 Neuron DLC Release Notes
Containers Trn1/Trn1n, Inf1, Inf2 aws-neuronx-k8-plugin (.deb, .rpm) aws-neuronx-k8-scheduler (.deb, .rpm) aws-neuronx-oci-hooks (.deb, .rpm) Neuron K8 Release Notes Neuron Containers Release Notes
NeuronPerf (Inference only) Trn1/Trn1n, Inf1, Inf2 neuronperf (.whl) NeuronPerf 1.x Release Notes
TensorFlow Model Server Neuron Trn1/Trn1n, Inf1, Inf2 tensorflow-model-server-neuronx (.deb, .rpm) TensorFlow-Model-Server-Neuron (tensorflow-modeslserver-neuronx) Release Notes
Neuron Documentation Trn1/Trn1n, Inf1, Inf2 Neuron Documentation Release Notes

Trn1/Trn1n and Inf2 only packages#

Component Instance/s Package/s Details
PyTorch Neuron Trn1/Trn1n, Inf2 torch-neuronx (.whl) PyTorch Neuron (torch-neuronx) release notes PyTorch Neuron (torch-neuronx) - Supported Operators
TensorFlow Neuron Trn1/Trn1n, Inf2 tensorflow-neuronx (.whl) TensorFlow 2.x (tensorflow-neuronx) Release Notes
Neuron Compiler (Trn1/Trn1n, Inf2 only) Trn1/Trn1n, Inf2 neuronx-cc (.whl) Neuron Compiler (neuronx-cc) release notes
Collective Communication library Trn1/Trn1n, Inf2 aws-neuronx-collective (.deb, .rpm) Neuron Collectives Release Notes
Neuron Custom C++ Operators Trn1/Trn1n, Inf2 aws-neuronx-gpsimd-customop (.deb, .rpm) aws-neuronx-gpsimd-tools (.deb, .rpm) Neuron Custom C++ Library Release Notes Neuron Custom C++ Tools Release Notes
Transformers Neuron Trn1/Trn1n, Inf2 transformers-neuronx (.whl) Transformers Neuron (transformers-neuronx) release notes
Neuron Distributed Trn1/Trn1n, Inf2 neuronx-distributed (.whl) NxD Core Release Notes (neuronx-distributed)
AWS Neuron Reference for NeMo Megatron Trn1/Trn1n neuronx-nemo-megatron github repo AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes

Note

In next releases aws-neuronx-tools and aws-neuronx-runtime-lib will add support for Inf1.

Inf1 only packages#

Component Instance/s Package/s Details
PyTorch Neuron Inf1 torch-neuron (.whl) PyTorch Neuron (torch-neuron) release notes PyTorch Neuron (torch-neuron) Supported operators
TensorFlow Neuron Inf1 tensorflow-neuron (.whl) TensorFlow Neuron (tensorflow-neuron (TF1.x)) Release Notes TensorFlow Neuron (tensorflow-neuron (TF1.x)) Supported operators TensorFlow 2.x (tensorflow-neuron) Release Notes
Apache MXNet Inf1 mx_neuron (.whl) Apache MXNet Neuron Release Notes Neuron Apache MXNet Supported operators
Neuron Compiler (Inf1 only) Inf1 neuron-cc (.whl) Neuron Compiler (neuron-cc) for Inf1 Release Notes Neuron Supported operators

Neuron 2.18.2 (04/25/2024)#

Patch release with minor Neuron Compiler bug fixes and enhancements. See more in Neuron Compiler (neuronx-cc) release notes

Neuron 2.18.1 (04/10/2024)#

Neuron 2.18.1 release introduces Continuous batching(beta) and Neuron vLLM integration(beta) support in Transformers NeuronX library that improves LLM inference throughput. This release also fixes hang issues related to Triton Inference Server as well as updating Neuron DLAMIs and DLCs with this release(2.18.1). See more in Transformers Neuron (transformers-neuronx) release notes and Neuron Compiler (neuronx-cc) release notes

Neuron 2.18.0 (04/01/2024)#

Table of contents

What’s New#

Neuron 2.18 release introduces stable support (out of beta) for PyTorch 2.1, introduces new features and performance improvements to LLM training and inference, and updates Neuron DLAMIs and Neuron DLCs to support this release (Neuron 2.18).

Training highlights: LLM model training user experience using NeuronX Distributed (NxD) is improved by introducing asynchronous checkpointing. This release also adds support for auto partitioning pipeline parallelism in NxD and introduces Pipeline Parallelism in PyTorch Lightning Trainer (beta).

Inference highlights: Speculative Decoding support (beta) in TNx library improves LLM inference throughput and output token latency(TPOT) by up to 25% (for LLMs such as Llama-2-70B). TNx also improves weight loading performance by adding support for SafeTensor checkpoint format. Inference using Bucketing in PyTorch NeuronX and NeuronX Distributed is improved by introducing auto-bucketing feature. This release also adds a new sample for Mixtral-8x7B-v0.1 and mistralai/Mistral-7B-Instruct-v0.2 in TNx.

Neuron DLAMI and Neuron DLC support highlights: This release introduces new Multi Framework DLAMI for Ubuntu 22 that customers can use to easily get started with latest Neuron SDK on multiple frameworks that Neuron supports as well as SSM parameter support for DLAMIs to automate the retrieval of latest DLAMI ID in cloud automation flows. Support for new Neuron Training and Inference Deep Learning containers (DLCs) for PyTorch 2.1, as well as a new dedicated GitHub repository to host Neuron container dockerfiles and a public Neuron container registry to host Neuron container images.

More release content can be found in the table below and each component release notes.

What’s New Details Instances
Transformers NeuronX (transformers-neuronx) for Inference [Beta] Support for Speculative Decoding API. See developer guide and Llama-2-70B sample Support for SafeTensors checkpoint format with improved weight loading performance. See developer guide Support for running Top-K sampling on Neuron Device for improved performance. See developer guide Code Llama model inference sample with 16K input seq length. See sample [Beta] Support for streaming API and stopping criteria API. See developer guide Support for Mixtral-8x7B-v0.1 model inference. See sample [Beta] Support for mistralai/Mistral-7B-Instruct-v0.2 model inference. See sample See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Training Support for Pipeline Parallelism training using PyTorch Lightning. See api guide , developer guide and tutorial Support for auto partitioning pipeline parallel stages when training large models. See api guide and Developer guide for Pipeline Parallelism Support for asynchronous checkpointing to improve the time it takes to save the checkpoint. See api guide , Developer guide for save/load checkpoint and Training Llama-3.1-70B, Llama-3-70B or Llama-2-13B/70B with Tensor Parallelism and Pipeline Parallelism Tutorial to fine-tune Llama-2-7B model using PyTorch Lightning and running evaluation on the fine-tuned model using Hugging Face optimum-neuron. See tutorial codegen25-7b-mono model training tutorial and script. See Training CodeGen2.5 7B with Tensor Parallelism and ZeRO-1 Optimizer See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Inference Support for auto bucketing in inference using a custom bucket kernel that can be passed as a bucket configuration to Tracing API. See api guide and Developer guide for Neuronx-Distributed Inference Support for inference with bf16 data type using XLA_USE_BF16=1 flag. See sample ([html] [notebook]) See more at NxD Core Release Notes (neuronx-distributed) Inf2,Trn1/Trn1n
PyTorch NeuronX (torch-neuronx) PyTorch 2.1 support is now stable (out of beta). See updated App Note and release notes for known issues. Support for auto bucketing in inference using a custom bucket kernel that can be passed as a bucket configuration to Tracing API. See [Autobucketing for Inference (|torch-neuronx )](../../frameworks/torch/torch-neuronx/programming-guide/inference/autobucketing-dev-guide.html#torch-neuronx-autobucketing-devguide) See more at PyTorch Neuron (torch-neuronx) release notes
NeuronX Nemo Megatron for Training Support for LoRa finetuning. See sample script Support for Mistral-7B training. See sample script Support for asynchronous checkpointing to improve the time it takes to save the checkpoint. See more at neuronx-nemo-megatron github repo and AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) New --enable-mixed-precision-accumulation compiler option to perform intermediate computations of an operation in FP32 regardless of the operation’s defined datatype. See Neuron Compiler CLI Reference Guide (neuronx-cc) See more at Neuron Compiler (neuronx-cc) release notes Trn1/Trn1n,Inf2
Neuron DLAMI and DLC New Neuron Multi Framework Deep Learning AMI (DLAMI) for Ubuntu 22 with separate virtual environments for PyTorch 2.1, PyTorch 1.13, Transformers NeuronX and Tensorflow 2.10. See setup guide and Neuron DLAMI User Guide Neuron Multi Framework Deep Learning AMI (DLAMI) is now the default Neuron AMI in QuickStart AMI list when launching Neuron instances for Ubuntu through AWS console. See setup guide Neuron DLAMIs for PyTorch 1.13 and Tensorflow 2.10 are updated with 2.18 Neuron SDK for both Ubuntu 20 and AL2. See Neuron DLAMI User Guide SSM parameter support for Neuron DLAMIs to find the DLAMI id with latest Neuron release SDK. See Neuron DLAMI User Guide New Neuron Deep Learning Containers(DLCs) for PyTorch 2.1 Inference and Training. See Neuron Containers PyTorch 1.13 Inference and Training DLCs are updated with latest 2.18 Neuron SDK and now also comes with pre-installed NeuronX Distributed library. See Neuron Containers Neuron DLCs are now hosted both in public Neuron ECR and as private images. Private images are only needed when using with Sagemaker. See Neuron Containers New Neuron Github Repository to host dockerfiles for Neuron DLCs. See neuron deep learning containers github repo Inf1,Inf2,Trn1/Trn1n
Other Documentation Updates App Note on snapshotting models with PyTorch NeuronX 2.1 to support dumping debug information. See How to debug models in PyTorch NeuronX Added announcement for Maintenance mode of TensorFlow 1.x. See Tensorflow-Neuron 1.x enters maintenance mode See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Known Issues and Limitations See 2.18.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

2.18.0 Known Issues and Limitations#

Neuron 2.17.0 (02/13/2024)#

What’s New#

Neuron 2.17 release improves small collective communication operators (smaller than 16MB) by up to 30%, which improves large language model (LLM) Inference performance by up to 10%. This release also includes improvements in Neuron Profiler and other minor enhancements and bug fixes.

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.16.1 (01/18/2024)#

Patch release with compiler bug fixes, updates to Neuron Device Plugin and Neuron Kubernetes Scheduler .

Neuron 2.16.0 (12/21/2023)#

Table of contents

What’s New#

Neuron 2.16 adds support for Llama-2-70B training and inference, upgrades to PyTorch 2.1 (beta) and adds new support for PyTorch Lightning Trainer (beta) as well as performance improvements and adding Amazon Linux 2023 support.

Training highlights: NeuronX Distributed library LLM models training performance is improved by up to 15%. LLM model training user experience is improved by introducing support of PyTorch Lightning Trainer (beta), and a new model optimizer wrapper which will minimize the amount of changes needed to partition models using NeuronX Distributed primitives.

Inference highlights: PyTorch inference now allows to dynamically swap different fine-tuned weights for an already loaded model, as well as overall improvements of LLM inference throughput and latency with Transformers NeuronX. Two new reference model samples for LLama-2-70b and Mistral-7b model inference.

User experience: This release introduces two new capabilities: A new tool, Neuron Distributed Event Tracing (NDET) which improves debuggability, and the support of profiling collective communication operators in the Neuron Profiler tool.

More release content can be found in the table below and each component release notes.

What’s New Details Instances
Transformers NeuronX (transformers-neuronx) for Inference [Beta] Support for Grouped Query Attention(GQA). See developer guide [Beta] Support for Llama-2-70b model inference using Grouped Query Attention. See tutorial [Beta] Support for Mistral-7B-Instruct-v0.1 model inference. See sample code See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Training [Beta] Support for PyTorch Lightning to train models using tensor parallelism and data parallelism . See api guide , developer guide and tutorial Support for Model and Optimizer Wrapper training API that handles the parallelization. See api guide and Developer guide for model and optimizer wrapper New save_checkpoint and load_checkpoint APIs to save/load checkpoints during distributed training. See Developer guide for save/load checkpoint Support for a new Query-Key-Value(QKV) module that provides the ability to replicate the Key Value heads and adds flexibility to use higher Tensor parallel degree during Training. See api guide and tutorial See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n
NeuronX Distributed (neuronx-distributed) for Inference Support weight-deduplication amongst TP shards by giving ability to save weights separately than in NEFF files. See developer guide Llama-2-7B model inference script ([html] [notebook]) See more at NxD Core Release Notes (neuronx-distributed) and Distributed Strategies APIs Inf2,Trn1/Trn1n
PyTorch NeuronX (torch-neuronx) [Beta]Support for] PyTorch 2.1. See introduce-pytorch-2-1 . See llama-2-13b inference sample. Support to separate out model weights from NEFF files and new replace_weights API to replace the separated weights. See PyTorch Neuron (torch-neuronx) Weight Replacement API for Inference and PyTorch NeuronX Tracing API for Inference [Beta] Script for training stabilityai/stable-diffusion-2-1-base and runwayml/stable-diffusion-v1-5 models . See script [Beta] Script for training facebook/bart-large model. See script [Beta] Script for stabilityai/stable-diffusion-2-inpainting model inference. See script Trn1/Trn1n,Inf2
Neuron Tools New Neuron Distributed Event Tracing (NDET) tool to help visualize execution trace logs and diagnose errors in multi-node workloads. See Neuron Distributed Event Tracing (NDET) User Guide Support for multi-worker jobs in neuron-profile . See Neuron Profile User Guide See more at Neuron System Tools Inf1/Inf2/Trn1/Trn1n
Documentation Updates Added setup guide instructions for AL2023 OS. See Setup Guide Added announcement for name change of Neuron Components. See Announcing Name Change for Neuron Components Added announcement for End of Support for PyTorch 1.10 . See Announcing End of Support for PyTorch Neuron version 1.10 Added announcement for End of Support for PyTorch 2.0 Beta. See Announcing End of Support for PyTorch NeuronX version 2.0 (beta) See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Known Issues and Limitations See 2.16.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

2.16.0 Known Issues and Limitations#

Neuron 2.15.2 (11/17/2023)#

Patch release that fixes compiler issues related to performance when training using neuronx-nemo-megatron library.

Neuron 2.15.1 (11/09/2023)#

Patch release to fix execution overhead issues in Neuron Runtime that were inadvertently introduced in 2.15 release.

Neuron 2.15.0 (10/26/2023)#

Table of contents

What’s New#

This release adds support for PyTorch 2.0 (Beta), increases performance for both training and inference workloads, adding ability to train models like Llama-2-70B using neuronx-distributed. With this release, we are also adding pipeline parallelism support for neuronx-distributed enabling full 3D parallelism support to easily scale training to large model sizes. Neuron 2.15 also introduces support for training resnet50, milesial/Pytorch-UNet and deepmind/vision-perceiver-conv models using torch-neuronx, as well as new sample code for flan-t5-xl model inference using neuronx-distributed, in addition to other performance optimizations, minor enhancements and bug fixes.

What’s New Details Instances
Neuron Distributed (neuronx-distributed) for Training Pipeline parallelism support. See Distributed Strategies APIs , Developer guide for Pipeline Parallelism and Pipeline Parallelism Overview Llama-2-70B model training script (sample script) (tutorial) Mixed precision support. See Developer guide for Pipeline Parallelism Support serialized checkpoint saving and loading using save_xser and load_xser parameters. See Distributed Strategies APIs See more at NxD Core Release Notes (neuronx-distributed) Trn1/Trn1n
Neuron Distributed (neuronx-distributed) for Inference flan-t5-xl model inference script (tutorial) See more at NxD Core Release Notes (neuronx-distributed) and Distributed Strategies APIs Inf2,Trn1/Trn1n
Transformers Neuron (transformers-neuronx) for Inference Serialization support for Llama, Llama-2, GPT2 and BLOOM models . See developer guide and tutorial See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
PyTorch Neuron (torch-neuronx) Introducing PyTorch 2.0 Beta support. See introduce-pytorch-2-0 . See llama-2-7b training , bert training and t5-3b inference samples. Scripts for training resnet50[Beta] ,milesial/Pytorch-UNet[Beta] and deepmind/vision-perceiver-conv[Beta] models. Trn1/Trn1n,Inf2
AWS Neuron Reference for Nemo Megatron library (neuronx-nemo-megatron) Llama-2-70B model training sample using pipeline parallelism and tensor parallelism ( tutorial ) GPT-NeoX-20B model training using pipeline parallelism and tensor parallelism See more at AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes and neuronx-nemo-megatron github repo Trn1/Trn1n
Neuron Compiler (neuronx-cc) New llm-training option argument to --distribution_strategy compiler option for optimizations related to distributed training. See more at Neuron Compiler CLI Reference Guide (neuronx-cc) See more at Neuron Compiler (neuronx-cc) release notes Inf2/Trn1/Trn1n
Neuron Tools alltoall Collective Communication operation for intra node(with in the instance), previously released in Neuron Collectives v2.15.13, was added as a testable operation in nccom-test. See NCCOM-TEST User Guide See more at Neuron System Tools Inf1/Inf2/Trn1/Trn1n
Documentation Updates New App Note and Developer Guide about Activation memory reduction using sequence parallelism and activation recomputation in neuronx-distributed Added a new Model Samples and Tutorials summary page. See Model samples and tutorials Added Neuron SDK Classification guide. See Neuron Software Classification See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.14.1 (09/26/2023)#

This is a patch release that fixes compiler issues in certain configurations of Llama and Llama-2 model inference using transformers-neuronx.

Neuron 2.14.0 (09/15/2023)#

Table of contents

What’s New#

This release introduces support for Llama-2-7B model training and T5-3B model inference using neuronx-distributed. It also adds support for Llama-2-13B model training using neuronx-nemo-megatron. Neuron 2.14 also adds support for Stable Diffusion XL(Refiner and Base) model inference using torch-neuronx . This release also introduces other new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

Note

This release deprecates --model-type=transformer-inference compiler flag. Users are highly encouraged to migrate to the --model-type=transformer compiler flag.

What’s New Details Instances
AWS Neuron Reference for Nemo Megatron library (neuronx-nemo-megatron) Llama-2-13B model training support ( tutorial ) ZeRO-1 Optimizer support that works with tensor parallelism and pipeline parallelism See more at AWS Neuron Reference for Nemo Megatron(neuronx-nemo-megatron) Release Notes and neuronx-nemo-megatron github repo Trn1/Trn1n
Neuron Distributed (neuronx-distributed) for Training pad_model API to pad attention heads that do not divide by the number of NeuronCores, this will allow users to use any supported tensor-parallel degree. See Distributed Strategies APIs Llama-2-7B model training support (sample script) (tutorial) See more at NxD Core Release Notes (neuronx-distributed) and Distributed Strategies APIs Trn1/Trn1n
Neuron Distributed (neuronx-distributed) for Inference T5-3B model inference support (tutorial) pad_model API to pad attention heads that do not divide by the number of NeuronCores, this will allow users to use any supported tensor-parallel degree. See Distributed Strategies APIs See more at NxD Core Release Notes (neuronx-distributed) and Distributed Strategies APIs Inf2,Trn1/Trn1n
Transformers Neuron (transformers-neuronx) for Inference Introducing --model-type=transformer compiler flag that deprecates --model-type=transformer-inference compiler flag. See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
PyTorch Neuron (torch-neuronx) Performance optimizations in torch_neuronx.analyze API. See PyTorch NeuronX Analyze API for Inference Stable Diffusion XL(Refiner and Base) model inference support ( sample script) Trn1/Trn1n,Inf2
Neuron Compiler (neuronx-cc) New --optlevel``(or ``-O) compiler option that enables different optimizations with tradeoff between faster model compile time and faster model execution. See more at Neuron Compiler CLI Reference Guide (neuronx-cc) See more at Neuron Compiler (neuronx-cc) release notes Inf2/Trn1/Trn1n
Neuron Tools Neuron SysFS support for showing connected devices on trn1.32xl, inf2.24xl and inf2.48xl instances. See Neuron Sysfs User Guide See more at Neuron System Tools Inf1/Inf2/Trn1/Trn1n
Documentation Updates Neuron Calculator now supports multiple model configurations for Tensor Parallel Degree computation. See Neuron Calculator Announcement to deprecate --model-type=transformer-inference flag. See Announcing deprecation for --model-type=transformer-inference compiler flag See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.13.2 (09/01/2023)#

This is a patch release that fixes issues in Kubernetes (K8) deployments related to Neuron Device Plugin crashes and other pod scheduling issues. This release also adds support for zero-based Neuron Device indexing in K8 deployments, see the Neuron K8 release notes for more details on the specific bug fixes.

Updating to latest Neuron Kubernetes components and Neuron Driver is highly encouraged for customers using Kubernetes.

Please follow these instructions in setup guide to upgrade to latest Neuron release.

Neuron 2.13.1 (08/29/2023)#

This release adds support for Llama 2 model training (tutorial) using neuronx-nemo-megatron library, and adds support for Llama 2 model inference using transformers-neuronx library (tutorial) .

Please follow these instructions in setup guide to upgrade to latest Neuron release.

Note

Please install transformers-neuronx from https://pip.repos.neuron.amazonaws.com to get latest features and improvements.

This release does not support LLama 2 model with Grouped-Query Attention

Neuron 2.13.0 (08/28/2023)#

Table of contents

What’s New#

This release introduces support for GPT-NeoX 20B model training in neuronx-distributed including Zero-1 optimizer capability. It also adds support for Stable Diffusion XL and CLIP models inference in torch-neuronx. Neuron 2.13 also introduces AWS Neuron Reference for Nemo Megatron library supporting distributed training of LLMs like GPT-3 175B. This release also introduces other new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

What’s New Details Instances
AWS Neuron Reference for Nemo Megatron library Modified versions of the open-source packages NeMo and Apex that have been adapted for use with AWS Neuron and AWS EC2 Trn1 instances. GPT-3 model training support ( tutorial ) See more at neuronx-nemo-megatron github repo Trn1/Trn1n
Transformers Neuron (transformers-neuronx) for Inference Latency optimizations for Llama and GPT-2 models inference. Neuron Persistent Cache support (developer guide) See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
Neuron Distributed (neuronx-distributed) for Training Now Stable, removed beta support ZeRO-1 Optimizer support with tensor parallel. (tutorial) Sequence Parallel support. (api guide) GPT-NeoX model training support. (sample script) (tutorial) See more at NxD Core Release Notes (neuronx-distributed) and Distributed Strategies APIs Trn1/Trn1n
Neuron Distributed (neuronx-distributed) for Inference KV Cache Support for LLM Inference (release notes) Inf2,Trn1/Trn1n
PyTorch Neuron (torch-neuronx) Seedable dropout enabled by default for training KV Cache inference support ( tutorial ) camembert-base training script. (sample script) New models inference support that include Stable Diffusion XL , CLIP (clip-vit-base-patch32 , clip-vit-large-patch14 ) , Vision Perceiver , Language Perceiver and T5 Trn1/Trn1n,Inf2
Neuron Tools New data types support for Neuron Collective Communication Test Utility (NCCOM-TEST) –check option: fp16, bf16, (u)int8, (u)int16, and (u)int32 Neuron SysFS support for FLOP count(flop_count) and connected Neuron Device ids (connected_devices). See Neuron Sysfs User Guide See more at Neuron System Tools Inf1/Inf2/Trn1/Trn1n
Neuron Runtime Runtime version and Capture Time support to NTFF Async DMA copies support to improve Neuron Device copy times for all instance types Logging and error messages improvements for Collectives timeouts and when loading NEFFs. See more at Neuron Runtime Release Notes Inf1, Inf2, Trn1/Trn1n
End of Support Announcements and Documentation Updates Announcing End of support for AWS Neuron reference for Megatron-LM starting Neuron 2.13. See more at AWS Neuron reference for Megatron-LM no longer supported Announcing end of support for torch-neuron version 1.9 starting Neuron 2.14. See more at Announcing end of support for torch-neuron version 1.9 Added TensorFlow 2.x (tensorflow-neuronx) analyze_model API section. See more at TensorFlow 2.x (tensorflow-neuron) analyze_model API Upgraded numpy version to 1.21.6 in various training scripts for Text Classification Updated bert-japanese training Script to use multilingual-sentiments dataset. See `hf-bert-jp <aws-neuron/aws-neuron-samples> `_ See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Known Issues and Limitations See 2.13.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

2.13.0 Known Issues and Limitations#

Neuron 2.12.2 (08/19/2023)#

Patch release to fix a jemalloc conflict for all Neuron customers that use Ubuntu 22. The previous releases shipped with a dependency on jemalloc that may lead to compilation failures in Ubuntu 22 only. Please follow these instructions in setup guide to upgrade to latest Neuron release.

Neuron 2.12.1 (08/09/2023)#

Patch release to improve reliability of Neuron Runtime when running applications on memory constrained instances. The Neuron Runtime has reduced the contiguous memory requirement for initializing the Neuron Cores associated with applications. This reduction allows bringup when only small amounts of contiguous memory remain on an instance. Please upgrade to latest Neuron release to use the latest Neuron Runtime.

Neuron 2.12.0 (07/19/2023)#

Table of contents

What’s New#

This release introduces ZeRO-1 optimizer for model training in torch-neuronx , introduces beta support for GPT-NeoX, BLOOM , Llama and Llama 2(coming soon) models in transformers-neuronx. This release also adds support for model inference serving on Triton Inference Server for Inf2 & Trn1 instances, lazy_load API and async_load API for model loading in torch-neuronx, as well as other new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

What’s New Details Instances
ZeRO-1 optimizer for model training in torch-neuronx Support of ZeRO-Stage-1 optimizer ( ZeroRedundancyOptimizer() API) for training models using torch-neuronx See tutorial at ZeRO-1 Tutorial Inf2, Trn1/Trn1n
Support for new models and Enhancements in transformers-neuronx [Beta] Support for inference of GPT-NeoX, BLOOM and Llama models. [Beta] Support for Llama 2 coming soon. Please monitor the transformers-neuronx repository for updates. Removed constraints on tp_degree in tensor-parallel configurations for GPT2, OPT, and BLOOM . See more at Transformers Neuron (transformers-neuronx) release notes Added multi-query / multi-group attention support for GPT2. See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
Support for Inf2 and Trn1 instances on Triton Inference Server Support for Model Inference serving on Triton for Inf2 and Trn1 instances. See more at Triton Server Python Backend See tutorial at Triton on SageMaker - Deploying on Inf2 Inf2, Trn1
Support for new computer vision models Performance optimizations in Stable Diffusion 2.1 model script and added [beta] support for Stable Diffusion 1.5 models. [Beta] Script for training CLIP model for Image Classification. [Beta] Script for inference of Multimodal perceiver model Please check aws-neuron-samples repository Inf2, Trn1/Trn1n
New Features in neuronx-distributed for training Added parallel cross entropy loss function. See more at tp_api_guide Trn1/Trn1n
lazy_load and async_load API for model loading in inference and performance enhancements in torch-neuronx Added lazy_load and async_load API to accelerate model loading for Inference. See more at PyTorch NeuronX Lazy and Asynchronous Loading API Optimize DataParallel API to load onto multiple cores simultaneously when device IDs specified are consecutive. See more at PyTorch Neuron (torch-neuronx) release notes Inf2, Trn1/Trn1n
[Beta] Asynchronous Execution support and Enhancements in Neuron Runtime Added beta asynchronous execution feature which can reduce latency by roughly 12% for training workloads. See more at NeuronX Runtime Configuration AllReduce with All-to-all communication pattern enabled for 16 ranks on TRN1/TRN1N within the instance (intranode) See more at Neuron Runtime Release Notes Inf1, Inf2, Trn1/Trn1n
Support for distribution_strategy compiler option in neuronx-cc Support for optional --distribution_strategy compiler option to enable compiler specific optimizations based on distribution strategy used. See more at Neuron Compiler CLI Reference Guide (neuronx-cc) Inf2, Trn1/Trn1n
New Micro Benchmarking Performance User Guide and Documentation Updates Added best practices user guide for benchmarking performance of Neuron devices. See more at Benchmarking Guide and Helper scripts Announcing end of support for Ubuntu 18. See more at Announcing end of support for Ubuntu 18 Removed support for Distributed Data Parallel(DDP) Tutorial. Improved sidebar navigation in Documentation. See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Known Issues and Limitations See 2.12.0 Known Issues and Limitations Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

2.12.0 Known Issues and Limitations#

Known Issues in Ubuntu 22 Support#

Known issues in certain resnet models on Ubuntu 20#

Neuron 2.11.0 (06/14/2023)#

Table of contents

What’s New#

This release introduces Neuron Distributed, a new python library to simplify training and inference of large models, improving usability with features like S3 model caching, standalone profiler tool, support for Ubuntu22, as well as other new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

What’s New Details Instances
New Features and Performance Enhancements in transformers-neuronx Support for int8 inference. See example at int8 weight storage support Improved prompt context encoding performance. See more at Transformers NeuronX (transformers-neuronx) Developer Guide Improved collective communications performance for Tensor Parallel inference on Inf2 and Trn1. See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
Neuron Profiler Tool Profiling and visualization of model execution on Trainium and Inferentia devices now supported as a stand-alone tool. See more at Neuron Profile User Guide Inf1, Inf2, Trn1/Trn1n
Neuron Compilation Cache through S3 Support for sharing compiled models across Inf2 and Trn1 nodes through S3 See more at PyTorch NeuronX neuron_parallel_compile CLI Inf2, Trn1/Trn1n
New script to scan a model for supported/unsupported operators Script to scan a model for supported/unsupported operators before training, scan output includes supported and unsupported operators at both XLA operators and PyTorch operators level. See a sample tutorial at Analyze for Training Tutorial Inf2, Trn1/Trn1n
Neuron Distributed Library [Beta] New Python Library based on PyTorch enabling distributed training and inference of large models. Initial support for tensor-parallelism. See more at NxD Core Inf2, Trn1/Trn1n
Neuron Calculator and Documentation Updates New Neuron Calculator Documentation section to help determine number of Neuron Cores needed for LLM Inference. Added App Note Generative LLM inference with Neuron See more at Neuron Documentation Release Notes Inf1, Inf2, Trn1/Trn1n
Enhancements to Neuron SysFS Support for detailed breakdown of memory usage across the NeuronCores See more at Neuron Sysfs User Guide Inf1, Inf2, Trn1/Trn1n
Support for Ubuntu 22 See more at Setup Guide for setup instructions on Ubuntu22 Inf1, Inf2, Trn1/Trn1n
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.10.0 (05/01/2023)#

Table of contents

What’s New#

This release introduces new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

What’s New Details Instances
Initial support for computer vision models inference Added Stable Diffusion 2.1 model script for Text to Image Generation Added VGG model script for Image Classification Task Added UNet model script for Image Segmentation Task Please check aws-neuron-samples repository Inf2, Trn1/Trn1n
Profiling support in PyTorch Neuron(torch-neuronx) for Inference with TensorBoard See more at Profiling PyTorch Neuron (torch-neuronx) with TensorBoard Inf2, Trn1/Trn1n
New Features and Performance Enhancements in transformers-neuronx Support for the HuggingFace generate function. Model Serialization support for GPT2 models. (including model saving, loading, and weight swapping) Improved prompt context encoding performance. See Transformers NeuronX (transformers-neuronx) for examples and usage See more at Transformers Neuron (transformers-neuronx) release notes Inf2, Trn1/Trn1n
Support models larger than 2GB in TensorFlow 2.x Neuron (tensorflow-neuronx) See Special Flags for details. (tensorflow-neuronx) Trn1/Trn1n, Inf2
Support models larger than 2GB in TensorFlow 2.x Neuron (tensorflow-neuron) See Special Flags for details. (tensorflow-neuron) Inf1
Performance Enhancements in PyTorch C++ Custom Operators (Beta) Support for using multiple GPSIMD Cores in Custom C++ Operators See Custom Operators API Reference Guide [Beta] Trn1/Trn1n
Weight Deduplication Feature (Inf1) Support for Sharing weights when loading multiple instance versions of the same model on different NeuronCores. See more at NeuronX Runtime Configuration Inf1
nccom-test - Collective Communication Benchmarking Tool Supports enabling benchmarking sweeps on various Neuron Collective Communication operations. See NCCOM-TEST User Guide for more details. Trn1/Trn1n , Inf2
Announcing end of support for tensorflow-neuron 2.7 & mxnet-neuron 1.5 versions See Announcing end of support for tensorflow-neuron versions 2.7 See Announcing end of support for mxnet-neuron versions 1.5 Inf1
Minor enhancements and bug fixes. See components-rn Trn1/Trn1n , Inf2, Inf1
Release Artifacts see Release Content Trn1/Trn1n , Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.9.1 (04/19/2023)#

Minor patch release to add support for deserialized torchscript model compilation and support for multi-node training in EKS. Fixes included in this release are critical to enable training and deploying models with Amazon Sagemaker or Amazon EKS.

Neuron 2.9.0 (03/28/2023)#

Table of contents

What’s New#

This release adds support for EC2 Trn1n instances, introduces new features, performance optimizations, minor enhancements and bug fixes. This release introduces the following:

What’s New Details Instances
Support for EC2 Trn1n instances Updated Neuron Runtime for Trn1n instances Overall documentation update to include Trn1n instances Trn1n
New Analyze API in PyTorch Neuron (torch-neuronx) A new API that return list of supported and unsupported PyTorch operators for a model. See PyTorch NeuronX Analyze API for Inference Trn1, Inf2
Support models that are larger than 2GB in PyTorch Neuron (torch-neuron) on Inf1 See separate_weights flag to torch_neuron.trace() to support models that are larger than 2GB Inf1
Performance Improvements Up to 10% higher throughput when training GPT3 6.7B model on multi-node Trn1
Dynamic Batching support in TensorFlow 2.x Neuron (tensorflow-neuronx) See Special Flags for details. Trn1, Inf2
NeuronPerf support for Trn1/Inf2 instances Added Trn1/Inf2 support for PyTorch Neuron (torch-neuronx) and TensorFlow 2.x Neuron (tensorflow-neuronx) Trn1, Inf2
Hierarchical All-Reduce and Reduce-Scatter collective communication Added support for hierarchical All-Reduce and Reduce-Scatter in Neuron Runtime to enable better scalability of distributed workloads . Trn1, Inf2
New Tutorials added Added tutorial to fine-tune T5 model Added tutorial to demonstrate use of Libtorch with PyTorch Neuron (torch-neuronx) for inference [html] Trn1, Inf2
Minor enhancements and bug fixes. See components-rn Trn1, Inf2, Inf1
Release included packages see Release Content Trn1, Inf2, Inf1

For more detailed release notes of the new features and resolved issues, see components-rn.

To learn about the model architectures currently supported on Inf1, Inf2, Trn1 and Trn1n instances, please see model_architecture_fit.

Neuron 2.8.0 (02/24/2023)#

Table of contents

What’s New#

This release adds support for EC2 Inf2 instances, introduces initial inference support with TensorFlow 2.x Neuron (tensorflow-neuronx) on Trn1 and Inf2, and introduces minor enhancements and bug fixes.

This release introduces the following:

What’s New Details
Support for EC2 Inf2 instances Inference support for Inf2 instances in PyTorch Neuron (torch-neuronx) Inference support for Inf2 instances in TensorFlow 2.x Neuron (tensorflow-neuronx) Overall documentation update to include Inf2 instances
TensorFlow 2.x Neuron (tensorflow-neuronx) support This releases introduces initial inference support with TensorFlow 2.x Neuron (tensorflow-neuronx) on Trn1 and Inf2
New Neuron GitHub samples New sample scripts for deploying LLM models with transformer-neuronx under aws-neuron-samples GitHub repository. New sample scripts for deploying models with torch-neuronx under aws-neuron-samples repository GitHub repository.
Minor enhancements and bug fixes. See components-rn
Release included packages see Release Content

For more detailed release notes of the new features and resolved issues, see components-rn.

Neuron 2.7.0 (02/08/2023)#

Table of contents

What’s New#

This release introduces new capabilities and libraries, as well as features and tools that improves usability. This release introduces the following:

What’s New Details
PyTorch 1.13 Support of PyTorch 1.13 version for PyTorch Neuron (torch-neuronx). For resources see PyTorch Neuron
PyTorch DistributedDataParallel (DDP) API Support of PyTorch DistributedDataParallel (DDP) API in PyTorch Neuron (torch-neuronx). For resources how to use PyTorch DDP API with Neuron, please check neuronx-ddp-tutorial.
Inference support in torch-neuronx For more details please visit pytorch-neuronx-main` page. You can also try Neuron Inference samples aws-neuron/aws-neuron-samples in the aws-neuron-samples GitHub repo.
Neuron Custom C++ Operators[Beta] Initial support for Neuron Custom C++ Operators [Beta] , with Neuron Custom C++ Operators (“CustomOps”) you can now write CustomOps that run on NeuronCore-v2 chips. For more resources please check Neuron Custom C++ Operators [Beta] section.
transformers-neuronx [Beta] transformers-neuronx is a new library enabling LLM model inference. It contains models that are checkpoint-compatible with HuggingFace Transformers, and currently supports Transformer Decoder models like GPT2, GPT-J and OPT. Please check aws-neuron-samples repository
Neuron sysfs filesystem Neuron sysfs filesystem exposes Neuron Devices under /sys/devices/virtual/neuron_device providing visibility to Neuron Driver and Runtime at the system level. By performing several simple CLIs such as reading or writing to a sysfs file, you can get information such as Neuron Runtime status, memory usage, Driver info etc. For resources about Neuron sysfs filesystem visit Neuron Sysfs User Guide.
TFLOPS support in Neuron System Tools Neuron System Tools now also report model actual TFLOPs rate in both neuron-monitor and neuron-top. More details can be found in the Neuron Tools documentation.
New sample scripts for training This release adds multiple new sample scripts for training models with torch-neuronx, Please check aws-neuron-samples repository
New sample scripts for inference This release adds multiple new sample scripts for deploying models with torch-neuronx, Please check aws-neuron-samples repository
Neuron GitHub samples repository for Amazon EKS A new AWS Neuron GitHub samples repository for Amazon EKS, Please check aws-neuron-samples repository

For more detailed release notes of the new features and resolved issues, see components-rn.

Neuron 2.6.0 (12/12/2022)#

This release introduces the support of PyTorch 1.12 version, and introduces PyTorch Neuron (torch-neuronx) profiling through Neuron Plugin for TensorBoard. Pytorch Neuron (torch-neuronx) users can now profile their models through the following TensorBoard views:

This release introduces the support of LAMB optimizer for FP32 mode, and adds support for capturing snapshots of inputs, outputs and graph HLO for debugging.

In addition, this release introduces the support of new operators and resolves issues that improve stability for Trn1 customers.

For more detailed release notes of the new features and resolved issues, see components-rn.

Neuron 2.5.0 (11/23/2022)#

Neuron 2.5.0 is a major release which introduces new features and resolves issues that improve stability for Inf1 customers.

Component New in this release
PyTorch Neuron (torch-neuron) PyTorch 1.12 support Python 3.8 support LSTM support on Inf1 R-CNN support on Inf1 Support for new API for core placement Support for improved logging Improved torch_neuron.trace() performance when using large graphs Reduced host memory usage of loaded models in libtorchneuron.so Additional operators support
TensorFlow Neuron (tensorflow-neuron) tf-neuron-auto-multicore tool to enable automatic data parallel on multiple NeuronCores. Beta support for tracing models larger than 2GB using extract-weights flag (TF2.x only), see TensorFlow 2.x (tensorflow-neuron) Tracing API tfn.auto_multicore Python API to enable automatic data parallel (TF2.x only)

This Neuron release is the last release that will include torch-neuron versions 1.7 and 1.8, and that will include tensorflow-neuron versions 2.5 and 2.6.

In addition, this release introduces changes to the Neuron packaging and installation instructions for Inf1 customers, see Introducing Neuron packaging and installation changes for Inf1 customers for more information.

For more detailed release notes of the new features and resolved issues, see components-rn.

Neuron 2.4.0 (10/27/2022)#

This release introduces new features and resolves issues that improve stability. The release introduces “memory utilization breakdown” feature in both Neuron Monitor and Neuron Top system tools. The release introduces support for “NeuronCore Based Sheduling” capability to the Neuron Kubernetes Scheduler and introduces new operators support in Neuron Compiler and PyTorch Neuron. This release introduces also additional eight (8) samples of models’ fine tuning using PyTorch Neuron. The new samples can be found in the AWS Neuron Samples GitHub repository.

Neuron 2.3.0 (10/10/2022)#

This Neuron 2.3.0 release extends Neuron 1.x and adds support for the new AWS Trainium powered Amazon EC2 Trn1 instances. With this release, you can now run deep learning training workloads on Trn1 instances to save training costs by up to 50% over equivalent GPU-based EC2 instances, while getting the highest training performance in AWS cloud for popular NLP models.

The following workloads were tested in this release:

This document is relevant for: Inf1, Inf2, Trn1, Trn2