AWS Inferentia

AWS Inferentia is a machine learning inference chip designed to deliver high performance at low cost. AWS Inferentia will support the TensorFlow, Apache MXNet, and PyTorch deep learning frameworks, as well as models that use the ONNX format.

Description

AWS Inferentia is a machine learning inference chip designed to deliver high performance at low cost. AWS Inferentia will support the TensorFlow, Apache MXNet, and PyTorch deep learning frameworks, as well as models that use the ONNX format.

Version

unkown

Documentation

https://aws.amazon.com/machine-learning/inferentia

SLA

https://aws.amazon.com/compute/sla/

Data

User data is stored on Amazon servers under the associated AWS account and is protected under the AWS shared responsibility model as detailed here https://aws.amazon.com/compliance/shared-responsibility-model/