AWS Inferentia
Categories:
AWS Inferentia is a machine learning inference chip designed to deliver high performance at low cost. AWS Inferentia will support the TensorFlow, Apache MXNet, and PyTorch deep learning frameworks, as well as models that use the ONNX format.
Description
AWS Inferentia is a machine learning inference chip designed to deliver high performance at low cost. AWS Inferentia will support the TensorFlow, Apache MXNet, and PyTorch deep learning frameworks, as well as models that use the ONNX format.
Version
unkown
Documentation
https://aws.amazon.com/machine-learning/inferentia
SLA
https://aws.amazon.com/compute/sla/
Data
User data is stored on Amazon servers under the associated AWS account and is protected under the AWS shared responsibility model as detailed here https://aws.amazon.com/compliance/shared-responsibility-model/