Overview
The algorithm delivers a fully automated, end-to-end pipeline for training and deploying YOLOX small which is a high-accuracy, anchor-free object detection network optimized for Ambarella edge AI processors. Users supply a custom labeled dataset and configure key hyperparameters (image resolution, number of classes, epochs, batch size, learning rate, and target chipset), and the algorithm manages the complete training workflow on AWS GPU instances.
Upon training completion, the best checkpoint is automatically compiled into Ambarella CVFlow artifacts. The tools generate an ambapb checkpoint (compiled artifacts represented in ONNX format) for host evaluation and a cavalry binary to deploy on Ambarella CV-series chipsets. All compiled outputs are stored as SageMaker model artifacts, ready for deployment.
This SageMaker algorithm provides an end-to-end workflow to train and deploy a YOLOX small variant, a high-accuracy, anchor-free object detection network, for Ambarella edge AI processors. Users can bring their own labeled dataset, configure core hyperparameters (image resolution, number of classes, epochs, batch size, learning rate, and chipset), and run training on AWS GPU infrastructure using a managed pipeline. After training, the best checkpoint is automatically converted into Ambarella CVFlow deployment artifacts. The toolchain produces an AmbaPB checkpoint for host-side evaluation and a cavalry binary for deployment on Ambarella CV-series chipsets, including CV72, CV75, CV7, and N1-655. Generated outputs are packaged as SageMaker model artifacts for straightforward deployment and lifecycle management.
For inference, the bundled server loads CVFlow artifacts and exposes REST endpoints compatible with SageMaker real-time and asynchronous inference. Input images are preprocessed, executed through Ambarella runtime, and returned as JSON detections, enabling low-latency and scalable object detection workflows from model training to production deployment.
Highlights
- End-to-end YOLOX small variant in SageMaker: Train with configurable hyperparameters on custom datasets using AWS GPU compute. Ambarella CVFlow deployment ready: Automatically generate AmbaPB and cavalry artifacts for CV72, CV75, CV7, and N1-655. Production inference flexibility: Deploy the same model package to both real-time and asynchronous SageMaker endpoints.
Details
Introducing multi-product solutions
You can now purchase comprehensive solutions tailored to use cases and industries.
Features and programs
Financing for AWS Marketplace purchases
Pricing
Vendor refund policy
This product is offered at no charge, so no software usage fees are collected and refunds do not apply. Customers are responsible for any AWS infrastructure charges incurred while using the product.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker algorithm
An Amazon SageMaker algorithm is a machine learning model that requires your training data to make predictions. Use the included training algorithm to generate your unique model artifact. Then deploy the model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
This is the first release of the YOLOX small training and inference algorithm on Amazon SageMaker.
Features
End-to-end training pipeline for the YOLOX small object detection network on custom datasets using AWS GPU instances.
Configurable hyper-parameters including image resolution, number of classes, epochs, batch size, learning rate, warmup epochs, and target Ambarella chipset.
Automatic compilation of the best training checkpoint into Ambarella CVFlow deployment artifacts, including an AmbaPB checkpoint for host evaluation and a cavalry binary for on-device deployment.
Support for Ambarella CV-series chipsets: CV72, CV75, CV7, and N1-655.
Real-time and asynchronous SageMaker inference endpoints via a bundled REST inference server
Additional details
Inputs
- Summary
The inference endpoint accepts a single JPEG image per request submitted as raw binary data in the HTTP request body.
Content type: image/jpg Format: Raw JPEG binary data in the HTTP request body Color space: RGB, 3-channel Spatial resolution: Any resolution. The server automatically resizes and letterbox-pads the image to the model's configured input size. Preprocessing: Handled on the server-side.
- Limitations for input type
- File format: JPEG only (image/jpg). Other image formats (PNG, BMP, TIFF, etc.) are not supported. Requests with unsupported content types will return HTTP 415.
Support
Vendor support
Please email to inquiries@ambarella.com for support related questions.
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products
