Maintained with ☕️ by
IcePanel logo

Announcing latency-optimized inference for Amazon Nova Pro foundation model in Amazon Bedrock

Share

Services

Amazon Nova Pro foundation model now supports latency-optimized inference in preview on Amazon Bedrock, enabling faster response times and improved responsiveness for generative AI applications. Latency-optimized inference speeds up response times for latency-sensitive applications, improving the end-user experience and giving developers more flexibility to optimize performance for their use case. Accessing these capabilities requires no additional setup or model fine-tuning, allowing for immediate enhancement of existing applications with faster response times. Latency optimized inference for Amazon Nova Pro is available via [cross-region inference](https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference.html) in US West (Oregon), US East (Virginia), and US East (Ohio) regions. Learn more about Amazon Nova foundation models at the [AWS News Blog](https://aws.amazon.com/blogs/aws/introducing-amazon-nova-frontier-intelligence-and-industry-leading-price-performance), the [Amazon Nova product page](https://aws.amazon.com/nova/), or the [Amazon Nova user guide](https://docs.aws.amazon.com/nova/latest/userguide). Learn more about latency optimized inference on Bedrock in [documentation](https://docs.aws.amazon.com/bedrock/latest/userguide/latency-optimized-inference.html). You can get started with Amazon Nova foundation models in Amazon Bedrock from the [Amazon Bedrock console](https://console.aws.amazon.com/bedrock/).