Announcing latency-optimized inference for Amazon Nova Pro foundation model in Amazon Bedrock
Share
Services
Amazon Nova Pro foundation model now supports latency-optimized inference in preview on Amazon Bedrock, enabling faster response times and improved responsiveness for generative AI applications. Latency-optimized inference speeds up response times for latency-sensitive applications, improving the end-user experience and giving developers more flexibility to optimize performance for their use case. Accessing these capabilities requires no additional setup or model fine-tuning, allowing for immediate enhancement of existing applications with faster response times.
Latency optimized inference for Amazon Nova Pro is available via [cross-region inference](https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference.html) in US West (Oregon), US East (Virginia), and US East (Ohio) regions. Learn more about Amazon Nova foundation models at the [AWS News Blog](https://aws.amazon.com/blogs/aws/introducing-amazon-nova-frontier-intelligence-and-industry-leading-price-performance), the [Amazon Nova product page](https://aws.amazon.com/nova/), or the [Amazon Nova user guide](https://docs.aws.amazon.com/nova/latest/userguide). Learn more about latency optimized inference on Bedrock in [documentation](https://docs.aws.amazon.com/bedrock/latest/userguide/latency-optimized-inference.html). You can get started with Amazon Nova foundation models in Amazon Bedrock from the [Amazon Bedrock console](https://console.aws.amazon.com/bedrock/).
What else is happening at Amazon Web Services?
Read update
Services
Share
Read update
Services
Share
AWS Firewall Manager is now available in the AWS Asia Pacific (Thailand) and AWS Mexico (Central) regions
about 22 hours ago
Services
Share
Amazon CloudWatch RUM now supports monitoring multiple domains with a single App Monitor
about 22 hours ago
Services
Share
Read update
Services
Share