Skip to main content

Amazon unveils its new family of Nova foundational models

pasta city
AWS

Amazon CEO Andy Jassy took to the stage at the company’s re:Invent conference on Tuesday to show off six new text, image, and video generation models that it calls Amazon Nova.

Recommended Videos

This new family of multimodal generative AIs includes Nova Micro, a text-only model built for low-cost, low-latency responses; Nova Lite, a low-cost multimodal model for processing image, video, and text inputs; and Nova Pro, its general purpose multimodal model that combines “accuracy, speed, and cost for a wide range of tasks,” per the company’s announcement post. Nova Premier is Amazon’s “most capable … multimodal models for complex reasoning tasks,” while Nova Canvas is a dedicated text-to-image engine and Nova Reel is purpose-built to generate video.

The text-based models have been optimized on 15 different languages. Micro offers a 128,000-token context window while both Lite and Pro can handle up to 300,000 tokens (around 225,000 words or 30 minutes of video). The company plans to expand the context windows of its larger models up to 2 million tokens by early next year. 

Canvas enables users to generate and edit images using natural language prompts. Reels, which will compete with the likes of Gen-3 Alpha, Kling, and Dall-E 3, can generate clips up to six seconds in length from both text prompts and reference images. The video generator also offers camera motion control including pans and zooms.

Pasta City, created with Amazon Nova Reel by Amazon Ads

“We’ve continued to work on our own frontier models,” Jassy told the assembled crowd, “and those frontier models have made a tremendous amount of progress over the last four to five months. And we figured, if we were finding value out of them, you would probably find value out of them.”

Jassy also says that these models are both among the least expensive to operate and fastest in their class, though the company has yet to post benchmark data supporting those claims. “We’ve optimized these models to work with proprietary systems and APIs, so that you can do multiple orchestrated automatic steps — agent behavior — much more easily with these models,” he said. “So I think these are very compelling.”

The Micro, Lite, and Pro models (as well as Canvas and Reels) are all currently available to AWS customers. Premiere is set to arrive in Q1 2025.

Andrew Tarantola
Former Digital Trends Contributor
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
Anthropic’s new Claude model offers both real-time and long-pondered responses
Claude code homescreen

OpenAI's o3 and DeepSeek's R1 models have some new competition. Anthropic announced Monday the release of its new "hybrid reasoning" model, Claude 3.7 Sonnet.

Existing reasoning models like o3, R1, and Google’s Gemini 2.0 Flash Thinking are designed to break down complex problems into smaller tasks, then deduce and verify their answers before responding, a process that returns more accurate answers at the cost of higher compute usage and longer inference times. Claude 3.7 Sonnet, on the other hand is capable of providing either "near-instant responses or extended, step-by-step thinking that is made visible to the user," according to the company's announcement post.

Read more
Meta rolls out its AI chatbot to nearly a dozen Middle Eastern nations
Meta AI in the Middle East

Millions of Facebook, Instagram, WhatsApp, and Messenger users throughout the Middle East now enjoy access to Meta's self-named AI chatbot platform, the company announced on Monday. The chatbot is rolling out to users in Algeria, Egypt, Iraq, Jordan, Libya, Morocco, Saudi Arabia, Tunisia, United Arab Emirates, and Yemen.

"AI just got even more accessible than ever before, as we officially launched Meta AI in the Middle East and North Africa with Arabic capabilities," Meta wrote in its announcement blog post. At launch, these users will have access to only some of Meta AI's generative capabilities -- specifically, text and image generation, as well as image animation. The company plans to expand those offerings to include simultaneous dubbing for Reels, AI image editing, and the "Imagine Me" feature (which generates a user's portrait based on uploaded reference photos) in the near future.

Read more
Spotify now offers its listeners AI-narrated audiobooks
The Spotify Audiobooks on the desktop app showing Lord of the Rings.

In a move expected to dramatically increase the quantity of available audiobooks, Spotify announced on Thursday that will begin accepting titles narrated by vocal AIs from ElevenLabs.

"For authors looking for a cost-effective way to create high-quality audiobooks, digital voice narration by ElevenLabs is a great option," the company wrote in its announcement post, pointing out that AI narration has been one of its most requested features. "Authors can use the ElevenLabs platform to narrate their audiobooks in 29 languages, with complete control over voice and intonation."

Read more