Amazon is rolling out a new Alexa+ feature on Fire TV that can take you to a specific moment in a given movie on Prime Video based on a natural language voice command. The company says that, when you describe a certain scene, quote or character action, Alexa+ can start playing that part of the film. The company previewed this feature at its Devices and Services event in September.
According to Amazon, you can say something like “Jump to the card scene in Love Actually" or “Jump to the Ozdust ballroom scene in Wicked with Glinda,” to quickly get to that moment. Alexa+ can apparently figure out which movie you're referring to if you don't say the title. So if you say, for instance, “Jump to the scene when John McClane says ‘come out to the coast, we’ll get together, have a few laughs,’” Prime Video will start playing that bit in Die Hard where McClane is in an air duct.
To make this work, Alexa+ uses "visual understanding" and captions to determine what's happening in each scene so it can take you to the one you're looking for. It's all processed through the X-Ray feature in Prime Video. As with Alexa+, it's built on Amazon Bedrock and it harnesses large language models such as Amazon Nova and Anthropic Claude.
Alexa+ has indexed tens of thousands of scenes across thousands of movies on Prime Video so far, including many that you can purchase or rent. Amazon plans to expand this feature to more films and scenes, as well as TV shows, in the near future.
While this is pretty interesting from a tech perspective and how Amazon’s able to make it work, I’d be interested to know how many people actually end up using it. This isn’t how most people who genuinely love cinema watch movies — maybe just start at the beginning of a film and take it from there? Besides, if you really want to watch a specific scene, YouTube exists.
This article originally appeared on Engadget at https://www.engadget.com/ai/amazon-rolls-out-a-find-a-scene-alexa-feature-for-prime-video-150557530.html?src=rss
Amazon appears to have quietly removed its terrible AI-generated English dubs for several anime shows currently streaming on Prime Video, following widespread ridicule from viewers and industry professionals. AI dubs were recently added to Banana Fish, No Game, No Life and Vinland Saga, where they were labeled "AI beta" in the Languages section of the app.
As shows that previously only offered English subtitles, the option of a dub for those who prefer it could have been seen as a win for Amazon. But it quickly became clear that the dubs were really quite bad, completely devoid of any emotion or convincing intonation in dramatic moments. Particularly awful clips of the AI English dub for Banana Fish soon started circulating on social media, and the National Association of Voice Actors released a statement in which it branded the dubs "AI slop."
In his own statement, voice actor Daman Mills called the AI-generated dub for Banana Fish a "massive insult to us as performers." In a post on X, which at the time of writing has been liked 14,000 times, he said: "Voice Actors deserve the same level of respect as on camera performers. Anime already pays talent very little. Dub production costs shouldn’t make a dent in these companies’ pocket books. Using AI for a dub of a show that released nearly 8 YEARS AGO AND HAD NO RUSHED SCHEDULE just spits in our faces, has infuriated the consumer, and completely destroys the art."
Decision-makers at Amazon apparently noted the backlash, as the English dub options no longer show up, as acknowledged by Mills and others yesterday. An AI-generated Spanish dub for Vinland Saga appears to have survived the silent cull, but otherwise it’s back to Japanese language-only and subtitles for the other shows.
The company is clearly committed to introducing more AI to Prime Video — along with its various other services — despite this latest public shaming. It launched an "AI-aided" dubbing program for Prime Video earlier this year, piloting English and Latin American Spanish dubs in 12 licensed series and movies on Prime Video in March. Last month, it also introduced video recaps that summarize shows’ "most pertinent plot points" using generative AI. The feature is currently in beta for select English language Prime Original shows in the US.
This article originally appeared on Engadget at https://www.engadget.com/entertainment/amazon-halts-ai-anime-dub-beta-after-widespread-ridicule-141501051.html?src=rss
Superhuman, the AI-powered mail app, is heading in a more agentic direction with its latest update. Its "write with AI" feature, which you could previously activate when drafting an email, now works across your inbox, calendar, and the web. This means it can now pull in information from other emails or research a topic online. The AI will think for as long as it needs before responding to a prompt and will open its Ask AI tool if it needs clarification.
Ask AI now lives in a left sidebar when you’re on desktop, so it’s always accessible should you need to draft a note, ask a question or quickly schedule a meeting without digging around in your emails. You can also now check your Ask AI history on iOS and desktop for previous conversations. Write with AI is also now available on Android, which will soon gain the other new features too.
Superhuman
Superhuman was acquired by Grammarly earlier this year, with the latter recently rebranding so all of its AI apps now sit under the Superhuman umbrella. The mail service is seemingly primarily targeted at business rather than consumers, with its most advanced version of Write with AI and Ask AI being included in Business and Enterprise plans. The more basic standard version of Write with AI is rolled into the Starter plan for desktop and mobile.
Superhuman is promising further agentic updates in the near future.
This article originally appeared on Engadget at https://www.engadget.com/ai/superhuman-formerly-grammarly-has-some-ai-updates-for-its-superhuman-mail-app-140017716.html?src=rss
It's that time of year again, when all of our favorite streaming platforms start dropping personalized lists of what we've been consuming. Spotify Wrapped is perhaps the biggest of the bunch and it's available for perusal right now.
As always, users can access Wrapped to find their most listened-to genres, artists, songs, albums and podcasts from the past year. This information is shareable via social media if you want random bald eagle avatars to comment on your music taste, but there's a new interactive feature called Wrapped Party.
Spotify
This is a game of a sort. Spotify says it "turns your listening data into a live competition." Wrapped Party hands out awards for stuff like listening to smaller artists and obsession with a particular artist, in addition to total minutes streamed. Finally, friends can settle the age-old debate of "who listens to music more."
Spotify Wrapped is also about the platform itself, so we have plenty of little tidbits from the global user base. Bad Bunny was named the most streamed artist in the world, just ahead of his Super Bowl performance that internet bozos have turned into a controversy for some reason. This is the fourth time he's come out on top in the past five years. He also had the most popular album of the year.
Spotify
The global top song is something of a surprise, as it's not Bad Bunny or even Taylor Swift. It's the Lady Gaga and Bruno Mars duet "Die With a Smile." The top podcast is, as always, The Joe Rogan Experience. At least Spotify is getting what it paid for with Rogan.
This article originally appeared on Engadget at https://www.engadget.com/entertainment/music/spotify-wrapped-2025-is-here-and-now-its-a-competition-130052418.html?src=rss
Uber has made a big push to offer robotaxis as an option for its rideshare services in more markets this year. Starting today, the company is offering autonomous vehicles as an option for customers in Dallas. The move is in partnership with Avride.
At the start, the AVs providing rides will have a person in the front seat, but Uber plans to have fully driverless operation "in the future." The company will have a small fleet of Avride's Hyundai Ioniq 5 vehicles to start, but it plans to eventually have hundreds of these AVs working in Dallas. Riders can set their preferences to increase their chances of being paired with a robotaxi in the Uber app. If someone is assigned an AV for their ride, they will have the option to switch to a traditional rideshare driver.
Uber started a partnership with Avride in October 2024, but the rideshare company has cast a wide net for collaborators. It has also worked to bring robotaxis to markets with Waymo in Austin and Atlanta, with Lucid in the Bay Area, with WeRide in Abu Dhabi, and with Momenta in Europe.
This article originally appeared on Engadget at https://www.engadget.com/transportation/uber-is-launching-robotaxis-in-dallas-120000411.html?src=rss
It's time for the annual Rockefeller Christmas tree lighting! The Christmas in Rockefeller Center tree lighting special will air tonight, Dec. 3 from 8-10 PM ET — though coverage will start an hour prior, at 7 PM ET. The Voice and Happy's Place star Reba McEntire will host and perform at the Rockefeller Tree lighting, which will also feature performances from Halle Bailey, Michael Bublé, Kristin Chenoweth, Laufey, the Radio City Rockettes and more. Here’s how to tune into the 2025 Rockefeller Tree lighting.
When is the 2025 Rockefeller Tree lighting?
The Rockefeller Center Christmas tree will be lit on Wednesday, Dec. 3, 2025.
2025 Rockefeller Tree lighting time:
Coverage of the 2025 Rockefeller Tree lighting will start at 7 p.m. ET. The official Christmas in Rockefeller Center tree lighting special will air from 8-10 p.m ET.
2025 Rockefeller Tree lighting channel:
The Christmas in Rockefeller Center special will air on NBC and stream on Peacock.
How to watch the 2025 Rockefeller Tree lighting without cable:
Who is hosting the 2025 Rockefeller Tree lighting?
Reba McEntire will host NBC's annual holiday special, and perform throughout the evening.
2025 Rockefeller Tree Lighting performers:
Alongside Reba McEntire, the tree lighting ceremony special will feature performances from Marc Anthony, Halle Bailey, Michael Bublé, Kristin Chenoweth, Laufey, New Edition, Brad Paisley, Carly Pearce, Gwen Stefani and the Radio City Rockettes, who are celebrating their 100th anniversary this year.
Ways to watch the Rockefeller Tree lighting for free:
This article originally appeared on Engadget at https://www.engadget.com/entertainment/streaming/how-to-watch-the-christmas-in-rockefeller-center-tree-lighting-special-tonight-111504725.html?src=rss
One of the iPhone’s many accessibility features is something Apple calls "Magnifier," which uses the smartphone's cameras to magnify and identify objects in the world around you. For Global Accessibility Awareness Day in May this year, Apple brought Magnifier to the Mac, opening up even more places the assistive tool can be used, like classroom or work environments where you might already have a MacBook pulled out.
Magnifier requires macOS 26 Tahoe and can work with a built-in webcam, a connected third-party camera or an iPhone via Apple's Continuity feature. Provided your MacBook can run Apple’s latest software update, it’s a natural fit for zooming in on a whiteboard at the back of a large lecture hall or getting a closer look at documents on a desk in front of you. You can use the app to both capture an individual image you want to refer to later, or to analyze text in a live video feed. But where to begin? Here’s how to set up and use Magnifier on your Mac.
How to use Magnifier to identify and display text
A MacBook using Magnifier and a connected iPhone to identify and format text from a book.
Apple
Magnifier's most powerful feature uses the MacBook's machine learning capabilities to identify, display and format text that your camera captures. This works with text your camera can see in the room around you, and things it captures via macOS' Desk View feature. For example, to view documents on your desk:
Open Magnifier.
Click on the Camera section in Magnifier's menu bar and then select your Desk View camera from the dropdown menu.
Click on the Reader icon (a simple illustration of a document) near the top-right of your Magnifier window.
Click on the sidebar menu icon to access settings to format text.
Apple gives you options to change the color, font and background of text Magnifier identifies, among other customization options. If you'd prefer to capture faraway text, you can position a webcam or iPhone camera facing away from you and swap to it via the Camera section in Magnifier's menu bar.
You can also listen to any text Magnifier has identified by clicking on the Play button in the top-right corner of Magnifier's reader mode. Clicking the Pause button will pause playback, clicking the Skip Forward or Skip Backward buttons skip through lines of text, and if you want to adjust playback speed, you can click on the 1x button and pick a speed from the dropdown menu.
How to use Magnifier to zoom in on yourself
Magnifier can identify text, but it also works as a way to get a zoomed in view of your own face.
Ian Carlos Campbell for Engadget
By default, Magnifier uses your MacBook's built-in webcam, which means you'll see a view of yourself and whatever's behind you if you don't have another camera selected. This might not be usual for seeing faraway text, but it is handy if you're applying makeup, putting in contacts or doing anything else where you need a detailed view of your face.
In my tests, using Magnifier worked the best with my MacBook's built-in webcam or an iPhone. When I tried using a third-party webcam from Logitech, my live camera feed was noticeably laggy. Your mileage may vary, but if you experience any issues with your own webcam, it's worth trying your built-in webcam to see if that helps. You can swap between cameras and zoom in to your camera feed inside the Magnifier app:
Open Magnifier.
In the top menu bar, select Camera and then click on the camera you'd like to use in the dropdown menu.
Use the slider in the top center of the Magnifier window to zoom in on yourself.
You can see a live feed of your zoomed in view in Magnifier's main window. If you click on the Camera button in the bottom-left corner of the app, you can also snap a photo to review later. Any photos you capture will appear in Magnifier's left sidebar menu. Clicking on them lets you view them, zoom in on them and adjust their visual appearance (Brightness, Contrast and other visual settings) via the Image section in Magnifier's menu bar.
This article originally appeared on Engadget at https://www.engadget.com/computing/laptops/how-to-use-magnifier-on-a-macbook-to-zoom-in-on-faraway-text-080100677.html?src=rss
Artificial intelligence is showing up everywhere in Google's services these days, whether or not people want them and sometimes in places where they really don't make a lick of sense. The latest trial from Google appears to be giving articles the AI treatment in Google Discover. The Verge noticed that some articles were being displayed in Google Discover with AI-generated headlines different from the ones in the original posts. And to the surprise of absolutely no one, some of these headlines are misleading or flat-out wrong.
For instance, one rewritten headline claimed "Steam Machine price revealed," but the Ars Technicaarticle's actual headline was "Valve's Steam Machine looks like a console, but don’t expect it to be priced like one." No costs have been shared yet for the hardware, either in that post or elsewhere from Valve. In our own explorations, Engadget staff also found that Discover was providing original headlines accompanied by AI-generated summaries. In both cases, the content is tagged as "Generated with AI, which can make mistakes." But it sure would be nice if the company just didn't use AI at all in this situation and thus avoided the mistakes entirely.
The instances The Verge found were apparently "a small UI experiment for a subset of Discover users," Google rep Mallory Deleon told the publication. "We are testing a new design that changes the placement of existing headlines to make topic details easier to digest before they explore links from across the web." That sounds innocuous enough, but Google has a history of hostility towards online media its frequent role as middleman between publishers and readers. Web publishers have made multiple attempts over the years to get compensation from Google for displaying portions of their content, and in at least twoinstances, Google has responded by cutting out those sources from search results and later claiming that showing news doesn't do much for the bottom line of its ad business.
For those of you who do in fact want more AI in your Google Search experience, you're in luck. AI Mode, the chatbot that's already been called outright "theft" by the News Media Alliance, is getting an even more symbiotic integration into the mobile search platform. Google Search's Vice President of Product Robby Stein posted yesterday on X that the company is testing having AI Mode accessible on the same screen as an AI Overview rather than the two services existing in separate tabs.
This article originally appeared on Engadget at https://www.engadget.com/ai/google-discover-is-testing-ai-generated-headlines-and-they-arent-good-234700720.html?src=rss
You can tell a lot about a company by what they're willing to sue over. Take Instacart, which just filed a lawsuit against New York City. Its beef? The company doesn't like five new city laws, set to take effect in January. They would require Instacart to pay workers more and give customers a tipping option of at least 10 percent.
Reutersreports that Instacart's suit targets Local Law 124, which mandates that grocery delivery workers receive the same minimum pay as restaurant delivery workers. It also challenged Local Law 107, which mandates 10 percent or higher tipping options (or a place to enter one manually). The lawsuit also takes aim at other laws requiring extra recordkeeping and disclosures. The new rules are set to take effect on January 26.
As is typical of companies griping about regulations that hurt their bottom lines, Instacart framed the issue as a noble fight for what's right. "When a law threatens to harm shoppers, consumers, and local grocers — and especially when it does so unlawfully — we have a responsibility to act," the company proclaimed in a blog post. "This legal challenge is about standing up for fairness, for the independence that tens of thousands of New York grocery delivery workers rely on and for affordable access to groceries for the people who need it most."
Instacart's suit reportedly claims that Congress banned state and local governments from regulating prices on platforms such as its own. It also alleges that New York's state legislature "has long taken charge" of minimum pay, and that the US Constitution doesn't allow states and cities to discriminate against out-of-state companies.
The company warns that everyone will lose if it's forced to comply. Should the laws take effect, "Instacart will be forced to restructure its platform, restrict shoppers' access to work, disrupt relationships with consumers and retailers and suffer constitutional injuries with no adequate legal remedy," it claimed in the filing.
Instacart CEO Chris Rogers, elevated to the post in May, has an estimated net worth of at least $28.6 million. His predecessor, Fidji Simo, who chairs the board and is now with OpenAI, is reportedly worth around $72.7 million. If NYC’s minimum pay laws will be as catastrophic as Instacart claims, maybe they could chip in to help.
This article originally appeared on Engadget at https://www.engadget.com/big-tech/instacart-sues-new-york-city-over-minimum-pay-tipping-laws-220205207.html?src=rss
ExpressVPN, one of the best VPNs, is launching two brand-new features that sound confusingly like things it already does. Users on Android, Mac and iOS (but apparently not Windows, Linux or smart TVs) can now use Fastest Location to automatically pick the VPN server with the fastest download speed and lowest latency. Mac users are also getting an overhauled ExpressVPN app designed to work natively with MacOS.
If you've used ExpressVPN before, your first reaction probably went something like "Wait, didn't it already have a Fastest Location button and a Mac app?" You're not wrong, but there's still a meaningful difference with these new features. In the past, ExpressVPN didn't technically pick the fastestlocation, but the Smart Location, which picks the best available server using "metrics such as download speed, latency, and distance" (emphasis mine). Those are the same metrics as the new feature, but the such as makes me think there are, or were, other ingredients in the "smart location" algorithm.
My guess is that ExpressVPN is rebranding "smart" to "fastest" in response to customer complaints that "smart" was picking sub-optimal server locations. That's not a behavior I noticed when I last reviewed ExpressVPN — the smart location was always plenty fast for me — but I'm just one user. Only testing can show whether they actually changed the algorithm or just the name.
The new Mac app is a more straightforward upgrade. While ExpressVPN has always had a client for Mac, it's thus far been a port of an app originally developed for iPad. This makes its otherwise-excellent interface feel a bit like, well, a phone app you use on your desktop. In contrast, the new app was built using Project Catalyst, which lets Mac developers turn their iOS apps into desktop-native software. The new interface looks a lot richer, using the screen space a lot like Proton VPN does. And being more like Proton VPN is rarely a bad thing.
This article originally appeared on Engadget at https://www.engadget.com/cybersecurity/vpn/expressvpn-adds-a-fastest-location-button-and-launches-a-new-native-mac-app-205837728.html?src=rss