Google will use machine learning to try and tell if a user is under 18

Google will start testing a feature this year that uses machine learning to weed out children trying to access adult content on YouTube. The “machine learning-based age estimation model” will try to predict whether a user is under 18 and, if so, apply appropriate age filter settings to their account. The announcement came amid a flurry of Google child safety announcements as the US Senate considers a bill that would ban pre-teens from social media.

YouTube CEO Neal Mohan first mentioned the ML age restriction feature on Tuesday in his letter about the platform’s “bets” for the coming year. “We’ll use machine learning in 2025 to help us estimate a user’s age — distinguishing between younger viewers and adults — to help provide the best and most age-appropriate experiences and protections,” he wrote.

A YouTube spokesperson clarified to Engadget that the model will make its predictions using data like the types of info the person searches for, the categories of videos they watch and the age of their accounts. For example, if someone frequently searches for info about mortgage lending or taxes, that probably points to the person being over 18. Ditto for an account that’s over 20 years old.

If the ML model predicts that someone is underage, YouTube will apply its standard protections to deliver a more age-appropriate experience. Of course, the platform blocks explicit (and otherwise age-restricted) videos from under-18 users. Google’s underage restrictions also include its SafeSearch Filter, which (as its name suggests) blocks explicit content from search results.

YouTube told Engadget it will begin testing the filtration tool by the end of this year. The platform plans to roll it out globally in 2026. Although YouTube’s CEO first announced the feature, Google will test the feature in other areas, too. However, we don’t yet know where else the new ML feature will pop up. Google isn’t alone in trialing such a feature.

Google isn’t alone in trialing such a moderation feature. Last year, Meta said it would use an “adult classifier” tool to identify underage Instagram users posing as adults.

Google

Google also said on Wednesday that School Time, a feature previously only available as a smartwatch app on the Fitbit Ace LTE and Galaxy Watch for Kids, will roll out to Android phones and tablets. Part of Google’s Family Link parental controls app, School Time will let parents determine what phone features and apps their children can use during school hours. Parents can choose which apps remain active (like, say, learning-friendly apps) while allowing messages and calls from certain contacts. The idea is to minimize screen time and help kids focus on their work while still green-lighting emergency contacts.

Along similar lines, the Android Family Link app will let parents approve or deny contacts to add to their children’s devices. (This is another feature making its way over from the Galaxy Watch for Kids.) Parents can then limit calls and texts to only approved contacts. However, that feature isn’t quite here yet: Google says it will roll out “in the coming months.”

Finally, this spring, parents using Google parental control features can add tap-to-pay to their children’s Android phones. (Google said this was coming last year.) Parents will be able to approve a payment card, add or remove further cards and view the child’s transactions. Google says the feature will also work for things like gift cards and concert tickets.

This article originally appeared on Engadget at https://www.engadget.com/ai/google-will-use-machine-learning-to-try-and-tell-if-a-user-is-under-18-204713279.html?src=rss 

OpenAI will offer free ChatGPT users unlimited access to GPT-5

OpenAI’s upcoming GPT-5 release will integrate its o3 reasoning model and be available to free users, CEO Sam Altman revealed in a roadmap he shared on X. He said the company is also working to simplify how users interact with ChatGPT. 

“We want AI to ‘just work’ for you; we realize how complicated our model and product offerings have gotten,” Altman wrote. “We hate the model picker as much as you do and want to return to magic unified intelligence.” 

In its current iteration, forcing ChatGPT to use a specific model, such as o3-mini, involves either tapping the “Reason” button in the prompt bar or one of the options present in the model picker, which appears after the chatbot answers a question. If you pay for ChatGPT Plus or Pro, that dropdown menu can get pretty long, with multiple models and intelligence settings to choose from.  

OPENAI ROADMAP UPDATE FOR GPT-4.5 and GPT-5:

We want to do a better job of sharing our intended roadmap, and a much better job simplifying our product offerings.

We want AI to “just work” for you; we realize how complicated our model and product offerings have gotten.

We hate…

— Sam Altman (@sama) February 12, 2025

As for the company’s roadmap, Altman says GPT-4.5 will be OpenAI’s “last non-chain-thought model,” meaning everything that comes after will feature the capability to solve problems by breaking them down into a series of intermediate steps. Following the release of GPT 4.5, OpenAI’s primary goal is “to unify o-series models and GPT-series models by creating systems that can use all our tools, know when to think for a long time or not, and generally be useful for a very wide range of tasks.”

Looking ahead to GPT-5, Altman says OpenAI will release the model “as a system that integrates a lot of our technology,” including o3 and its recently released Deep Research feature. In a change of plans, OpenAI won’t release o3 as a standalone model. Previously, Altman had said the new system would arrive “shortly after” o3-mini, which OpenAI made available for public use at the end of last month. 

Once GPT-5 arrives, OpenAI plans to offer free users unlimited access to the model, “subject to abuse thresholds,” at “the standard intelligence setting.” Plus users will get to run GPT-5 “at a higher level of intelligence,” while Pro users will get to run the model at “an even higher level of intelligence.” 

Altman did not provide an exact timeline for either GPT-4.5 or GPT-5, other than to say they could arrive within weeks or months.  

This article originally appeared on Engadget at https://www.engadget.com/ai/openai-will-offer-free-chatgpt-users-unlimited-access-to-gpt-5-211935734.html?src=rss 

Overwatch 2’s next update overhauls the game with perks and loot boxes

Blizzard is making some major changes to Overwatch 2 for its Season 15 update, including adding all new gameplay in the form of a perks system and the return of loot box rewards, on top of the new heroes and skins.

Overwatch 2 has been through several changes since it left early access in 2023, many of which brought the game from the revamped structure that made it a sequel to something that works more like a free-to-play version of the original Overwatch. Perks are legitimately new, though. Now during a match you’ll be given two opportunities to pick a perk for your hero. One minor perk after leveling up for the first time, which adds smaller upgrades like a passive ability or a cooldown reduction, and one major perk that can alter the play style of your hero. For example, Torbjorn can pick from turrets that stick to ceilings and walls (a frankly diabolical upgrade) or an even more powerful “Level 3 turret.”

Blizzard Entertainment

As a reward for your hard work in-game, Blizzard is also adding to Overwatch 2‘s battle pass system by reintroducing loot boxes as another way to get cosmetics. You can earn loot boxes by completely weekly and event rewards, and you’ll get one Legendary Loot Box for having the Free Battle Pass or two Legendary Loot Boxes for having the Premium Battle Pass, as well. Beyond publishing drop rates for boxes, Blizzard is guaranteeing that “a Rare or better item will drop in every single box, with an Epic item within five consecutive boxes and a Legendary item within twenty consecutive boxes.”

Blizzard Entertainment

Coming a bit later in Season 16, Overwatch 2 will get an entirely new “Stadium” game mode that draws clear inspiration from competing shooters Counter-Strike 2 and Valorant. In a Stadium match you’ll be placed on a team of five fighting to earn currency across seven possible rounds. In-between rounds you can spend that currency to unlock upgrades and customize your abilities as you play. Stadium will also let you pick between the traditional first-person mode or a new third-person mode, which should make it easier to aim some of the wilder abilities you can unlock during a match.

The cherry on top to all these changes are new quality of life features Blizzard plans to add this year, like the ability to ban specific heroes from a match, and vote on maps you want to play. Of course, there are also two new heroes in the works: Freja, “a former search and rescue operative turned bounty hunter” with an explosive crossbow who will be playable in a free trial weekend during Season 15 before joining the game in Season 16, and a hero currently being called “Aqua” who will land in Season 18 and can control water.

Season 15 launches on February 18, while features like hero bans and map voting are coming later this year, in that order, according to Blizzard.

This article originally appeared on Engadget at https://www.engadget.com/gaming/overwatch-2s-next-update-overhauls-the-game-with-perks-and-loot-boxes-212409206.html?src=rss 

Acclaimed designer Jeff Minter is back with a remake of the ’80s arcade curio I, Robot

Iconic game designer Jeff Minter is back with another modern take on a long-forgotten Atari title. Minter has turned his psychedelic eye toward the 1984 arcade cabinet I, Robot. His version ups the visuals and takes serious liberties with the original design, adopting techno music and some new game modes.

The original I, Robot was a flop, despite being made by Dave Theurer, the guy behind Missile Command and Tempest. Maybe it was just ahead of its time. After all, it was the very first commercial video game to use real-time, flat-shaded 3D polygon graphics. The game came out a full eight years before Sega’s Virtua Racing and more than a decade before the PlayStation and N64 brought 3D gaming into the mainstream.

The original title had players control a robot as it jumped around mazes to flip the color of tiles and shoot enemies. There was also a giant eyeball that had to be avoided at all costs. Minter’s version keeps the core gameplay mechanic, but increases the speed and incorporates new gameplay elements. For instance, there are new tube shooter levels and an exploration mode called Ungame. It looks pretty nifty.

If the name Jeff Minter seems vaguely familiar, he’s the person who made Tempest 2000 and its various sequels. More recently, he remade an unreleased Atari prototype called Akka Arrh that originally dates back to 1982. He also provided the visuals for a Nine Inch Nails video.

Minter’s version of I, Robot will be released this spring on pretty much every platform, including Nintendo Switch, PC, PS4, PS5, Xbox One and Xbox Series X/S. The design team is also working on a port for PlayStation VR2.

This article originally appeared on Engadget at https://www.engadget.com/gaming/acclaimed-designer-jeff-minter-is-back-with-a-remake-of-the-80s-arcade-curio-i-robot-194037105.html?src=rss 

An Apple TV app is finally available for Android devices

After years of waiting, the Apple TV app is finally available natively on Android devices. This version was “built from the ground up” to take advantage of the Android operating system and is currently ready for download from the Google Play Store. It was designed for smartphones, tablets and foldables.

This is a full version of the app, with all of the stuff Apple users have come to expect. There’s a feature for picking up a show where you left off and a tool for downloading content to watch offline. It works via Wi-Fi or cellular. 

The app allows access to Apple TV+ content, so you can finally watch the absolute boatload of original shows the platform pumps out. It’s a veritable cornucopia of sci-fi goodness, with hit shows like Severance, Silo and For All Mankind. The platform also airs non-genre fare like Ted Lasso, Shrinking and Slow Horses. New users get a free seven-day trial to the platform. Otherwise, pricing is the same as always and folks can sign up for subscriptions via Google Play.

The app even grants access to MLS Season Pass content, just in time for Major League Soccer’s 2025 season. Sports fans can also check out a weekly MLB double-header each Friday with no local restrictions.

There have been workarounds to get Apple TV content on Android devices. In the past, users were forced to enter credentials via the web app or use a Prime Video bundle. Android TVs have long had a dedicated Apple TV app, but there was no way to subscribe on the television itself. That also changed today, as Google Play now offers the ability to subscribe on Android TVs.

There’s no casting support on this first version of the app, but we hope it’ll come soon via a future update. The Apple TV app only works on devices running Android 10 or later.

This article originally appeared on Engadget at https://www.engadget.com/entertainment/streaming/an-apple-tv-app-is-finally-available-for-android-devices-200044050.html?src=rss 

MyFitnessPal is going to let AI play your nutrionist

MyFitnessPal, a popular, free food logging app, is getting into AI, specifically letting AI plan your meals for you. Following an acquisition of a startup called Intent, MyFitnessPal is adding an AI-generated meal planning feature for subscribers.

The meal plans the updated MyFitnessPal app will be able to create take into account subscribers’ “goals, preferences, dietary habits, and routines,” according to MyFitnessPal, via information the app already has on you, and what looks like a survey. The feature will also be part of an end-to-end experience. Any plan the app creates can be automatically translated into a grocery list for the next time you’re at the store, or an order to a grocery delivery service if you’d prefer to do as little as possible.

MyFitnessPal

Given the sheer number of recipe and diet videos you can stumble upon across social media, trying to simplify the process of planning what you’re going to eat makes sense. Using AI to do it does seem like it could have some unintended consequences, though. Knowing AI’s ability to hallucinate, even if MyFitnessPal isn’t generating recipes from scratch, it still feels like you could end up with some oddball meal combinations or incorrect grocery lists.

The only way to know for sure is to try the AI meal planner for yourself, and if you’re interested, you can already sign-up for a waitlist so you’re able to try it out as soon as its available. MyFitnessPal says meal planning will officially launch in the spring in the US, UK, Ireland, Canada, Australia and New Zealand, and require joining a new $99.99 annual subscription tier called “Premium+” to use.

This article originally appeared on Engadget at https://www.engadget.com/apps/myfitnesspal-is-going-to-let-ai-play-your-nutrionist-191239496.html?src=rss 

Watch Sony’s State of Play showcase here at 5PM ET

Roll up, gamers. It’s time for your latest influx of PlayStation knowledge. Sony is hosting its first State of Play stream of 2025 today (February 12) at 5PM ET. The showcase will focus on PS5 games. You’ll be able to watch it on the PlayStation website or the brand’s YouTube and Twitch channels. Or, you know, by hitting the big play button on the YouTube video embedded above.

The State of Play will run for about 40 minutes, so there should be lots of juicy details here. Ghost of Yotei and Death Stranding 2: On the Beach are both slated for 2025 debuts, so we could get release dates for those. For what it’s worth, Death Stranding 2 director Hideo Kojima recently teased that he was working on a new trailer and he’s appearing at an SXSW panel next month to talk about the game.

Other first-party projects such as Marvel’s Wolverine, Marathon and Fairgame$ could get a look in during the State of Play. It seems far too early for us to get another peek at Naughty Dog’s Intergalactic: The Heretic Prophet, though there’s a chance that we’ll finally find out the exact release date for season two of The Last of Us on HBO.

As for third-party games, it’s a pretty safe bet that we’ll see a trailer and release date for Metal Gear Solid Delta: Snake Eater since, whoops, Sony itself leaked those last week. Forza Horizon 5 and Indiana Jones and the Great Circle are making the leap over from Xbox this spring, so release date reveals for those during the State of Play would make sense. 

Rumors also suggest that Senua’s Saga: Hellblade II, a remastered collection of the first three Gears of War games and even the Halo series, Xbox’s crown jewel, will cross the great divide to PlayStation. Perhaps we’ll learn about some of those during the stream too. In any case, you’ll be able to find coverage of all the big State of Play news on a cool website called Engadget dot com.

This article originally appeared on Engadget at https://www.engadget.com/gaming/playstation/watch-sonys-state-of-play-showcase-here-at-5pm-et-165148390.html?src=rss 

The Weather Channel app can show how likely you are to get specific snowfall amounts

We’re in the tail end of winter here, but it’s snowing as I write this. Not a lot, but just enough to be a potential frustration to the move I have planned for later this week. Basically, winter weather sucks, and one of the worst things about it is how unpredictable it can be. The Weather Company, owners and developers of the Weather Channel app (but not the TV station) know this, and have just released a feature to try and make planning for storms a little easier.

It’s called probabilistic snowfall, and it aims to give you greater insight into just how severe or impactful a storm might be. You’ll still see the standard “50 percent chance of snowfall” in your forecasts, but in the 72 hours leading up to snow, you’ll also see the percent change for specific snowfall amounts. For example, you’ll see that there’s an 80 percent chance of snow overall, but you’ll also see bars detailing how likely more specific snowfall amounts are. In this screenshot from the app, you can see that five to eight inches is the most likely outcome, followed by eight to 12 inches — enough snow that you might want to change your plans to deal with it.

According to Rachel Chukura, head of product at The Weather Company, that was precisely the vision behind the feature. “The idea is that users can see the different probabilities and know there could be variations, she said. “That’s important because of all the decisions we make in terms of what weather we think is coming. What time do you need to leave for your commute, will there be a school delay, and so on.” And while the company is starting with snowfall forecasts, they also hope to bring it to more forecasting tools over time.

I’ve been using the feature for about a week, just in time to see it pop up for the small storm we had in Philadelphia last night. I obviously can’t say how accurate it is, as I’m not out there measuring snowfall with a ruler. But the few inches we got overnight were in line with the probabilistic forecast, and it was helpful leading up to the storm to know I likely didn’t have to worry about the weather truly being an inconvenience.

The feature is part of the Weather Channel app’s premium subscription, which costs $30 per year. Besides probabilistic snowfall, that also removes ads, shows longer-term forecasts on the radar, adds a 30-mile lightning strike map and provides more detailed 15-minute weather breakdowns. While most people out there will stick with the default weather app on their phones, or just use the free version of the Weather Channel or other similar apps, there’s definitely something to be said for greater insight into how likely you are to get crushed by a snowstorm.

This article originally appeared on Engadget at https://www.engadget.com/apps/the-weather-channel-app-can-show-how-likely-you-are-to-get-specific-snowfall-amounts-165900316.html?src=rss 

The iPad mini 7 is back on sale for $100 off

The iPad mini 7 is once again on sale for $100 off via Amazon. This applies to every design variation, so the 128GB standard model is now $400 and the 512GB Wi-Fi/cellular version is $850. However, not all of the colors have been blessed by the discount fairy. Most colorways get a discount of some kind, but it may not apply to all of the hardware variations.

Apple’s latest and greatest diminutive tablet easily found a place on our list of the best iPads. It’s a solid refresh without any real tradeoffs when compared to the full-sized tablet. It supports the Apple Pencil Pro, the A17 Pro chip is plenty powerful and the overall design is comfortable to hold. It’s an iPad mini.

We called it a “solid and necessary update to the best small tablet on the market” in our official review. We came away impressed by the stereo speakers, which sound much better than one would expect. The display is nice, despite maxing out at a 60Hz refresh rate, and the camera array takes surprisingly decent shots, even in low light. It’s great for everyday computing tasks and, of course, for playing mobile games like Balatro.

This model doesn’t include an M-series chip, though the A17 Pro is plenty powerful. Also, the battery life isn’t going to be winning any awards. We struggled to squeeze ten hours of it. This model also lacks Face ID, which could be an issue for some users. Beyond these minor nitpicks, however, this is a near-perfect teensy tablet.

Follow @EngadgetDeals on Twitter and subscribe to the Engadget Deals newsletter for the latest tech deals and buying advice.

This article originally appeared on Engadget at https://www.engadget.com/deals/the-ipad-mini-7-is-back-on-sale-for-100-off-162536671.html?src=rss 

Sony A1 II review: A powerhouse camera that falls short of its own high standards

How do you follow up a product that has reigned as the king of mirrorless cameras for the last four years? For Sony, the answer with the A1 was simple: just improve everything. The result is the $6,500 A1 II, a powerful pro-level hybrid mirrorless camera that’s Sony’s best to date.

The A1 II is equipped with a 50-megapixel stacked sensor, not a global sensor like the A9 III, so it’s not quite as fast. However, its blend of speed and resolution makes it the most versatile camera in Sony’s lineup. It also impresses on the video side with 8K 30 fps 10-bit capture, along with 4K slo-mo at up to 120 fps.

I expect a lot out of a Sony flagship camera, though, and while the A1 II has lots of small improvements, there’s nothing terribly innovative. It’s also up against impressive (and cheaper) competition like the $4,300 Canon EOS R5 II and $4,000 Nikon Z8. Plus, there’s the similarly priced pro-level Canon R1 and Nikon Z9. So despite its strengths, I found Sony’s new flagship to be slightly disappointing.

Body and handling

The A1 II is Sony’s best-handling camera to date. In terms of design and control placement, it’s nearly identical to the company’s high-end A9 III. At 743 grams (1.64 pounds), it’s heavier than the 617-gram (1.36 pound) A1, but still a featherweight compared to other professional models like the three-pound Z9. The new grip has slightly softer materials and more rounded contours, so it’s more comfortable and secure to hold than the hard-edged A1. That’s a big help to working pros with heavy lenses.

If you like manual controls, the A1 II is a dream. It has three primary dials, compared to just two on rival flagships. There’s a mode selector ring to choose photo, video and so-called Slow & Quick modes, along with a dual-control wheel for both shooting and autofocus. Around back, there’s yet another dial, plus a joystick and numerous customizable buttons.

The A1 II’s menu system has been refreshed to match the A9 III. It now offers a “home” setting for your most-used features, and everything is neatly organized into sub-menus. Sony also added an easy-to-use section for customizing buttons, dials, the quick menu and more.

Steve Dent for Engadget

Despite the price difference, it’s natural to compare the A1 II’s design and setup with the Canon R5 II as both are high-performance, high-resolution cameras. The R5 II does feel a bit nicer thanks to the more rounded shape and softer materials. But the A1 II has a larger array of controls, so it’s easier and faster to shoot with it once you get used to them. Sony also has a more intuitive menu system that makes it easier to find key settings.

Then there’s the A1 II’s sublime electronic viewfinder (EVF). Though it has the same resolution as the A1 at 9.44 million dots, it no longer drops in resolution at its default 120 fps frame rate. With that, it’s so sharp and bright that it blows away any argument for an optical viewfinder. Sony even provides two eyepiece cups, including what it calls a “squishy” option that molds around your eye to block out light, giving me a fully immersive experience. With all of that, this is easily the best EVF I’ve ever used.

Vloggers complained about the lack of a fully articulating display on the A1. Sony addressed that by outfitting the A1 II with a bright, 3-inch, 2.1-million-dot screen that not only tilts up and down but also fully swivels — ideal for both photographers and content creators.

As before, the A1 II has a dual card slot system with support for both SD UHS II and (sigh) Sony’s CFexpress Type A cards that no one else uses. Those are faster than SD, but not as fast as the CFexpress Type B slots found on the R5 II, Z8 and others.

Steve Dent for Engadget

The Z-type battery delivers a decent 520 shots on a charge, which is just slightly less than the 530 on the A1. It comes with a nice array of ports, including full-sized HDMI and high-speed USB-C 3.2 Gen 2 that allows rapid 10Gbps file transfers. You also get mic and headphone ports, a 2.5 Gbps ethernet connection and a flash sync port. Even the Wi-Fi was upgraded with 2×2 MIMO support that allows 2.5Gbps transfer speeds instead of 1Gbps as before.

Performance

The A1 II is the fastest high-resolution camera available, letting you hit burst speeds up to 30 fps when shooting RAW with continuous autofocus enabled in electronic mode, or 10 fps with the mechanical shutter. Those speeds match or beat its main rivals, but aren’t an improvement over the four-year-old A1. That’s a bit of a surprise, given that the A1 II has a much faster Bionz XR image processor borrowed from the A9 III.

The autofocus is a big improvement, though. It locks onto subjects quicker than before, so I saw fewer out-of-focus shots when firing off bursts. The A1 II also has Sony’s latest algorithms and AI features from the A9 III that make subject-recognition faster and more user-friendly. It can now recognize human bodies, faces and eyes, along with animals, birds, insects and several types of vehicles. On top of that, the A1 II is Sony’s first camera to automatically identify those subjects so you don’t have to tediously select them yourself. In other words, if you’re photographing birds and happen to see a bear, you won’t waste valuable time switching modes.

Sony is a bit late to the game with this auto subject selection, as Canon and other brands have had it for a while. However, the A1 II has the best implementation I’ve seen to date, as it chooses and locks onto subjects more quickly. Like other cameras, it can get tripped up in situations with multiple subjects, occasionally selecting the wrong one. With human subjects, Canon has an edge thanks to its face-memorization feature on the R1 and R5 that will always try to lock onto a specific person.

Steve Dent for Engadget

While the A1 II has both a mechanical and electronic shutter, the stacked sensor’s fast readout speeds mean you can use the latter almost exclusively. Even in sports like golf with high speed movements, skewing and distortion is rarely an issue when using the electronic shutter, and you can shoot photos silently without disturbing participants.

Pre-capture is a key new feature, letting you keep up to 70 frames when half-pressing the shutter button, before fully pressing it. That reduces the odds of missing a decisive moment in a scene.

When it comes to stabilization, the A1 II also represents a first for the company. It offers 8.5 stops, beating the A9 III by a half-stop and matching the EOS R5 II and R1. Thanks to that, I was able to nail sharp photos even at ridiculously slow shutter speeds, like a full second. That’s ideal when you need to shoot in low light or deliberately blur moving subjects without using a tripod.

While the A1 II is quick, it didn’t bring the generational leap I expected compared to a four-year-old camera. My professional photographer friends (who already own A1s) felt the same way and won’t be upgrading.

Image quality

The A1 II is not just fast, but also delivers high resolution, and that combination is why you pay the big bucks. Images are sharper than just about any full-frame camera, with the exception of Sony’s own A7R V. Dynamic range is excellent and colors are accurate when shooting JPEGs, though I prefer Canon’s warmer images straight out of the camera.

The new model has the same sensor as the A1, but Sony managed to boost dynamic range very slightly at higher ISOs. As a result, you get minimal noise up to around ISO 6,400 and images are very usable at ISO 12,800. This is outstanding for a 50MP camera, and better than Canon’s EOS R5 II. Beyond that, and up to the camera’s maximum ISO 32,000, noise starts to become distracting.

RAW files give photographers ample room to dial down blown out highlights or boost detail in shadow areas. If you need the maximum amount of dynamic range, you’ll want to shoot using the lossless compressed RAW format, as lossy compressed RAW files are noticeably worse in highly detailed images. The downside to the latter is that shooting them limits burst speeds to 20 fps and fills your memory card much faster.

While image quality is outstanding, it’s not much of an improvement over the A1. For studio and landscape photographers who value image quality above all and already own an A1, the A1 II likely isn’t worth the investment.

Video

On paper, the A1 II is a powerhouse for video, offering up to 8K 30 fps recording and 4K 60 fps(up to 4K 120 fps with no sound), with 10-bit log shooting available. It comes with a new feature that lets you load custom LUTs when shooting in log (S-Log3) mode, so you can better judge your exposure. It also has a setting that lets you better match footage with Sony’s professional cinema cameras. That’s on top of the other improvements I already mentioned, like the fully articulating display and upgraded stabilization, that are also beneficial when recording video.

Frame of 8K video from the Sony A1 II

Steve Dent for Engadget

Autofocus is much improved with more reliable tracking of moving subjects, as I found when shooting a fast-moving soccer event. It also supports the same subject tracking modes available when capturing photos (eyes, faces, bodies, birds, animals and vehicles).

Looking closer, though, the video capabilities are slightly disappointing. Unlike the much cheaper Nikon Z8 and Canon EOS R5 II cameras, the A1 II doesn’t offer internal RAW recording, possibly because the CFexpress Type A slots can’t handle the required bandwidth. That can make a big difference with video quality and the ability to adjust colors and levels when editing.

Still, when recording 8K or 4K 10-bit log video at the highest data rates, image quality and dynamic range are solid. The S-Log3 and 10-bit features give content creators a lot of room to tweak footage shot in dim interiors or a bright sunny day.

The in-body stabilization may be the best I’ve seen on any camera to date. The optical mode does a great job removing any hand jitters if you don’t move around too much. And for walking or running, the electronic stabilization (not available in 8K) provides near GoPro-like smoothing, with none of the sudden jolts seen on other models. If you do need to move the camera quickly, rolling shutter is usually a non-issue. I only saw any distortion when shooting 8K video, and even then, it was minimal.

Wrap-up

Steve Dent for Engadget

The A1 II is an awesome camera and near the top of the food chain in terms of speed and power. Nearly every aspect has been improved over the A1, including the stabilization, body design and autofocus.

However, the innovation that Sony is known for is lacking here, and on top of that, the A1 II is very expensive. With that taken into account, I think it falls short. The A1 II is only a mild upgrade over the A1 and when it comes to video, it’s lagging behind the Nikon Z8 and Z9 as well as the Sony R5 II.

If you already have the A1, I don’t think the A1 II offers enough extra to justify a new investment. And anyone who is more serious about video should consider either the Canon R5 II or Nikon Z8, as both offer higher-quality RAW video at up to 8K 60p for less money. However, for pros or serious amateurs going into the A1 lineup for the first time, who focus on photos and have no qualms about the $6,500 price, the Sony A1 II is a great pick.

This article originally appeared on Engadget at https://www.engadget.com/cameras/sony-a1-ii-review-a-powerhouse-camera-that-falls-short-of-its-own-high-standards-150053430.html?src=rss 

Generated by Feedzy
Exit mobile version