Android’s Circle to Search can now help students solve math and physics homework

Google has introduced another capability for its Circle to Search feature at the company’s annual I/O developer conference, and it’s something that could help students better understand potentially difficult class topics. The feature will now be able to show them step-by-step instructions for a “range of physics and math word problems.” They just have to activate the feature by long-pressing the home button or navigation bar and then circling the problem that’s got them stumped, though some math problems will require users to be signed up for Google’s experimental Search Labs feature.

The company says Circle to Search’s new capability was made possible by its new family of AI models called LearnLM that was specifically created and fine-tuned for learning. It’s also planning to make adjustments to this particular capability and to roll out an upgraded version later this year that could solve even more complex problems “involving symbolic formulas, diagrams, graphs and more.” Google launched Circle to Search earlier this year at a Samsung Unpacked event, because the feature was initially available on Galaxy 24, as well as on Pixel 8 devices. It’s now also out for the Galaxy S23, Galaxy S22, Z Fold, Z Flip, Pixel 6 and Pixel 7 devices, and it’ll likely make its way to more hardware in the future. 

In addition to the new Circle to Search capability, Google has also revealed that devices that can support the Gemini for Android chatbot assistant will now be able to bring it up as an overlay on top of the application that’s currently open. Users can then drag and drop images straight from the overlay into apps like Gmail, for instance, or use the overlay to look up information without having to swipe away from whatever they’re doing. They can tap “Ask this video” to find specific information within a YouTube video that’s open, and if they have access to Gemini Advanced, they can use the “Ask this PDF” option to find information from within lengthy documents. 

Google is also rolling out multimodal capabilities to Nano, the smallest model in the Gemini family that can process information on-device. The updated Gemini Nano, which will be able to process sights, sounds and spoken language, is coming to Google’s TalkBack screen reader later this year. Gemini Nano will enable TalkBack to describe images onscreen more quickly and even without an internet connection. Finally, Google is currently testing a Gemini Nano feature that can alert users while a call is ongoing if it detects common conversation patterns associated with scams. Users will be alerted, for instance, if they’re talking to someone asking them for their PINs or passwords or to someone asking them to buy gift cards. 

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/androids-circle-to-search-can-now-help-students-solve-math-and-physics-homework-180223229.html?src=rss 

Google builds Gemini right into Android, adding contextual awareness within apps

Google just announced some nifty improvements to its Gemini AI chatbot for Android devices as part of the company’s I/O 2024 event. The AI is now part of the Android operating system, allowing it to integrate in a more comprehensive way.

The coolest new feature wouldn’t be possible without that integration with the underlying OS. Gemini is now much better at understanding context as you control apps on the smartphone. What does this mean exactly? Once the tool officially launches as part of Android 15, you’ll be able to bring up a Gemini overlay that rests on top of the app you’re using. This will allow for context-specific actions and queries.

Google gives the example of quickly dropping generated images into Gmail and Google Messages, though you may want to steer clear of historical images for now. The company also teased a feature called “Ask This Video” that lets users pose questions about a particular YouTube video, which the chatbot should be able to answer. Google says this should work with “billions” of videos. There’s a similar tool coming for PDFs. 

Google

It’s easy to see where this tech is going. Once Gemini has access to the lion’s share of your app library, it should be able to actually deliver on some of those lofty promises made by rival AI companies like Humane and Rabbit. Google says it’s “just getting started with how on-device AI can change what your phone can do” so we imagine future integration with apps like Uber and Doordash, at the very least.

Circle to Search is also getting a boost thanks to on-board AI. Users will be able to circle just about anything on their phone and receive relevant information. Google says people will be able to do this without having to switch apps. This even extends to math and physics problems, just circle for the answer, which is likely to please students and frustrate teachers.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/google-builds-gemini-right-into-android-adding-contextual-awareness-within-apps-180413356.html?src=rss 

Google’s Gemini Nano brings better image-description smarts to its TalkBack vision tool

The Google I/O event is here, and the company is announcing lots of great updates for your Android device. As we heard earlier, Gemini Nano is getting multimodal support, meaning your Android will still process text but with a better understanding of other factors like sights, sounds and spoken language. Now Google has shared that the new tool is also coming to it’s TalkBack feature.

TalkBack is an existing tool that reads aloud a description of an image, whether it’s one you captured or from the internet. Gemini Nano’s multimodal support should provide a more detailed understanding of the image. According to Google, TalkBack users encounter about 90 images each day that don’t have a label. Gemini Nano should be able to provide missing information, such as what an item of clothing looks like or the details of a new photo sent by a friend. 

Gemini Nano works directly on a person’s device, meaning it should still function properly without any network connection. While we don’t yet have an exact date for when it will arrive, Google says TalkBack will get Gemini Nano’s updated features later this year.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/googles-gemini-nano-brings-better-image-description-smarts-to-its-talkback-vision-tool-180759598.html?src=rss 

With Gemini Live, Google wants you to relax and have a natural chat with AI

While Google and OpenAI have been racing to win the AI crown over the past year, we’ve seemingly reverted away from the idea of speaking to virtual assistants. Generative AI products have typically launched with text-only inputs, and only later add the ability to search images and basic voice commands. At Google I/O today, the company showed off Gemini Live, a new mobile experience for natural conversations with its AI. 

Google offered up a few potential use cases; You could have a conversation with Gemini Live to help prepare for a job interview, where it could potentially ask you relevant questions around the positions. It could also give you public speaking tips if you want to research a speech. What makes Gemini Live unique is that you’ll be able to speak at your own pace, or even interrupt its responses if you’d like. Ideally, it should be more like having a conversation with a person, instead of just voicing smart assistant commands or generative AI queries.

At I/O, Google also showed off Project Astra, a next-generation virtual assistant that takes the concept of Gemini Live even further. Astra is able to view your camera feed and answer questions in real-time. It’s unclear how long that’ll take to arrive, but Google says some of Astra’s live video features will come to Gemini Live later this year. Gemini Live will be available for Gemini Advanced subscribers in the next few months.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/with-gemini-live-google-wants-you-to-relax-and-have-a-natural-chat-with-ai-181329788.html?src=rss 

Google announces new scam detection tools that provide real-time alerts during phone calls

Google just announced forthcoming scam detection tools coming to Android phones later this year, which is a good thing as these scammers keep getting better and better at parting people from their money. The toolset, revealed at Google I/O 2024, is still in the testing stages but uses AI to suss out fraudsters in the middle of a conversation.

You read that right. The AI will be constantly on the hunt for conversation patterns commonly associated with scams. Once detected, you’ll receive a real-time alert on the phone, putting to bed any worries that the person on the other end is actually heading over to deliver a court summons or whatever.

Google gives the example of a “bank representative” asking for personal information, like PINs and passwords. These are uncommon bank requests, so the AI would flag them and issue an alert. Everything happens on the device, so it stays private. This feature isn’t coming to Android 15 right away and the company says it’ll share more details later in the year. We do know that people will have to opt-in to use the tool. 

Google made a big move with Android 15, bringing its Gemini chatbot to actual devices instead of requiring a connection to the cloud. In addition to this scam detection tech, the addition of onboard AI will allow for many more features, like contextual awareness when using apps.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/google-announces-new-scam-detection-tools-that-provide-real-time-alerts-during-phone-calls-181442091.html?src=rss 

Google Gemini can power a virtual AI teammate with its own Workspace account

Google’s Gemini AI systems can do a lot, judging by today’s I/O keynote. That includes the option to set up a virtual teammate with its own Workspace account. You can configure the teammate to carry out specific tasks, such as to monitor and track projects, organize information, provide context, pinpoint trends after analyzing data and to play a role in team collaboration.

In Google Chat, the teammate can join all relevant rooms and you can ask it questions based on all the conversation histories, Gmail threads and anything else it has access to. It can tell team members whether their projects are approved or if there might be an issue based on conflicting messages. 

It seems like the virtual teammate was just a tech demo for now, however. Aparna Pappu, vice president and GM of Workspace, said Google has “a lot of work to do to figure out how to bring these agentive experiences, like virtual teammates, into Workspace.” That includes finding ways to let third parties make their own versions. 

While it doesn’t seem like this virtual teammate will be available soon, it could eventually prove to be a serious timesaver — as long as you trust it to get everything right first time around.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/google-gemini-can-power-a-virtual-ai-teammate-with-its-own-workspace-account-182809274.html?src=rss 

Castlevania is coming to Dead by Daylight later this year

Dead by Daylight fans can check out the Dungeons and Dragons chapter starting today, but Behaviour Interactive teased another high-profile crossover during an anniversary showcase on Tuesday. A Castlevania chapter is on the way to DbD. There aren’t any details yet about what that will include, but you just might get to play as Alucard or Simon Belmont in the fog. Behaviour plans to divulge more info about the Castlevania tie-up, which will arrive later this year, in August.

As for the Dungeons and Dragons chapter, which brings a dark fantasy element to DbD for the first time, Behaviour spilled the beans on that during the stream. PC players will be able to try out the chapter on the public test build today before it goes live for everyone on June 3.

Behaviour Interactive

The new killer is Vecna (the D&D version rather than the Stranger Things baddie) and stalwart video game actor and Critical Role mastermind Matt Mercer is voicing the character. The latest survivor is actually two identities in one. It’s a bard character that you can opt to play as an elf female or human male, bringing a slight element of D&D-style character creation to DbD.

The chapter will also include a new map, which (surprise, surprise) is a dungeon. Whenever you’re up against Vecna, you’ll be able to find treasure chests which will trigger a roll of a 20-sided die when opened. Rolling a one will net you a nasty surprise while getting a 20 grants you a powerful magical item. Roll any number in between and you’ll get a helpful item.

Speaking of maps, there will be larger ones to check out in an upcoming new mode. DbD has long pitted four survivors against one killer. A pair of killers will soon be able to team up and hunt eight survivors. They’ll be able to take advantage of team powers too.

There will be a lot of changes for this 2 vs. 8 mode, which will be around for a limited time at first. Perks will be jettisoned in favor of a class system, and there won’t be any hooks. Downed survivors will instead go straight to a cage. If a survivor is caged three times, they’re out of the game. Behaviour sees this as more of a party mode as opposed to the competitive nature of 1 vs. 4. The 2 vs. 8 mode is slated to arrive later this summer, and you can expect to find out more about it in July.

Behaviour also had some news about several DbD spinoff games that are in the works. The Casting of Frank Stone is a single-player, narrative-focused game set in the DbD universe and developer Supermassive has released the first gameplay trailer.

The spinoff tells the story of a group of young people who venture into a condemned steel mill in 1980 while attempting to film their own horror movie. There, they discover evidence of crimes committed by serial killer Frank Stone.

The gameplay sounds very familiar for those who have experience of previous Supermassive games like Until Dawn and The Quarry. The direction of the story will shift based on your narrative decisions and how you handle environmental puzzles and quick-time events. The Casting of Frank Stone, which is said to be about five to seven hours long, is slated to arrive later this year.

An untitled co-op shooter spinoff from Midwinter Entertainment is still in early development, but it now has a codename: Project T. It’ll be a third-person game and unlike the survivors in DbD, you’ll actually be able to fight back against enemies using guns. Fans who want to find out more can sign up for an insider program, which will include updates, closed playtests and the chance to provide feedback.

That’s not all though, as Behaviour announced yet another DbD spinoff. What the Fog is a two-person co-op roguelike that it developed in-house. The premise is that DbD survivors Claudette and Dwight are sucked into a cursed board game, Jumanji-style. The game is mainly played in third-person, but if you die you’ll move into a bird’s-eye support mode, where you can help your teammate survive. Just like in DbD, you’ll need to interact with a hook to revive your ally. There’s a single-player mode, while Feng Min is an unlockable character.

Behaviour Interactive

What the Fog shares some elements with DbD. You’ll need to pick up tokens called blood points by roaming the map and killing enemies. These let you activate generators so you can escape a room. You’ll get a buff from each generator and acquire a weapon upgrade after each round. There are bosses to take down too. What the Fog also has a more cartoony look than DbD‘s more realistic art style.

I’ve played a few rounds of the single-player mode and I’m enjoying it quite a bit so far. The metal soundtrack and monster-slaying chaos actually reminds me a bit of the Doom series. After unlocking a door, I’d suggest sticking around in the room a while longer to kill some more enemies and snag a bunch of blood points. That way, you’ll be able to repair the next room’s generators quickly and power up before taking on a fresh army of monsters.

What the Fog is available now on Steam. The first 2 million copies are available for free, though you’ll need a Behaviour account to claim one. If you’re not quick enough to snag a free copy or just feel like giving Behavior a few bucks, the game costs $5.

This article originally appeared on Engadget at https://www.engadget.com/castlevania-is-coming-to-dead-by-daylight-later-this-year-164509826.html?src=rss 

Ask Google Photos to get help making sense of your gallery

Google is inserting more of its Gemini AI into every single product it has and the next target in its sights is Photos. At its I/O developer conference today, the company’s CEO Sundar Pichai announced a feature called Ask Photos, which is designed to help you find specific images in your gallery by talking to Gemini. 

Ask Photos will show up as a new tab at the bottom of your Google Photos app. It’ll start rolling out to One subscribers first, starting in US English over the upcoming months. When you tap over to that panel, you’ll see the Gemini star icon and a welcome message above a bar that prompts you to “search or ask about Photos.”

According to Google, you can ask things like “show me the best photo from each national park I’ve visited,” which not only draws upon GPS information but also requires the AI to exercise some judgement in determining what is “best.” The company’s VP for Photos Shimrit Ben-Yair told Engadget that you’ll be able to provide feedback to the AI and let it know which pictures you preferred instead. “Learning is key,” Ben-Yair said.

You can also ask Photos to find your top photos from a recent vacation and generate a caption to describe them so you can more quickly share them to social media. Again, if you didn’t like what Gemini suggested, you can also make tweaks later on.

For now, you’ll have to type your query to Ask Photos — voice input isn’t yet supported. And as the feature rolls out, those who opt in to use it will see their existing search feature get “upgraded” to Ask. However, Google said that “key search functionality, like quick access to your face groups or the map view, won’t be lost.”

The company explained that there are three parts to the Ask Photos process: “Understanding your question,” “crafting a response” and “ensuring safety and remembering corrections.” Though safety is only mentioned in the final stage, it should be baked in the entire time. The company acknowledged that “the information in your photos can be deeply personal, and we take the responsibility of protecting it very seriously.”

To that end, queries are not stored anywhere, though they are processed in the cloud (not on device). People will not review conversations or personal data in Ask Photos, except “in rare cases to address abuse or harm.” Google also said it doesn’t train “any generative AI product outside of Google Photos on this personal data, including other Gemini models and products.”

Your media continues to be protected by the same security and privacy measure that cover your use of Google Photos. That’s a good thing, since one of the potentially more helpful ways to use Ask Photos might be to get information like passport or license expiry dates from pictures you might have snapped years ago. It uses Gemini’s multimodal capabilities to read text in images to find answers, too.

Of course, AI isn’t new in Google Photos. You’ve always been able to search the app for things like “credit card” or a specific friend, using the company’s facial and object recognition algorithms. But Gemini AI brings generative processing so Photos can do a lot more than just deliver pictures with certain people or items in them.

For example, you might also get Photos to tell you what themes you might have used for the last few birthday parties you threw for your partner or child. Gemini AI is at work here to study your pictures and figure out what themes you already adopted.

There are a lot of promising use cases for Ask Photos, which is an experimental feature at the moment and that is “starting to roll out soon.” Like other Photos tools, it might begin as a premium feature for One subscribers and Pixel owners before trickling down to all who use the free app. No official word yet on when or whether that might happen, though.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/ask-google-photos-to-get-help-making-sense-of-your-gallery-170734062.html?src=rss 

Google’s new Gemini 1.5 Flash AI model is lighter than Gemini Pro and more accessible

Google announced updates to its Gemini family of AI models at I/O, the company’s annual conference for developers, on Tuesday. It’s rolling out a new model called Gemini 1.5 Flash, which it says is optimized for speed and efficiency.

“[Gemini] 1.5 Flash excels at summarization, chat applications, image and video captioning, data extraction from long documents and tables, and more,” wrote Demis Hassabis, CEO of Google DeepMind, in a blog post. Hassabis added that Google created Gemini 1.5 Flash because developers needed a model that was lighter and less expensive than the Pro version, which Google announced in February. Gemini 1.5 Pro is more efficient and powerful than the company’s original Gemini model announced late last year.

Gemini 1.5 Flash sits between Gemini 1.5 Pro and Gemini 1.5 Nano, Google’s smallest model that runs locally on devices. Despite being lighter weight then Gemini Pro, however, it is just as powerful. Google said that this was achieved through a process called “distillation”, where the most essential knowledge and skills from Gemini 1.5 Pro were transferred to the smaller model. This means that Gemini 1.5 Flash will get the same multimodal capabilities of Pro, as well as its long context window – the amount of data that an AI model can ingest at once – of one million tokens. This, according to Google, means that Gemini 1.5 Flash will be capable of analyzing a 1,500-page document or a codebase with more than 30,000 lines at once. 

Gemini 1.5 Flash (or any of these models) aren’t really meant for consumers. Instead, it’s a faster and less expensive way for developers building their own AI products and services using tech designed by Google.

In addition to launching Gemini 1.5 Flash, Google is also upgrading Gemini 1.5 Pro. The company said that it had “enhanced” the model’s abilities to write code, reason and parse audio and images. But the biggest update is yet to come – Google announced it will double the model’s existing context window to two million tokens later this year. That would make it capable of processing two hours of video, 22 hours of audio, more than 60,000 lines of code or more than 1.4 million words at the same time.

Both Gemini 1.5 Flash and Pro are now available in public preview in Google’s AI Studio and Vertex AI. The company also announced today a new version of its Gemma open model, called Gemma 2. But unless you’re a developer or someone who likes to tinker around with building AI apps and services, these updates aren’t really meant for the average consumer.

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/googles-new-gemini-15-flash-ai-model-is-lighter-than-gemini-pro-and-more-accessible-172353657.html?src=rss 

Google’s Project Astra uses your phone’s camera and AI to find noise makers, misplaced items and more.

When Google first showcased its Duplex voice assistant technology at its developer conference in 2018, it was both impressive and concerning. Today, at I/O 2024, the company may be bringing up those same reactions again, this time by showing off another application of its AI smarts with something called Project Astra. 

The company couldn’t even wait till its keynote today to tease Project Astra, posting a video to its social media of a camera-based AI app yesterday. At its keynote today, though, Google’s DeepMind CEO Demis Hassabis shared that his team has “always wanted to develop universal AI agents that can be helpful in everyday life.” Project Astra is the result of progress on that front. 

What is Project Astra?

According to a video that Google showed during a media briefing yesterday, Project Astra appeared to be an app which has a viewfinder as its main interface. A person holding up a phone pointed its camera at various parts of an office and verbally said “Tell me when you see something that makes sound.” When a speaker next to a monitor came into view, Gemini responded “I see a speaker, which makes sound.”

The person behind the phone stopped and drew an onscreen arrow to the top circle on the speaker and said, “What is that part of the speaker called?” Gemini promptly responded “That is the tweeter. It produces high-frequency sounds.”

Then, in the video that Google said was recorded in a single take, the tester moved over to a cup of crayons further down the table and asked “Give me a creative alliteration about these,” to which Gemini said “Creative crayons color cheerfully. They certainly craft colorful creations.”

Wait, were those Project Astra glasses? Is Google Glass back?

The rest of the video goes on to show Gemini in Project Astra identifying and explaining parts of code on a monitor, telling the user what neighborhood they were in based on the view out the window. Most impressively, Astra was able to answer “Do you remember where you saw my glasses?” even though said glasses were completely out of frame and were not previously pointed out. “Yes, I do,” Gemini said, adding “Your glasses were on a desk near a red apple.”

After Astra located those glasses, the tester put them on and the video shifted to the perspective of what you’d see on the wearable. Using a camera onboard, the glasses scanned the wearer’s surroundings to see things like a diagram on a whiteboard. The person in the video then asked “What can I add here to make this system faster?” As they spoke, an onscreen waveform moved to indicate it was listening, and as it responded, text captions appeared in tandem. Astra said “Adding a cache between the server and database could improve speed.”

The tester then looked over to a pair of cats doodled on the board and asked “What does this remind you of?” Astra said “Schrodinger’s cat.” Finally, they picked up a plush tiger toy, put it next to a cute golden retriever and asked for “a band name for this duo.” Astra dutifully replied “Golden stripes.”

How does Project Astra work?

This means that not only was Astra processing visual data in realtime, it was also remembering what it saw and working with an impressive backlog of stored information. This was achieved, according to Hassabis, because these “agents” were “designed to process information faster by continuously encoding video frames, combining the video and speech input into a timeline of events, and caching this information for efficient recall.”

It was also worth noting that, at least in the video, Astra was responding quickly. Hassabis noted in a blog post that “While we’ve made incredible progress developing AI systems that can understand multimodal information, getting response time down to something conversational is a difficult engineering challenge.”

Google has also been working on giving its AI more range of vocal expression, using its speech models to “enhanced how they sound, giving the agents a wider range of intonations.” This sort of mimicry of human expressiveness in responses is reminiscent of Duplex’s pauses and utterances that led people to think Google’s AI might be a candidate for the Turing test.

When will Project Astra be available?

While Astra remains an early feature with no discernible plans for launch, Hassabis wrote that in future, these assistants could be available “through your phone or glasses.” No word yet on whether those glasses are actually a product or the successor to Google Glass, but Hassabis did write that “some of these capabilities are coming to Google products, like the Gemini app, later this year.”

Catch up on all the news from Google I/O 2024 right here!

This article originally appeared on Engadget at https://www.engadget.com/googles-project-astra-uses-your-phones-camera-and-ai-to-find-noise-makers-misplaced-items-and-more-172642329.html?src=rss 

Generated by Feedzy
Exit mobile version