10 May 2023

Google is changing it looks | Gizmodo

 

1  This article is part of a developing story. Our writers and editors will be updating this page as new information is released. Please check back again in a few minutes to see the latest updates. Meanwhile, if you want more news coverage, check out our techscience, or io9 front pages. And you can always see the most recent Gizmodo news stories at gizmodo.com/latest.

Everything Google Announced About AI, Bard, and PaLM 2 at I/O 2023

Google is sticking AI into Search, its Workspace apps, Maps, and more. It's also updated its language model.

Google showed off its text-to-image generator along with music from Dan Deacon at Google I/O 2023

Music artist Dan Deacon constantly joked about a ‘duck with lips’ before the start of Google I/O 2023. His performance included AI-generated content from Google’s new text-to-video models.
Screenshot: Google / Gizmodo

"Before Google’s latest I/O event could even begin, the tech giant was already trying to stake a claim for its new AI. As music artist Dan Deacon played trippy, chip tune-esque electronic vocals to the gathered crowd, the company displayed the work of its latest text-to-video AI model with a psychedelic slideshow featuring, people in office settings, a trip through space, and mutating ducks surrounded by mumbling lips.

The spectacle was all to show that Google is bringing out the big guns for 2023, taking its biggest swing yet in the ongoing AI fight. At Google I/O, the company displayed a host of new internal and public-facing AI deployments. CEO Sundar Pichai said the company wanted to make AI helpful “for everyone.”

What this really means is—simply—Google plans to stick some form of AI into an increasing number of user-facing software products and apps on its docket.

Google’s sequel to its language model, called PaLM 2

Google unveiled its latest large language model that’s supposed to kick its AI chatbots and other text-based services into high gear. The company said this new LLM is trained on more than 100 languages and has strong coding, mathematics, and creative writing capabilities.

The company said there are four different versions of PaLM, with the smallest “Gecko” version small enough to work on mobile devices. Pichai showed off one example where a user asked PaLM 2 to give comments on a line of code to another developer while translating it into Korean. PaLM will be available in a preview sometime later this year.

PaLM 2 is a continuation of 2022’s PaLM and March’s PaLM-E multimodal model already released earlier this year. In addition, its limited medical-based LLM caled Med-PaLM 2 is supposed to accurately answer medical questions. Google claimed this system achieved 85% accuracy on the US Medical Licensing Examination. Though there’s still enormous ethical questions to work out with AI in the medical field, Pichai said Google hopes Med-PaLM will be able to identify medical diagnoses by looking at images like an X-ray.

Google’s Bard upgrades

Google’s “experiment” for conversational AI has gotten some major upgrades, according to Google. It’s now running on PaLM 2 and has a dark mode, but beyond that, Sissie Hsiao, the VP for Google Assistant and Bard, said the team has been making “rapid improvements” to Bard’s capabilities. She specifically cited its ability to code or debug programs, since it’s now trained on more than 20 programing languages.

Google announced it’s finally removing the waitlist for its Bard AI, letting it out into the open in more than 180 countries. It should also work in both Japanese and Korean, and Hsiao said it should be able to accept around 40 languages “soon.”

Hsiao used an example where Bard creates a script in Python for doing a specific move in a game of Chess. The AI can also explain parts of its code and suggest additions to its base.

> Bard can also integrate directly into Gmail, Sheets, and Docs, able to export text directly to those programs. Bard also uses Google Search to give images and descriptions in its responses.

> Bard is also gaining connections to third-party apps, including Adobe Firefly AI image generator.

AI in Google Search

Google’s bread and butter, Search, is going to integrate the company’s AI to create a “snapshot” to users’ queries. These AI-generated summaries appear at the top of a page on the browser version of Google Search, while a set of links relating to the AI-generated text appear on the right.

Users can click to expand that view, where each line of text gets its own set of links exploring the topic in greater detail. The company said this can act as a “jumping off point” for users and their searches while still giving them access to official sources as well as users’ blogs.

The AI snapshot also works when people search for products, as it displays prices and commentary on a search. For example, a search for “good bike for a 5 mile commute with hills” will generate a few bullet points about design and motor assistance, then rank a number of different brands based on those criteria.

The system also works on mobile, through a “Converse” tab sitting alongside Video, Images, News, and so on. On Android, the snapshot takes up most of the screen while the expanded search results are pushed down to the bottom. All of users’ previous prompts remain above, and users can scroll up to find previous results. Users who continue to scroll down on the page can see more links like a regular Search results.

Google is still calling this an “experiment” with AI in Search. The availability is limited, and there’s currently a waitlist on the new Search Labs platform for those who want to access this new AI in search and to use the AI to help with coding.

Even More AI in the Workspace apps and Gmail

Google has already talked up adding AI content generation in Gmail and Google Docs, but now the company said it’s expanding the so-called “Workspace AI collaborator” to add even more generative capabilities in its cloud-based apps. These generative AI and “sidekick” features are being released on a limited basis, but will be handed out to a more broad userbase later this year.

Much like Microsoft announced with its own 365 apps earlier this year, Google’s adding generative AI into its office-style applications. Aparna Pappu, the VP of Google Workspace, said in addition to the limited deployment of a “help me write” feature on Gmail and Docs, the company is adding new AI features to several Workspace apps, including Slides and Sheets. The spreadsheet application can generate generic templates based on user prompts, such as a client list for a dog walking business.

.Google announced its new “Help me write” feature, which uses an AI to generate a full email response based on previous emails. Users can then iterate on those emails to make them more or less elaborate. Pichai used the example of a user asking customer service for a refund on a canceled flight. This is on top of long-existing content generation abilities in Gmail like Smart Compose and Smart Reply.

For Slides, users can now use text-to-image generation to add to a slideshow. The generator creates multiple instances of an image, and users can further refine those prompts with different styles.

In Gmail, the AI “sidekick” can automatically summarize an email thread and can find cite earlier documents relating to that conversation. As far as Docs goes, the existing AI is getting more elaborate. It now suggests extra prompts based on generated text, and it can now add in AI-generated images directly within Docs. This also works within Slides, letting users generate speaker notes based on AI-generated summaries of each individual slide.

New Magic Editor AI features in Google Photos

Google showcased new updates to its Magic Eraser feature, now calling a suite a AI-based editing features “Magic Editor.” Along with removing extraneous people, items, or other elements of a picture, the company said users will be able to transform objects inside each photo. The company showed how users might go about removing a bag strap or displacing an object, even the photo subject, to a different location in an image.

Gif: Google

Photos users should also be able to punch up elements of an image, such as the lighting or the clouds in the sky. Google said these updates will help those elements blend “seamlessly” into the rest of the image.

All Pixel phones should gain early access to Magic Editor later this year, though the company explicitly said things might not be as clean-looking as their promotional videos.

AI Will Soon Make Google Maps Even More Immersive

Gif: Google

With the help of Google’s new AI tools, Google Maps will soon be getting a much more elaborate Immersive View feature (which previously focused on landmarks) that relies on the billions of images that have been captured for Street View, as well as satellite imagery, to let users further explore a suggested route. It almost turns route planning into an interactive video game, giving cyclists a birds eye view of the bike lanes they’ll be relying on. Drivers will be able to preview intersections, and even the amount of parking available around their destination. The previewed route even shows what the weather will be like depending on the time of day, so you’ll not only know every last turn ahead of time, but what you should be wearing.

Immersive View for routes will be available in the coming months for cities including “Amsterdam, Berlin, Dublin, Florence, Las Vegas, London, Los Angeles, New York, Miami, Paris, Seattle, San Francisco, San Jose, Tokyo, and Venice.”

Gif: Google

On the developer side of things, Google is introducing a new “Aerial View API” allowing app makers to add more immersive points-of-interest that include detailed, 3D overhead views allowing users to see not just the specific locale, but the area around it, too. One great use of this API is for realtor or apartment hunting apps, allowing prospective buyers or renters to see the surrounding neighborhood and what amenities are nearby such as parks, or even access to major roads.

Google is also making the “high-res, 3D imagery behind Google Earth” available to developers as part of an experimental new release allowing app makers to make similarly immersive experiences, without needing to source their own satellite imagery or 3D model generation. In a blog post today, Google suggests it would potentially be a great tool for tourism, allowing 3D maps of national parks to be easily created, complete with 3D mountains, trees, and rivers, allowing potential tourists to see a destination in more detail before planning a trip, or as a way to enhance the experience of an architectural landmark, by introducing interactive overlays with interesting facts about the site and its history." 

No comments:

Fury in Russia at 'serious escalation' of Ukraine missile move | BBC News

  'Kyiv Stands': Sullivan says U.S. role in defending Ukraine defines Biden's legacy