30 March 2023

Unbelievably realistic fake images could take over the internet

 Take the time to read it > 

How unbelievably realistic fake images could take over the internet

AI image generators like DALL-E and Midjourney are getting better and better at fooling us.

Sara Morrison is a senior Vox reporter who covers data privacy, antitrust, and Big Tech’s power over us all.

"Last week, a 31-year-old construction worker took a few psychedelics and thought it might be fun to use AI image generator Midjourney to create a photorealistic image of Pope Francis wearing a big white Balenciaga-style puffer jacket. A lot of people who saw it thought it was fun, too, so they spread it around social media. Most of them probably had no idea that it wasn’t real.

Now, the Pope having that drip isn’t the worst nor most dangerous deployment of photorealistic AI-generated art, in which new images are created from text prompts. But it is an example of just how good this technology is becoming, to the point that it can even trick people who are usually more discerning about spreading misinformation online. You might even call it a turning point in the war against mis- and disinformation, which the people fighting were, frankly, already losing simply because social media exists. Now we have to deal with the prospect that even the people who are fighting that war may inadvertently help spread the disinformation they’re trying to combat. And then what?

It’s not just Coat Pope. In the last two weeks, we’ve seen several ominous AI-image stories. 

We had Trump’s fake arrest and attempted escape from the long AI-generated arm of the law, which was capped by a set of poorly rendered fingers. 

We had Levi’s announcing it would “supplement” its human models with AI-generated ones in the name of diversity (hiring more diverse human models was apparently not an option). 

Microsoft unleashed its Bing Image Creator in its new AI-powered Bing and Edge browser, and Midjourney, known for its photorealistic images, released its latest version.

Finally, there’s the news that AI image generators are getting better at drawing hands, which had been one of the tell-tale signs to detect if an image is fake. Even as convincing as Coat Pope appeared, a close look at his right hand would have revealed its AI origins. But soon, we may not even have that. Levi’s will be able to use AI models to show off its gloves, while the rest of us might be thrown into a new world where we have absolutely no idea what we can trust — one that’s even worse than the world we currently inhabit.

“We’ve had this issue with text and misinformation on social platforms. People are conditioned to be skeptical with text,” said Ari Lightman, a professor of digital media and marketing at Carnegie Mellon University. “An image ... adds some legitimacy in the user’s mind. An image of video creates more resonance. I don’t think our blinders are up yet.”

In just a few short years, AI-generated images have come a long way. In a more innocent time (2015) Google released “DeepDream,” which used Google’s artificial neural network programs — that is, artificial intelligence that’s been trained to learn in a way that mimics a human brain’s neural networks — to recognize patterns in images and make new images from them. You’d feed it an image, and it would spit back something that resembled it but with a bunch of new images weaved in, often things approximating eyeballs and fish and dogs. It wasn’t meant to create images so much as to show, visually, how the artificial neural networks detected patterns. The results looked like a cross between a Magic Eye drawing and my junior year of college. Not particularly useful in practice, but pretty cool (or creepy) to look at.

These programs got better and better, training on billions of images that were usually scraped from the internet without their original creators’ knowledge or permission. In 2021, OpenAI released DALL-E, which could make photorealistic images from text prompts. It was a “breakthrough,” says Yilun Du, a PhD student at MIT’s Computer Science and Artificial Intelligence Laboratory who studies generative models. Soon, not only was photorealistic AI-generated art shockingly good, but it was also very much available. OpenAI’s Dall-E 2, Stability AI’s Stable Diffusion, and Midjourney were all released to the general public in the second half of 2022.

The expected ethical concerns followed, from copyright issues to allegations of racist or sexist bias to the possibility that these programs could put a lot of artists out of work to what we’ve seen more recently: convincing deepfakes used to spread disinformation. And while the images are very good, they still aren’t perfect. But given how quickly this technology has advanced so far, it’s safe to assume that we’ll soon be hitting a point where AI-generated images and real images are nearly impossible to tell apart.

Take Nick St. Pierre’s work, for example. St. Pierre, a New York-based 30-year-old who works in product design, has spent the last few months showing off his super-realistic AI art creations and explaining how he got them. He may not have the artistic skills to compose these images on his own, but he has developed a skill for getting them out of Midjourney, which he says he uses because he thinks it’s the best one out there. St. Pierre says he dedicated the month of January to 12-hour days of working in Midjourney. Now he can create something like this in just about two hours.

“When you see a digital image on the internet and it’s AI generated, it can be cool, but it doesn’t, like, shock you,” St. Pierre said. “But when you see an image that’s so realistic and you’re like, ‘wow, this is a beautiful image’ and then you realize it’s AI? It makes you question your entire reality.”

...READ MORE

No comments:

IDK

  Why Is Trump Denying He Banged a Porn Star? STORMY SEASON The former president claims all he did was pose for a photo with Stormy Daniels....