There are few things we write about here more than artificial intelligence, and for good reason — it’s core to the future economy and a foundational offering from our firm. We exist in that space between what’s known and what’s possible, between things like mobile apps and databases and innovations like voice-based U/I or AI integrations powering a complete corporate digital transformation. But one thing we don’t always delve too deeply into is the types of artificial intelligence and what they’re good for respectively. And one such branch of AI has gotten a lot of press recently — the Generative Adversarial Networks, or GANs. They’re awesome, powerful and full of promise… and just as awesomely terrifying for what they could mean going forward.
We did write a post about whether something is actually harnessing AI recently, but it didn’t get too deep into the most common types of AI out there today, but rather clarified the need to understand what exactly AI is. So here are some of the broader strokes of AI for reference:
Now, there are ever more subsets within each of these categories, and these categories are not necessarily prescriptive for all of AI, but they are good broad strokes to define what we’re talking about.
So, machine learning and deep learning (which harnesses neural networks) are great for recognizing and/or learning patterns in/from data (which don’t have to be numbers, the data can be anything — chess moves, photos, eye-tracking results, etc.). But what if you want to create data patterns instead of just identify them?
That’s where GANs come into play.
According to the MIT Technology Review, “GANs are having a bit of a cultural moment. They are responsible for the first piece of AI-generated artwork sold at Christie’s, as well as the category of fake digital images known as ‘deepfakes‘.”
In a GAN setup, unlike a traditional AI set up, you have two neural networks working in concert (but somewhat at cross purposes). Again according to MIT Technology Review:
“You start by feeding both neural networks a whole lot of training data and give each one a separate task. The first one, known as the generator, must produce artificial outputs, like handwriting, videos, or voices, by looking at the training examples and trying to mimic it. The second, known as the discriminator, then determines whether the outputs are real by comparing each one to the same training examples.
In this way, AI networks work against each other, together, to produce something incredibly lifelike. This can be terrifying in its ability to make presidents seem to say whatever you want them to in video, and as fake news spirals out of control, it’s not hard to imagine consumers not being able to tell the difference between a deepfake and the real thing.
The good news, at least for now, is that GANs have some very real limitations that will make them hard to weaponize too soon:
“They need quite a lot of computational power and narrowly-scoped data to produce something truly believable. In order to produce a realistic image of a frog, for example, it needs hundreds of images of frogs from a particular species, preferably facing a similar direction. Without those specifications, you get some really wacky results…”
Ultimately, GANs could prove incredibly useful in fields like image generation, media post-production, medicine, etc. But, they can just as easily be used (one day) to terrifying ends of unknown scope. As with so many things AI, it’s a mixed bag heavily dependent on the intentions of its wielder.