Science fiction did a great job preparing us for submarines and rockets. But it seems to be struggling lately. We donโ€™t know what to hope for, what to fear, or what genre weโ€™re even in.

Space opera? Seems unlikely. And now that weโ€™ve made it to 2021, the threat of zombie apocalypse is receding a bit. So itโ€™s probably some kind of cyberpunk. But there are many kinds of cyberpunk. Should we get ready to fight AI or to rescue replicants from a sinister corporation? It hasnโ€™t been obvious. Iโ€™m writing this, however, because recent twists in the plot seem to clear up certain mysteries, and I think itโ€™s now possible to guess which subgenre the 2020s are steering toward.

Clearly some plot twist involving machine learning is underway. Itโ€™s been hard to keep up with new developments: from BERT (2018) to GPT-3 (2020)โ€”which can turn a prompt into an imaginary news storyโ€”to, most recently, CLIP and DALL-E (2021), which can translate verbal descriptions into images. 

Output from DALL-E. If you prefer, you can have a baby daikon radish in a tutu walking a dog.

I have limited access to DALL-E, and canโ€™t test it in any detail. But if we trust the images released by Open AI, the model is good at fusing and extrapolating abstractions: it not only knows what it means for a lemur to hold an umbrella, but can produce a surprisingly plausible โ€œphoto of a television from the 1910s.โ€ All of this is impressive for a research direction that isnโ€™t much more than four years old.

The prompt here is โ€œa photo of a television from the โ€ฆ<fill in the decade>โ€

On the other hand, some AI researchers donโ€™t believe these models are taking the field in the direction it was supposed to go. Gary Marcus and Ernest Davies, for instance, doubt that GPT-3 is โ€œan important step toward artificial general intelligenceโ€”the kind that would โ€ฆ reason broadly in a manner similar to humans โ€ฆ [GPT-3] learns correlations between words, and nothing more.โ€ 

People who want to contest that claim can certainly find evidence on the other side of the question. Iโ€™m not interested in pursuing the argument here. I just want to know why recent advances in deep learning give me a shivery sense that Iโ€™ve crossed over into an unfamiliar genre. So letโ€™s approach the question from the other side: what if these models are significant because they donโ€™t reason โ€œin a manner similar to humansโ€?

It is true, after all, that models like DALL-E and GPT-3 are only learning (complex, general) patterns of association between symbols. When GPT-3 generates a sentence, it is not expressing an intention or an opinionโ€”just making an inference about the probability of one sentence in a vast โ€œlatent spaceโ€ of possible sentences implied by its training data. 

When I say โ€œa vast latent space,โ€ I mean really vast. This space includes, for instance, the thoughts Jerome K. Jerome might have expressed about Twitter if he had lived in our century.

Mario Klingemann gets GPT-3 to extrapolate from a title and a byline.

But a latent space, however vast, is still quite different from goal-driven problem solving. In a sense the chimpanzee below is doing something more like human reasoning than a language model can.

Primates, understandably, envision models of the world as things individuals create in order to reach bananas. (Ultimately from Wolfgang Kรถhler, The Mentality of Apes, 1925.)

Like us, the chimpanzee has desires and goals, and can make plans to achieve them. A language model does none of that by itselfโ€”which is probably why language models are impressive at the paragraph scale but tend to wander if you let them run for pages.

So where does that leave us? We could shrug off the buzz about deep learning, say โ€œitโ€™s not even as smart as a chimpanzee yet,โ€ and relax because weโ€™re presumably still living in a realist novel. 

And yes, to be sure, deep learning is in its infancy and will be improved by modeling larger-scale patterns. On the other hand, it would be foolish to ignore early clues about what itโ€™s good for. There is something bizarrely parochial about a view of mental life that makes predicting a nineteenth-century writerโ€™s thoughts about Twitter less interesting than stacking boxes to reach bananas. Perhaps itโ€™s a mistake to assume that advances in machine learning are only interesting when they resemble our own (supposedly โ€œgeneralโ€) intelligence. What if intelligence itself is overrated?

The collective symbolic system we call โ€œculture,โ€ for instance, coordinates human endeavors without being itself intelligent. What if models of the world (including models of language and culture) are important in their own rightโ€”and neednโ€™t be understood as attempts to reproduce the problem-solving behavior of individual primates? After all, people are already very good at having desires and making plans. We donโ€™t especially need a system that will do those things for us. But weโ€™re not great at imagining the latent space of (say) all protein structures that can be created by folding amino acids. We could use a collaborator there.

Storytelling seems to be another place where human beings sense a vast space of latent possibility, and tend to welcome collaborators with maps. Look at whatโ€™s happening to interactive fiction on sites like AI Dungeon. Tens of thousands of users are already making up stories interactively with GPT-3. Thereโ€™s a subreddit devoted to the phenomenon. Competitors are starting to enter the field. One startup, Hidden Door, is trying to use machine learning to create a safe social storytelling space for children. For a summary of what collaborative play can build, we could do worse than their motto: โ€œWorlds with Friends.โ€

Itโ€™s not hard to see how the โ€œsocial playโ€ model proposed by Hidden Door could eventually support the form of storytelling that grown-ups call fan fiction. Characters or settings developed by one author might be borrowed by others. Add something like DALL-E, and writers could produce illustrations for their story in a variety of stylesโ€”from Arthur Rackham to graphic novel.

Will a language model ever be as good as a human author? Can it ever be genuinely original? I donโ€™t know, and I suspect those are the wrong questions. Storytelling has never been a solitary activity undertaken by geniuses who invent everything from scratch. From its origin in folk tales, fiction has been a game that works by rearranging familiar moves, and riffing on established expectations. Machine learning is only going to make the process more interactive, by increasing the number of people (and other agents) involved in creating and exploring fictional worlds. The point will not be to replace human authors, but to make the universe of stories bigger and more interconnected.

Storytelling and protein folding are two early examples of domains where models will matter not because theyโ€™re โ€œintelligent,โ€ but because they allow usโ€”their creatorsโ€”to collaboratively explore a latent space of possibility. But I will be surprised if these are the only two places where that pattern emerges. Music and art, and other kinds of science, are probably open to the same kind of exploration.

This collaborative future could be weirder than either science fiction or journalism have taught us to expect. News stories about ML invariably invite readers to imagine autonomous agents analogous to robots: either helpful servants or inscrutable antagonists like the Terminator and HAL. Boring paternal condescension or boring dread are the only reactions that seem possible within this script.

We need to be considering a wider range of emotions. Maybe a few decades from now, autonomous AI will be a reality and weโ€™ll have to worry whether itโ€™s servile or inscrutable. Maybe? But thatโ€™s not the genre weโ€™re in at the moment. Machine learning is already transforming our world, but the things that should excite and terrify us about the next decade are not even loosely analogous to robots. We should be thinking instead about J. L. Borgesโ€™ Library of Babelโ€”a vast labyrinth containing an infinite number of books no eye has ever read. There are whole alternate worlds on those shelves, but the Library is not a robot, an alien, or a god. It is just an extrapolation of human culture.

Eric Desmazieres, โ€œThe Library of Babel.โ€

Machine learning is going to be, letโ€™s say, a thread leading us through this Libraryโ€”or perhaps a door that can take us to any bookshelf we imagine. So if the 2020s are a subgenre of SF, I would personally predict a mashup of cyberpunk and portal fantasy. With sinister corporations, of course. But also more wardrobes, hidden doors, encylopedias of Tlรถn, etc., than weโ€™ve been led to expect in futuristic fiction.

Iโ€™m not saying this will be a good thing! Human culture itself is not always a good thing, and extrapolating it can take you places you donโ€™t want to go. For instance, movements like QAnon make clear that human beings are only too eager to invent parallel worlds. Armored with endlessly creative deepfakes, those worlds might become almost impenetrable. So weโ€™re probably right to fear the next decade. But letโ€™s point our fears in a useful direction, because we have more interesting things to worry about than a servant who refuses to โ€œopen the pod bay doors.โ€ We are about to be in a Borges story, or maybe, optimistically, the sort of portal fantasy where heroines create doors with a piece of chalk and a few well-chosen words. I have no idea how our version of that story ends, but I would put a lot of money on โ€œnot boring.โ€

http://www.tedunderwood.com
Contributor

Recently Published

Key Takeaway: Leading scientists and technologists often make terrible predictions about the direction of innovation, leading to misalignments between a company’s economic incentives to profit from its proprietary AI model and society’s interests in how the AI model should be monetised and deployed. Focusing on the economic risks from AI is not just about preventing […]

Top Picks

Key Takeaway: Quantum computing, which uses entanglement to represent information, has the potential to revolutionize everyday life. However, the development of quantum computers has been slow due to the need to demonstrate an advantage over classical computers. Only a few notable quantum algorithms have been developed, such as the BB84 protocol and Shor’s algorithm, which […]
Key Takeaway: China’s leaders have declared a GDP growth target of 5% in 2024, despite facing economic problems and a property crisis. The country’s rapid economic growth has been attributed to market incentives, cheap labor, infrastructure investment, exports, and foreign direct investment. However, none of these drivers are working effectively. The government’s determination to deflate […]
Key Takeaway: Neuralink, founded by Elon Musk, aims to implant a brain-computer interface (BCI) in people’s brains, allowing them to control computers or phones by thought alone. This technology holds the promise of alleviating human suffering and allowing people with disabilities to regain lost capacities. However, the long-term aspirations of Neuralink include the ability to […]

Trending

I highly recommend reading the McKinsey Global Instituteโ€™s new report, โ€œReskilling China: Transforming The Worldโ€™s Largest Workforce Into Lifelong Learnersโ€, which focuses on the countryโ€™s biggest employment challenge, re-training its workforce and the adoption of practices such as lifelong learning to address the growing digital transformation of its productive fabric. How to transform the country […]

Join our Newsletter

Get our monthly recap with the latest news, articles and resources.

Login

Welcome to Empirics

We are glad you have decided to join our mission of gathering the collective knowledge of Asia!
Join Empirics