The evolution timeline depicting man evolving from a primate to a human to a walking robot

The Evolution of AI

by Jon Chun and Katherine Elkins

Generative AI is a transformative force, reshaping both arts and humanities computing. Its recent evolution retraces our own human evolution, only on a vastly accelerated scale.

In the Beginning Was the Word

Working with language seemed like it would be a simple task for computers. Given how easily machines compute complex math, how hard could it be? The answer turned out to be very.

Traditional or Good Old Fashioned AI (GOFAI) focused on statistical patterns that required models carefully crafted by experts. The advent of Deep Learning, powered by advancements like Moore’s Law and big data, enabled a paradigm shift that eliminated the need for explicit rules.

A chart depicting the GOFAI model and the Deep Learning model
Good Old Fashioned AI (GOFAI) vs Deep Learning

With this new paradigm, some argued these Large Language Models (LLMs) only simulate understanding, much like “stochastic parrots.” 

Applied at scale, however, Deep Learning LLMs exhibit complex emergent properties, just as human biology arises from simple DNA. With scale, new capabilities appear, from summarization to theory of mind.

Exploring the World

Could true intelligence arise from a disembodied “brain in a vat”? How much of our world can be captured by the words on which LLMs are trained? It turns out, more than we thought. ChatGPT, for instance, has demonstrated the ability to generate spatial maps.

But to realize general intelligence, AI agents need embodiment to explore unconstrained environments and learn firsthand. This entails multi-modal AI capable of learning through multiple inputs like sight and sound. Today, AI robots are beginning to explore and interact with the world in just this way.

Mastery of Tools

LLMs, like humans, also have intrinsic limitations. Like us, their evolution leverages tool use and creation to overcome these. Current integration with external databases and search engines reduces so-called “hallucinations.” Plug-ins and APIs enable LLMs to write programs, send email, and manipulate complex symbolic models.

Just as animal intelligence exploded after mastering tools, this stage could unlock an exponential takeoff in intelligence. With tools, AI could more easily engineer its own self-improvement and augment its capabilities beyond human comprehension.

AI Gets Creative (and Emotional)

AI now ventures into creative domains: art, music, and writing. Until recently, automation had largely replaced manual labor characterized as dirty and dangerous, difficult and dull.

Generative AI can write and design, analyze and program. Intellectual property lawsuits and Hollywood writers’ strikes are only isolated examples of looming disruptions.

Alongside these creative powers comes emotional intelligence. Emotional Intelligence and AffectiveAI delve into recognizing, modeling, and influencing human emotions, forging a path toward more empathetic AI systems.

Already, our connected devices feed data to hungry AffectiveAI models, and AI may soon know us better than we know ourselves. Once algorithms excel at modeling collective psychology, we may find our minds and behaviors shaped by forces beyond our awareness.

AI as Social Agent Collectives

Just as social collaboration propelled human civilization forward, a society of collaborative AI agents can orchestrate complex tasks with a symphony of specialized skills. Rather than one monolithic model, the latest natural language model, GPT-4, is likely a coordinated mixture of experts.

Some theorists envision a future in which humans directly interface with AI collectives through brain-computer interfaces. More immediate questions involve how to align AI agents with human values.

Conclusion

Addressing the Human-AI value alignment problem is crucial to ensure the AIs we develop are in service to humanity. It’s a pivotal moment for those of us in humanities and arts computing to have a voice in shaping our AI future.

As the articles in this special volume on times of crisis make manifest, part of this influence lies in educating the next generation. There are a variety of interconnected questions that span the range of human experience and organization, and to study any of these in isolation is insufficient.

But we also have a role to play in the coming conversations about the future of work, the limitations we may wish to place on artificial agents, and the ethical, legal and societal questions we now face. To ensure that the AIs we develop are in service to humanity, we must first solve the human-human alignment dilemma. Before we decide on values for our machines, we must first agree, however much we are able, on which values are shared by humans.

We are grateful for journals like the International Journal Humanities and Arts Computing that provide forums to consider these questions. Arts and humanities computing is having a moment, and there is no better time, with the spotlight upon us, to do something with the moment that’s been given us.


IJHAC cover
IJHAC

About the journal

IJHAC is one of the world’s premier multi-disciplinary, peer-reviewed forums for research on all aspects of arts and humanities computing. The journal focuses both on conceptual and theoretical approaches, as well as case studies and essays demonstrating how advanced information technologies can further scholarly understanding of traditional topics in the arts and humanities.

Sign up for TOC alerts,  recommend to your library, learn how to submit an article and subscribe to IJHAC.


About the authors

Headshot of Jon Chun
Jon Chun

Jon Chun is co-founder of the world’s first human-centered AI curriculum at Kenyon College, where he also teaches. He is an engineer, computer scientist and former Silicon Valley Startup CEO who has worked in medicine, finance, and network security in the US, Asia, and Latin America. His most recent AI research focuses on NLP, multimodal AffectiveAI, LLMs and XAI/FATE. He also consults for both industry and startups.

Headshot of Katherine Elkins
Katherine Elkins

Katherine Elkins is Professor of Humanities and faculty in Computing at Kenyon College, where she teaches in the Integrated Program in Humane Studies. She writes about the age-old conversation between philosophy and literature as well as the more recent conversation about AI, language and art. Recent books include The Shapes of Stories with Cambridge University Press, which explores the emotional arc underlying narrative. In addition to articles on authors ranging from Plato and Sappho to Woolf and Kafka, her latest research explores AI ethics and explainability in Large Language Models like GPT4. She consults on AI regulation and fairness, and can also be heard talking about new developments in AI on podcasts like Radio AI and streaming networks like Al Jazeera.

Don’t forget to sign up to our mailing list to keep up to date with all of our free content and latest releases!

Edinburgh University Press
Edinburgh University Press
Articles: 137

Leave a Reply

Your email address will not be published. Required fields are marked *