Resilient Models
Resilience, adaptability and a high-rate of learning are not only values projected onto individuals, they also act as benchmarks against which the training success of AI models themselves can then be measured. Former OpenAI CTO Mira Murati delineated how users can participate in the “collective intelligence” of ChatGPT: by releasing the model to the public at an early stage, there is time for “society to adapt”, and for the model to learn from “the friction” with “reality”. Learning from this “friction” suggests that perturbations and volatility function as a necessary source of learning for the model, as it feeds it with the complexities, non-linearities and ambiguities of human and “world” behavior. Of course, the “reality” that Murati hints at, is not limited to the “real world” as such, but takes place in all of the many shades in-between reality and the synthetic and at differing grades of complexity.
It is not a coincidence that AI companies have focused so adamantly on images, videos, text and other data: not just because they are easily accessible on the web in large quantities, or because they are central to a culture where knowledge is seen as largely visual and textual, but also because they are centered around objects, rather than processes. The generation of images, artworks and text do not require the model to have learned the processes that lead up to their creation, but only require it to imitate the results of them. Instead, when Murati speaks about the friction with the real world, she is speaking about training the model on the complexities that emerge out of interaction, which, one could speculate, is why OpenAI is collecting data on all “conversations” with the model by default.
The distinction between objects and processes becomes much clearer with the training necessary for embodied and physicalised AI, which require the ability to “capture the indeterminacies of the real”. To train AI for such processes, simulated training environments such as OpenAI Gym or GoogleDeepMind Lab were developed and made publically available. In these simulations, AI agents are trained on simulated scenes and in competition, cooperation or co-existence with other non-human and human agents, in scenes where supposedly “complexity emerges from end-to-end training in a rich environment.”
In their video essay “Vivarium”, researchers from the think-tank Antikythera propose a future virtual training ground and simulation engine for embodied AI, where AI models are trained in “toy worlds” in various human-AI configurations, encouraging “learning from basic movement to interactive negotiation, adversarial feints to stigmergic coordination”. Nonetheless, the researchers also refer to the phenomenon of the Sim2Real Gap, which describes the issues that arise when attempting to transfer capabilities learned in a simulated environment to the “real world”. Complexity in AI goes both ways: volatile and complex data is necessary for training purposes, but it also poses a limit for what models can adapt to: training embodied AI in the real world is not only extremely expensive, but is also “impossible to parallelize, and difficult to control.” Nonetheless, the idea of training models directly in the real world is a recurring one: in 2018 an MIT Technology Review Column proposed that India would not only need an AI revolution to push forward its economy, but also that “India’s mess of complexity is just what AI needs”. India, as the author argues, could function as an ideal training ground on which AI could “mature” and which would make it “more resilient”.
As Luciana Parisi argues with reference to Gilles Deleuze, AI models as a “machine ecology infected with randomness” not only represent “an interactive system of learning and continuous adaptation”, but also brings forth a specific “logic of governance driven by the variable mesh of continuous variability”. More recently, Louise Amoore proposed a machine learning political order, a specific style of governance that is grounded in the transformation towards the “productive generation of turbulence and division from which algorithmic systems are derived.” Importantly, this machine learning political order is not the result of a “causal relationship where ideas from computer science bleed into the state and sovereign logics”, nor is it the same as considering AI models as political decision-makers or describing the automation of previously human governmental processes. Instead, it becomes relevant to investigate the broader “epistemic and political transformations” and the generation of “new norms and thresholds”. In this sense, this order is enframed by a much older shift in world picture from rule-based orders to a world that operates at the edge of chaos and which makes it possible, according to Amoore, to profit from “the volatilities of fractured disorder”.