AI shares with earlier socially transformative technologies a reliance on limiting models of the “human” that embed racialized metrics for human achievement, expression, and progress. Many of these fundamental mindsets about what constitutes humanity have become institutionally codified, continuing to mushroom in design practices and research development of devices, applications, and platforms despite the best efforts of many well-intentioned technologists, scholars, policy-makers, and industries. This essay argues why and how AI needs to be much more deeply integrated with the humanities and arts in order to contribute to human flourishing, particularly with regard to social justice. Informed by decolonial, disability, and gender critical frameworks, some AI artist-technologists of color challenge commercial imperatives of “personalization” and “frictionlessness,” representing race, ethnicity, and gender not as normative self-evident categories nor monetized data points, but as dynamic social processes always indexing political tensions and interests.
This page was updated on May 3, 2022. To view the original essay published on April 13, 2022, click here.
As he grew accustomed to the great gallery of machines, he began to feel the forty-foot dynamos as a moral force, much as the early Christians felt the Cross. The planet itself felt less impressive, in its old-fashioned, deliberate, annual or daily revolution, than this huge wheel, revolving within arm’s length at some vertiginous speed, and barely murmuring–scarcely–humming an audible warning to stand a hair’s breadth further for respect of power, while it would not wake the baby lying close against its frame. Before the end, one began to pray to it; inherited instinct taught the natural expression of man before silent and infinite force. Among the thousand symbols of ultimate energy the dynamo was not so human as some, but it was the most expressive.
—Henry Adams, “The Virgin and the Dynamo”1
In astonishment of the new technologies at the turn into the twentieth century, the renowned historian Henry Adams found the Gallery of the Electric Machines “physics stark mad in metaphysics” and wondered at their profound hold on the cultural imagination.2 The dynamo that so moved and unsettled Adams was a new generator of unprecedented scale, a machine responsible for powering the first electrified world’s fair in 1893, a purportedly spectacular event presided over by President Glover Cleveland. Its power was invisible but the more potent for it: “No more relation could he discover between the steam and the electric current than between the cross and the cathedral. The forces were interchangeable if not reversible, but he could see only an absolute fiat in electricity as in faith.” For Adams, the dynamo’s effect in the world was akin to evidence of things unseen like the symbols of the Virgin or the cross, imperceptible but world-transforming currents with implications both worldly and spiritual.
I open with this discussion of the world’s fair at the fin de siècle because Adams’s dynamo is our GPT-3 (Generative Pre-trained Transformer 3), a language model that uses deep learning to produce text/speech/responses that can appear generated by a human. His exhilaration–hand-in-glove with his existential vertigo–and his internal conflict similarly speak to our contemporary aspirations for and anxieties about artificial intelligence. Adams understood that the turn to such formidable technology represented a thrilling but cataclysmic event, “his historical neck broken by the sudden irruption of forces entirely new.” Although human grappling with exponential leaps in technology dates at least to the medieval period, this particular historical precedent of a transformational moment is singularly relevant for our contemporary moment: there’s a direct line between Adams’s concern with the hagiography of tech, the devaluation of the arts and humanities, and the comingling of scientific development with (racialized, ableist) narratives of progress to current debates about those nearly identical phenomena today. The consequences of those fundamental mindsets and practices, institutionally codified over time, continue to mushroom in devices, applications, platforms, design practices, and research development. Unacknowledged or misunderstood, they will continue to persist despite the best efforts of many well-intentioned technologists, scholars, policy-makers, and industries that still tend to frame and limit questions of fairness and bias in terms of “safety,” which can mute or obscure attention to issues of equity, justice, or power.3
Significantly, Adams’s response to the dynamo is neither apocalyptic jeremiad nor in the genre of salvation: that is, his concerns fell beyond the pale of narratives of dystopia or deliverance. He was no technophobe; in fact, he deeply admired scientific advances of all kinds. Rather, his ambivalence has to do with the inestimable psychological and spiritual sway of machines so impressive that “the planet itself felt less impressive,” even “old-fashioned.”4 That something man-made might seem so glorious as to overshadow creation, seemed so evocative of the infinite that people felt out of step with their own times. For Adams, those experiences signaled an epistemic break that rendered people especially receptive and open to change, but also vulnerable to idolizing false gods of a sort. He saw that the dynamo was quickly acquiring a kind of cult status, inviting supplication and reverence by its followers. The latest technology, as he personified it in his poem “Prayer to the Dynamo,” was simultaneously a “Mysterious Power! Gentle Friend! Despotic Master! Tireless Force!”5 Adams experienced awe in the presence of the dynamo: “awe” as the eighteenth-century philosopher Edmund Burke meant the term, as being overcome by the terror and beauty of the sublime. And being tech awestruck, he also instantly presaged many of his generation’s–and I would argue, our generation’s–genuflection before it.
As part of his concern that sophisticated technology inspires a kind of secular idolatry, Adams also noted its increasing dominance as the hallmark of human progress. In particular, he presciently anticipated that it might erode the power of both religion and the arts as vehicles for and markers of humanity’s higher strivings. Indeed, his experience at the Gallery taught him firsthand how fascination with such potent technology could eclipse appreciation of the arts: more specifically, of technological innovation replacing other modes of creative expression as the pinnacle of human achievement. Adams bemoaned the fact that his friend, Langley, who joined him at the exposition, “threw out of the field every exhibit that did not reveal a new application of force, and naturally, to begin with, the whole art exhibit.” The progress of which technology increasingly claimed to be the yardstick extended beyond the valuation of art also extended to racial, ethnic, and gender scales. Most contemporary technological development, design, and impact continue to rely unquestioningly on enlightenment models of the “human,” as well as the nearly unchanged and equally problematic metrics for human achievement, expression, and progress.
These are not rhetorical analogies; they are antecedences to AI, historical continuities that may appear obscured because the tech-ecosystem tends to eschew history altogether: discourses about AI always situate it as future-facing, prospective not retrospective. It is an idiom distinguished by incantations about growth, speed, and panoptic capture. The messy, recursive, complex narratives, events, and experiences that actually make up histories are reduced to static data points necessary in training sets for predictive algorithms. Adams’s reaction offers an alternative framing of time in contrast to marketing imperatives that fetishize the next new thing, which by definition sheds its history.
This reframing is important to note because for all the contemporary talk of disruption as the vaulted and radical mode of innovation, current discourse still often presents so-called disruptive technologies as a step in an inexorable advance forward and upward. In that sense, tech disruption is in perfect keeping with the same teleological concept of momentum and progress that formed the foundational basis by which world’s fairs ranked not only modes of human achievement but also degrees of “human.” The exhibitions catalogued not just inventions but people, classifying people by emerging racialized typologies on a hierarchical scale of progress with the clear implication that some were more human than others.6 This scale was made vivid and visceral: whether it was the tableaux vivant “ethnic villages” of the 1893 world’s fair in Chicago’s “White City” or the 1900 Paris showcase of African American achievement in the arts, humanities, and industry (images of “racial uplift” meant to counter stereotyping), both recognized how powerfully influential were representations of races’ putative progress–or lack of it.
Carrying the international imprimatur of the fairs, the exhibitions were acts of racial formation, naturalizing rungs of humanness and, indeed, universalizing the imbrication of race and progress. Billed as a glimpse into the future, the fairs simultaneously defined what was not part of modernity: what or who was irrelevant, backward, regressive in relation. Technological progress, therefore, was not simply represented alongside what (arts/humanities) or who (non-whites) were considered less progressive; progress was necessarily measured against both, indeed constituted by its difference and distance from both.
For critical theorist Homi Bhabha, such notions of progress, and the technology and symbol of it, are inextricably tied to the exercise of colonial and cultural power. His essay “Signs Taken for Wonders: Questions of Ambivalence and Authority Under a Tree outside Delhi, May 1817” critiques the “wondrous” presence of the book, itself a socially transformative technology, by beginning with the premise that innovation cannot be uncoupled from the prerogatives of those who have the power to shape realities with it:
The discovery of the book is, at once, a moment of originality and authority, as well as a process of displacement, that paradoxically makes the presence of the book wondrous to the extent to which it is repeated, translated, misread, displaced. It is with the emblem of the English book–“signs taken as wonders”–as an insignia of colonial authority and an insignia of colonial desire and discipline that I begin this essay.7
Adams spoke of awe in the presence of the dynamo. Bhabha goes further in challenging such “signs taken as wonders,” in questioning technologies so valorized that they engender awe, obeyance, and reverence as if such a response was natural, innocent of invested political and economic interests, free of market value systems.
Like all tools, AI challenges the notion that the skull marks the border of the mind. . . . New tools breed new literacies, which can engender nascent forms of knowing, feeling and telling.
—Vanessa Chang, “Prosthetic Memories, Writing Machines”8
Art sits at the intersection of technology, representation, and influence. Literature, film, music, media, and visual and graphic arts are all crucial incubators for how publics perceive tech. Storytelling impacts, implicitly or explicitly, everything from product design to public policy. Many of these narratives bear traces of literature’s earliest engagement with technology, at least since medieval times, and others–either engaged with AI or AI-enabled–are also offering new plotlines, tropes, identity formations, historiographies, and speculative futurities. Moreover, because cultural storytelling helps shape the civic imagination, it can, in turn, animate political engagement and cultural change.9
Indeed, the arts are specially poised to examine issues in technological spaces (from industry to STEM education) of equity, diversity, social justice, and power more capaciously and cogently than the sometimes reductive industry-speak of inclusion, fairness, or safety (usually simply meaning minimization of harm or death–a low bar indeed). Even before GPT-3, powerful natural language processing was enabling explorations in AI-assisted poetry, AI-generated filmscripts, AI-informed musicals, AI-advised symphonies, AI-curated art histories, and AI-augmented music.10 Many are proposing new nomenclature for hybrid genres of art, design, and tech, and fresh subfields are blooming in both academe and entertainment.11 And during the COVID-19 pandemic and intensified movements for social justice, there has been a plethora of virtual exhibitions and articles about the hot debates over the status, meaning, and valuation of AI-generated or -augmented art.12
Amidst this explosion of artistic engagement with AI, social and political AI scholars Kate Crawford and Luke Stark, in “The Work of Art in the Age of Artificial Intelligence: What Artists Can Teach Us about the Ethic of Data Practice,” offer a not uncommon perspective on the need for interdisciplinary collaboration: “Rather than being sidelined in the debates about ethics in artificial intelligence and data practices more broadly, artists should be centered as practitioners who are already seeking to make public the political and cultural tensions in using data platforms to reflect on our social world.”13 However, they also close the article by recommending that arts practitioners and scholars would do well with more technical education and that without it, their engagements and critiques will have lesser insight into and standing regarding the ethics of data practice: “One barrier to a shared and nuanced understanding of the ethical issues raised by digital art practices is a lack of literacy regarding the technologies themselves Until art critics engage more deeply with the technical frameworks of data art, their ability to analyze and assess the merits of these works–and their attendant ethical dilemmas–may be limited.” They continued: “a close relationship to computer science seemed to offer some artists a clearer lens through which to consider the ethics of their work.”14
Certainly, continuing education is usually all to the good. But I would welcome the equivalent suggestion that those in data science, computer science, engineering, and technology, in turn, should continue to educate themselves about aesthetics and arts practices–including at least a passing familiarity with feminist, queer, decolonial, disability, and race studies approaches to AI often central to those practices–to better understand ethical debates in their respective fields.15 Without that balance, the suggestion that artists and nontechnical laypeople are the ones who primarily need education, that they require technical training and credentialing in order to have a valid(ated) understanding of and legitimate say in the political, ethical, social, and economic discussions about AI, is a kind of subtle gatekeeping that is one of the many often unacknowledged barriers to cross-disciplinary communication and collaboration. Given the differential status of the arts in relation to technology today, it is usually taken for granted that artists (not technologists, who presumably are doing more important and time-consuming work in and for the world) have the leisure and means not only to gain additional training in other fields but also to do the hard translational work necessary to integrate those other often very different disciplinary practices, vocabularies, and mindsets to their own creative work. That skewed status impacts who gains the funding, influence, and means to shape the world.
Instead of asking artists to adapt to the world models and pedagogies informing technological training–which, as with any education, is not simply the neutral acquisition of skills but an inculcation to very particular ways of thinking and doing–industry might do well to adapt to the broader vernacular cultural practices and techne of marginalized Black, Latinx, and Indigenous communities. Doing so might shift conversation in the tech industry from simply mitigating harm or liability from the differentially negative impact of technologies on these communities. Rather, it would require a mindset in which they are recognized as equal partners, cultural producers of knowledge(s), as the longtime makers, not just the recipients and consumers, of technologies.16 In fact, artist-technologist Amelia Winger-Bearskin, who is Haudenosaunee (Iroquois) of the Seneca-Cayuga Nation of Oklahoma, Deer Clan, makes a case that many of these vernacular, often generational, practices and values are what she calls “antecedent technologies,” motivated by an ethic that any innovation should honor its debt to those seven generations prior and pay it forward seven generations.17
In this way, many contemporary artist-technologists engage issues including, but also going beyond, ethics to explore higher-order questions about creativity and humanity. Some offer non-Western or Indigenous epistemologies, cosmologies, and theologies that insist on rethinking commonly accepted paradigms about what it means to be human and what ways of doing business emerge from that. Perhaps most profoundly, then, the arts can offer different, capacious ways of knowing, seeing, and experiencing worlds that nourish well-being in the now and for the future. It is a reminder of and invitation to world models and frameworks alternative to what can seem at times to be dominating or totalizing technological visions. In fact, one of the most oft-cited criticisms of AI discourse, design, and application concerns its univision, its implied omniscience, what scholar Alison Adams calls “the view from nowhere.” It is challenged by art that offers simultaneous, multiple, specifically situated, and sometimes competing points of view and angles of vision that enlarge the aperture of understanding.18
For instance, informed by disability culture, AI-augmented art has drawn on GANs (generative adversarial networks) to envision non-normative, including neurodivergent, subjects that challenge taken-for-granted understandings of human experience and capability. The presumption of a universal standard or normative model, against which “deviance” or “deviation” is measured, is nearly always implied to be white, cis-gendered, middle-classed, and physically and cognitively abled. That fiction of the universal subject–of what disability scholar and activist Rosemarie Garland-Thomson terms the “normate”–has historically shaped everything from medical practice and civil rights laws to built environments and educational institutions. It also often continues to inform technologies’ development and perceived market viability and use-value. Representations of “human-centered” technology that include those with mental or physical disabilities often call for a divestment from these usual ways of thinking and creating. Such a direct critique is posed in art exhibitions such as Recoding CripTech. As the curatorial statement puts it, the installations reimagine “enshrined notions of what a body can be or do through creative technologies, and how it can move, look or communicate. Working with a broad understanding of technology . . . this multidisciplinary community art exhibition explores how disability–and artists who identify as such–can redefine design, aesthetics and the relationship between user and interface.” Works included in Recoding CripTech that employ artificial intelligence, such as M Eifler’s “Prosthetic Memory” and “Masking Machine,” suggest a provocative reframing of “optimization” or “functionality” in technologies that propose to augment the human experience.19
Race–racism–is a device. No More. No less. It explains nothing at all. . . . It is simply a means. An invention to justify the rule of some men over others. [But] it also has consequences; once invented it takes on a life, a reality of its own. . . . And it is pointless to pretend that it doesn’t exist–merely because it is a lie!
—Tshembe in Les Blancs (1965) by Lorraine Hansberry
Rashaad Newsome’s installation Being represents another artistic provocation that reframes both the form and content of traditional technological historiographies often told from that “view from nowhere.” Newsome, a multimedia artist and activist, makes visible the erased contributions to technology and art by people of African descent. Newsome terms the interactive social humanoid Being 2.0 an AI “griot,” a storyteller. But unlike most social robots commanded to speak, Being is intentionally “uppity”: wayward, noncompliant, disobedient, with expressive gestures drawn Black Queer vogue dance repertoire meant as gestures of decolonial resistance to the labor and service that social robots are expected to perform. It upends the historical association of robots and slaves (in the etymology of the Czech word, “robot” translates to “slave”) in movement, affect, function, and speech. Taking aim at the limited training data sets used in natural language processing, Newsome draws on broader archives that include African American vernacular symbolic systems.20 And since language carries cultural knowledge, Being’s speech expands not just vocabularies but reimagines how the standardized expressions of emotion and behavior often deployed in AI are racially and culturally encoded.21 In fact, Being is an attempt to redress the historical violence of antiquated notions about race, the more disturbing because the representations of race, reduced to seemingly self-evident graduations of color and physiognomy, are being actively resurrected in AI development and application.
Race is always a negotiation of social ascription and personal affirmation, a process of what sociologists Michael Omi and Howard Winant term “racial formation.” Omi and Winant refer to racial formation as a way of historicizing the practices and circumstances that generate and renew racial categories and racializing structures:
We define racial formation as the sociohistorical process by which racial categories are created, inhabited, transformed, and destroyed. . . . Racial formation is a process of historically situated projects in which human bodies and social structures are represented and organized. Next we link racial formation to the evolution of hegemony, the way in which society is organized and ruled. . . . From a racial formation perspective, race is a matter of both social structure and cultural representation.22
The expression “racial formation” is therefore a reminder that race is not a priori. It is a reminder to analyze the structural and representational–not just linguistic–contexts in which race becomes salient: the cultural staging, political investments, institutional systems, and social witnessing that grant meanings and values to categories. A full accounting of race therefore involves asking in whose interest is it that a person or people are racialized in any given moment in time and space?
Overlooked, for instance, in many debates over racial bias, surveillance, and privacy in facial recognition technology is the practice of coding “race” or “ethnicity” as fixed, static programmable variables, something writ on the face or otherwise available as physically intelligible–an outdated approach to race that harkens back to nineteenth-century phrenology and other pseudoscience mappings of racial traits. Moreover, that practice renders opaque how categories are never merely descriptive, disinterested renderings of facts or things even though they cannot be purged of the value systems that animate their creation and make them intelligible for technological use–at least as currently developed–in the first place. Additionally, the claim to a universal objectivity is one of the “epistemic forgeries,” according to Yarden Katz, who describes it as one of the “fictions about knowledge and human thoughts that help AI function as a technology of power” because it enables “AI practitioners’ presumption that their systems represent a universal ‘intelligence’ unmarked by social context and politics.”23 That drive for comprehensive typing and classification, for a universal compendium, cannot easily accommodate race other than a technical problem in mapping variation of types.24
To illustrate why AI representations are so problematic, let me take a seemingly innocuous example in the new algorithmic application “Ethnicity Estimate,” part of the Gradient app, which purports to diagnose percentages of one’s ethnic heritage based on facial recognition technology (FRT). Such an app is significant precisely because popular data-scraping applications are so often pitched as convenient business solutions or benign creative entertainment, bypassing scrutiny because they seem so harmless, unworthy of research analysis or quantitative study. Critically examining on such issues would be a direct impediment to a seamless user experience with the product, thus designers and users are actively disincentivized from doing so. Like many such applications, Ethnicity Estimate problematically uses nationality as a proxy for ethnicity and reduces population demographics to blood quantum.
Or consider Generated Photos: an AI-constructed image bank of “worry-free” and “infinitely diverse” facial portraits of people who do not exist in the flesh, which marketers, companies, and individuals can use “for any purpose without worrying about copyrights, distribution rights, infringement claims or royalties.”26 In creating these virtual “new people,” the service offers a workaround for privacy concerns. Generated Photos bills itself as the future of intelligence, yet it reinscribes the most reductive characterizations of race: among other parameters users can define when creating the portraits, such as age, hair length, eye color, and emotion through facial expression, the racial option has a dropdown of the generic homogenizing categories Asian, African American, Black, Latino, European/white.
Skin color options are similarly presented as self-evident and unproblematic givens, a data set based on an off-the-shelf color chart. There is a long racializing history of such charts, from the von Luschan chromatic scale, used throughout the first half of the twentieth century to establish racial classifications, to the Fitzpatrick scale, still common in dermatologists’ offices today, which classifies skin types by color, symbolized by six smiling emoji modifiers. Although the latter makes no explicit claim about races, the emojis clearly evoke the visuals well as the language of race with the euphemism of “pigmentary phototype.”
All these types are readily serviceable as discrete data points, which makes them an easy go-to in algorithmic training, but the practice completely elides the fact that designations of “dark” or “light” are charged cultural and contextual interpretations that are always negotiated in context and in situ.25 The relevance and meaning of race emerge through social and cultural relations, not light frequencies. Fran Ross’s brilliant, satirical novel Oreo (1974) offers a wry send-up of attempts to apply color charts to social identities, shown as Figure 6.26
Although new AI technologies show promise in diagnosing medical conditions of the skin, thinking of racial identification primarily in terms of chromatic scales or dermatoscopic data deflects attention, to put it generously, from the long history of the damaging associations of skin color and race that gave rise to early technologies like this in the first place, whether it was the “science” of phrenology, IQ tests, or fingerprinting, and with implications, more recently, for the use of biometrics.27 At a minimum, it ignores the imbrication of “race” in pigmentocracies and colorism, the historical privileging of light skin, and the various rationales for identifying what counts as “light-skinned.” Colorism, a legacy of colonialism, continues to persist in contemporary hierarchies of value and social status, including aesthetics (who or what is ranked beautiful, according to white, Western standards), moral worth (the religious iconography of “dark” with evil and “light” with holy continues to saturate languages), social relations (for instance, the “paper bag test” of the twentieth century was used as a form of class gatekeeping in some African American social institutions),28 and the justice system (since social scientists have documented the perceptual equation of “blackness” with crime, and thus those perceived as having darker skin as a priori criminally suspect).29
Why does this matter? Because it suggests that the challenges in representing race in AI are not something technological advances in any near or far future could solve. Rather, they signal cultural and political, not technical, problems to address. The issue, after all, is not merely a question of bias (implicit or otherwise), nor of inaccuracy (which might lead some to think the answer is simply the generation of more granular categories), nor of racial misrecognition (which some might hear as simply a call for ever more sophisticated FRT), nor even of ending all uses of racial categorization.30 It matters because algorithms trained on data sets of racial types reinforce color lines, literally and figuratively remanding people back in their “place.” By contrast, as I have suggested, the increasingly influential rise of AI artist-technologists, especially those of color, are among those most dynamically questioning and reimagining the commercial imperatives of “personalization” and “frictionlessness.” Productively refusing colorblindeness, they represent race, ethnicity, and gender not as normative, self-evident categories nor monetizable data points, but as the dynamic social processes–always indexing political tensions and interests–which they are. In doing so, they make possible the chance to truly create technologies for social good and well-being.
Something has happened. Something very big indeed, yet something that we have still not integrated fully and comfortably into the broader fabric of our lives, including the dimensions–humanistic, aesthetic, ethical and theological–that science cannot resolve, but that science has also (and without contradiction) intimately contacted in every corner of its discourse and being.
—Stephen Jay Gould, The Hedgehog, the Fox, and the Magister’s Pox (2003)31
I cite what may seem minor examples of cultural ephemera because, counterintuitively, they hint at the grander challenges of AI. They are a thread revealing the pattern of “something very big indeed,” as historian of science Stephen Jay Gould put it. Certainly there are ethical, economic, medical, educational, and legal challenges facing the future of AI. But the grandest technological challenge may in fact be cultural: the way AI is shaping the human experience. Through that lens, the question becomes not one of automation versus augmentation, in which “augmenting” refers to economic productivity, but rather to creativity. That is, how can AI best augment the arts and humanities and thus be in service to the fullness of human expression and experience?
This essay opened with Henry Adams’s moment of contact with the Dynamo’s “silent and infinite force,” as he put it, which productively denaturalizes the world as he knows it, suspends the usual epistemological scripts about the known world and one’s place in it. It is a sentiment echoed almost verbatim two hundred years later by Gould, witnessing another profound technological and cultural upending. Writing at the turn into our own century, Gould, like Adams, cannot fully articulate the revelation except to say poignantly that “something has happened,” that every dimension of “the broader fabric of our lives” is intimately touched by a technology whose profound effect cannot be “solved” by it. That liminal moment for Adams, for Gould, and for us makes space for imagining other possibilities for human creativity, aesthetic possibilities that rub against the grain and momentum of current technological visions, in order to better realize the “magisteria of our full being.”32
© 2022 by Michele Elam. Published under a CC BY-NC 4.0 license.
- 1Henry Adams, “The Virgin and the Dynamo,” in The Education of Henry Adams (Boston: self published, 1907). All of Henry Adams’s quotes are taken from the unpaginated PDF hosted by the University of Chicago.
- 2Unlike in The Wonderful Wizard of Oz, L. Frank Baum’s children’s book published the same year as the Paris Exhibition, for Adams, there is no reveal of the “man behind the curtain,” no Oz orchestrating a show. His interest is not in the technologists but in what ontological truths their creations tap.
- 3While there is no clear ethical or legal consensus on what constitutes “fairness,” there are critiques of fairness models that assume an equal playing field thwarting access to opportunities, that presume equal discrimination equals fairness, or that understand fairness in the narrowest sense of preventing harm (the critique of situating “fairness” under concerns of “safety”). For a review of some of the debates about fairness, see Michele Elam and Rob Reich, “Stanford HAI Artificial Intelligence Bill of Rights” (Stanford, Calif.: Stanford Institute for Human-Centered Artificial Intelligence, 2022).
- 4Adams, “The Virgin and the Dynamo.”
- 5The “Prayer to the Dynamo” is a section within a longer poem by Henry Adams, Prayer to the Virgin of Chartres (1900), accessed through Teaching American History.
- 6On the question of humanness, and for an analysis of how technologies determine full humans, not-quite-humans, and nonhumans, see Alexander G. Weheliye’s excellent Habeas Viscus: Racializing Assemblages, Biopolitics, and Black Feminist Theories of the Human (Durham, N.C.: Duke University Press, 2014). See also Sylvia Wynter’s foundational essay, “Unsettling the Coloniality of Being/Power/Truth/Freedom: Towards the Human, After Man, Its Overrepresentation–An Argument,” CR: The New Centennial Review 3 (3) (2003): 257–337. Wynter critiques the overrepresentation of man (as white, Western) as the only imaginable mode of humanness, overwriting other ontologies, epistemologies, and imaginaries. See also Katherine McKittrick, ed., Sylvia Wynter: On Being Human as Praxis (Durham, N.C.: Duke University Press, 2015). A major influence on this essay, Wynter’s pioneering and prolific work draws on arts, humanities, natural sciences and neuroscience, philosophy, literary theory, and critical race theory.
- 7Homi Bhabha, “Signs Taken for Wonders: Questions of Ambivalence and Authority Under a Tree outside Delhi, May 1817,” Critical Inquiry 12 (1) (1985): 144.
- 8Vanessa Chang, “Prosthetic Memories, Writing Machines,” Noëma, December 3, 2020.
- 9For instance, see the comment made by the then head of Instagram, Adam Mosseri, that his recent policy eliminating public “likes”–because of his concern “about the unintended consequences of Instagram as an approval arbiter” as he put it–was partly informed by an episode of the science fiction anthology television series Black Mirror. Amy Chozick, “This Is the Guy Who’s Taking Away the Likes,” The New York Times, January 17, 2020.
- 10For example, see Google’s “Poem Portraits by Es Devlin.” A 2016 short-film script entitled Sunspring was made by an AI bot. The main character in the filmed version was played by Thomas Middleditch, the same actor who plays the lead, Richard Hendriks, in the TV series Silicon Valley 😊. See also AI-generated musicals and symphonies, as in Maura Barrett and Jacob Ward, “AI Can Now Compose Pop Music and Even Symphonies. Here’s How Composers Are Joining In,” MACH by NBC News, May 29, 2019.
- 11Artist-technologists working at these intersections include Amelia Winger-Bearskin, Legacy Russell, Stephanie Dinkins, Ian Chang, Rashaad Newsome, Jacolby Satterwhite, Joy Buolamwini, Martine Syms, and others–not to mention those writers in the long literary history of speculative fiction.
- 12See, for instance, “Art at a Time Like This.” I take up in detail the debates over the value and valuation of the arts and humanities in the age of artificial intelligence in Michele Elam, “GPT-3 in ‘Still I Rise!’: Why AI Needs Humanists,” American Literature (forthcoming spring 2022).
- 13Luke Stark and Kate Crawford, “The Work of Art in the Age of Artificial Intelligence: What Artists Can Teach Us about the Ethics of Data Practice,” Surveillance & Society 17 (3/4) (2019): 452.
- 14Ibid., 451–452.
- 15There are many academic and activist resources and collectives working in these areas, including Latinx in AI, Black in AI, Queer AI, Indigenous AI, and Accessible AI, to name but a few.
- 16See, for instance, Ruha Benjamin, “Introduction: Discriminatory Design, Liberating Imagination,” in Captivating Technology: Race, Carceral Technoscience, and Liberatory Imagination in Everyday Life, ed. Ruha Benjamin (Durham, N.C.: Duke University Press, 2019), 1–25; and Nettrice R. Gaskins, “Techno-Vernacular Creativity and Innovation across the African Diaspora and Global South,” in Captivating Technology, 252–253.
- 17Amelia Winger-Bearskin explains the concept in “Antecedent Technology: Don’t Colonize Our Future,” Immerse News, November 21, 2019; and Amelia Winger-Bearskin, “Before Everyone Was Talking about Decentralization, Decentralization Was Talking to Everyone,” Immerse News, July 2, 2018.
- 18Alison Adams quoted in Yarden Katz, Artificial Whiteness: Politics and Ideology in Artificial Intelligence (New York: Columbia University Press, 2020), specifically in Katz’s discussion of AI notions of the self: “Practitioners in the 1970s, for instance, offered visions of the self as a symbolic processing machine. In the late 1980s and early 1990s, by contrast, the prevailing ‘self’ started looking more like a statistical inference engine driven by sensory data. But these classifications mask more fundamental epistemic commitments. Alison Adams has argued that AI practitioners across the board have aspired to a ‘view from nowhere’–to build systems that learn, reason, and act in a manner freed from social context. The view from nowhere turned out to be a view from a rather specific, white, and privileged space.” Ibid., 6. See Alison Adams, Artificial Knowing: Gender and the Thinking Machine (Abingdon-on-Thames, United Kingdom: Routledge, 1998).
- 19See Rosemarie Garland-Thomson, Extraordinary Bodies: Representing Physical Disability in American Culture and Literature (New York: Columbia University Press, 1996). The curatorial statement in Recoding CripTech explains the terminology: “the term ‘crip’ reclaims the word for disability culture and recognizes disability as a political and cultural identity”; see Recoding CripTech. See also M Eifler’s Prosthetic Memory (2020); and Masking Machine (2018). Many thanks to Lindsey Felt and Vanessa Chang, who produced and curated Recoding CripTech, and especially for Lindsey’s suggestions for this section of the essay.
- 20Su Lin Blodgett and Brendan O’Connor, “A Racial Disparity in Natural Language Processing: A Case Study in Social Media African American English,” arXiv (2017).
- 21See Neda Atanasoski and Kalindi Vora, “The Surrogate Human Affect: The Racial Programming of Robot Emotion,” in Surrogate Humanity: Race, Robots and the Politics of Technological Futures (Durham, N.C.: Duke University Press, 2019), 108–133.
- 22Michael Omi and Howard Winant, Racial Formation in the United States, 2nd ed. (New York: Routledge, 1994), 55–56.
- 23Katz, Artificial Whiteness: Politics and Ideology in Artificial Intelligence. The claim to universality, according to Katz, is the first forgery: “The second is that AI systems have matched or exceeded the capabilities of human thought [drawing] on deep-seated notions in Western culture about hierarchies of intelligence. The third epistemic forgery suggests that these computational systems arrive at truth, or ‘knowledge,’ ‘on their own,’ AI practitioners being merely the ones who set off the necessary conditions for computational processes to properly unfold.” Ibid., 94–95.
- 24Directly related to the issue of racial classification is the issue of racial and gender “ambiguity,” often (mis)understood in AI as simply a technical issue of documenting (or not) variance and managing uncertainty. For extended discussion of the complicated technical and political challenges posed by social identities transgressing racial or gender boundaries in AI, see Michele Elam, “Recoding Gender-Binaries and Color-Lines: AI and the Erasure of Ambiguity” in Feminist AI (forthcoming 2022).
- 25By arguing for understanding words “in context,” I mean in the social science sense of meaning emerging through performative interaction, not meaning garnered by sentencelevel “context,” as is commonly understood in natural language processing: that is, of other words that appear near each other in a sentence. Thanks to my Stanford colleagues Chris Manning and Surya Ganguli for our conversations about natural language processing.
- 26Fran Ross, Oreo (Boston: Northeastern University Press, 2000), 5. First published in 1974 by Greyfalcon House, Inc.
- 27This reduction of race to data points for AI has many problematic medical implications, even in work aiming to mitigate bias. Using self-reported racial data along with other medical information to create a “phenotype,” Juan Banda acknowledges that racial porosity (people identifying with multiple races, differently over time, or differently in various contexts) as well as durational changes in aging mean up to 20 percent of data are thrown out in assessments. Moreover, as is long documented, race is often proxy for all sorts of other factors and inequities, so whether the data collected are self-reported or social ascription, it is still a highly problematic category. Juan Banda, “Phenotyping Algorithms Fair for Underrepresented Minorities within Older Adults,” Stanford HAI Weekly Seminar with Juan Banda, February 2, 2022.
- 28The “paper bag test,” and color discrimination broadly, is a complicated form of internalized racism dating to slavery. Early scholarship suggested lighter-skinned slaves were supposedly treated better, although that view has been challenged, since being “in the Big House” instead of the field often meant greater exposure to sexual abuse. Moreover, there are many documented accounts of mixed race children–the issue of white masters and slaves–being sold away, often at the insistence of the white mistress of the house, since they stood as corporeal testimony to miscegenation and adultery–in short, to the sins of the father.
- 29See Jennifer Eberhardt, Biased: Uncovering the Hidden Prejudice that Shapes What We See, Think, and Do (London: Penguin, 2020). Eberhardt documents how police are more likely to think a suspect is holding a gun if they are perceived as “dark,” and that juries are more likely to negatively regard, and thus weigh judgment against, a defendant if they are considered “dark,” among the many other consequences.
- 30For instance, racial classification is often employed in the United States (though not in France and other countries) for strategic purposes generally understood as for the larger social good: for instance, by the Office of Management and Budget for census purposes, which are tied to the distribution of resources and the tracking of civil rights violations, or by Federal Statistical Research Data Centers that enable the study of health and economic trends across demographics. In such cases, data collection of those gross categories is undertaken with the understanding that these are efforts to capture not inherent traits or types, but rather broad trends in the complex of race, gender, and socioeconomics, among other variables. More recently, efforts to collect immigration and citizen status, neither following the prior intent nor practice of the Census, has made vivid the potential for misuses of this national mode of data collection.
- 31Stephen Jay Gould, The Hedgehog, the Fox, and the Magister’s Pox: Minding the Gap Between Science and the Humanities (Cambridge, Mass.: Belknap Press, 2011), 15.
- 32I am following Stephen Jay Gould’s critiques in The Hedgehog, the Fox, and the Magister’s Pox of E. O. Wilson and C. P. Snow’s notion of two cultures, described in Snow’s The Two Cultures and the Scientific Revolution as a split between science and humanities. Gould found that duality artificial, simplistic, and ahistorical, offering instead more accurate accounts of four historical periods exploring continuities of creative thinking across the sciences and humanities.