The history of technological automation – its ideas and implementations – is brimming with examples of how human efforts have been made invisible or unimportant in order for technology to come across as self-sufficient. As a field, artificial intelligence (AI) has contributed widely to the downplaying of labor involved at development as well as production stages of systems. Given the grand visions of AI – cognitive performance on par with or beyond that of its makers and the professed technological unemployment resulting from such accomplishments – the strategic invisibility of human work is here particularly salient. This paper investigates these matters by studying the policies of prominent actors in contemporary AI labs and situating them in a longer history of making certain forms of work disappear. Specifically, the paper addresses two kinds of labor of critical importance to the current paradigm of machine learning. First, the manual labor required to classify and tag data in a process known as "labeling" and, second, the work of assessing and tuning model output during so-called reinforcement learning with human feedback (RLHF). Moreover, before models have any data to train their patterning and predictions on, Internet users have invested in the production of digital data for a variety of purposes which were subsequently exploited in pre-training processing. In addition to being principally veiled, this paper understands these types of work – the toil of labeling and the correcting of model output by RLHF – as disciplinary in nature. They limit the range of available categories to which human-made information can belong and they punish undesired behavior. The ultimate goal of these efforts, this paper argues, is to remove the kinds of mistakes that users of AI-models might find incoherent or offensive while masking that such labor was carried out, all in the pursuit of creating the impression of an artificial intelligence reliably performing in an automatic and autonomous manner.
QC 20240528