‘What the Robot Saw’ is an endless documentary by a social media robot. It depicts and describes the people and scenes it encounters — in the ways its algorithms have been trained to perceive them. It’s also a massively durational online robo-performance and archive.
A social media AI turned documentary filmmaker, the Robot continuously makes its way through the world of low engagement online video, carefully organizing and describing the people and scenes it features in its documentary. The film is constantly curated, edited, titled and archived algorithmically from among the least viewed and subscribed YouTube videos uploaded over the past several hours. While on the one hand a robot documentary and generative durational performance, “What the Robot Saw” also documents the complicated relationship between the world’s surveillant and curatorial AI robots and the the humans who are both their subjects and their stars.
‘What the Robot Saw’ streams at 1080p, so please check that the YouTube player is set for the highest resolution your network connection will handle (gear icon on lower right of the player.) Fullscreen or Theatre Mode is recommended (net conditions permitting.) If the stream isn’t live, you can find recent archives here.
An invisible audience of software robots continually analyze content on the Internet. Videos by non-“YouTube stars” that algorithms don’t promote to the top of the search rankings or the “recommended” sidebar may be seen by few or no human viewers. For these videos, robots may be the primary audience. In ‘What the Robot Saw,’ the Robot is AI voyeur turned director: classifying and magnifying the online personas of the subjects of a never-ending film.
Using computer vision, neural networks, and other robotic ways of seeing, hearing, and understanding, the Robot continually selects, edits, and identifies recently uploaded public YouTube clips from among those with low subscriber and view counts, focusing on personal videos. A loose, stream-of-consciousness narrative develops as the Robot drifts through neural network-determined groupings. As the Robot scans and magnifies the clips, it generates the film in a style fitting its own obsessions, inserting titles for sections and “interviewees,” and streaming it live back to YouTube for public viewing.
Robot meets Resting Bitch Face and Other Adventures. As it makes its way through the film, the Robot adds lower third supers: periodic section titles, derived from its image recognition-based groupings and interpreted through the Robot’s vaguely poetic perspective; and frequent identifiers, for the many human interviewees in its documentary. The identifiers — talking head style descriptions like “Confused-Looking Female, age 22-34” — are generated using Amazon Rekognition — a popular commercial face detection/recognition library. The feature set of Rekognition offers a glimpse into how computer vision robots, marketers, and others, choose to categorize humans. While attempting to adhere to Rekognition’s guidelines that differentiate a person’s appearance from their actual internal emotional state, the Robot titles each person as it analyzes/perceives them — and as marketers and others using similar software do. When you’re a computer vision robot, appearance is everything. Pixels don’t have internal states.
Dated archives are generated on YouTube for each daypart livestream, offering a theoretically endless, on-demand archive of the videos few humans get to see, as robots might, and sometimes do, see them.
* The live stream runs throughout day; there are “intermissions” every four hours (and as needed for maintenance.) Archives of recent streams are available on the Videos page or on the YouTube Channel.
Although the YouTube live stream is central to the project, the technical limitations of live streaming mean the image and sound quality are not ideal and may vary with network conditions. A high quality stream can be generated locally for art installations and screenings.