Meta and Twitter have discovered and dismantled a network of coordinated inauthentic behavior spreading pro-US (and anti-China/Russia/Iran) narratives in Central Asia and the Middle East (Al Jazeera, Axios stories). Undoubtedly, this kind of intervention bolsters the platforms’ image as neutral purveyors of information and entertainment, determined to enforce the rules of the game no matter what the ideological flavor of the transgression may be. In a way, paradoxically, such impartiality may even play well in Washington, where the companies would certainly welcome the support, given the current unfavorable political climate.
The type of universalism on display in this instance harkens back to an earlier era of the internet, the techno-libertarian heyday of the 1990s. Arguably, however, that early globalist vision of the world-wide web has already been eviscerated at the infrastructural level, with the growth of distinctive national versions of online life, in a long-term process that has only been made more visible by the conflict in Ukraine. Hence, the impartiality and universality of Meta and Twitter can be seen ultimately as an internal claim by and for the West, since users in countries like Russia, China, or Iran are unable to access these platforms in the first place. Of course, geopolitical splintering was one of the ills the web3 movement set out to counter. How much decentralization can resist the prevailing ideological headwinds, however, is increasingly unclear. Imperfect universalisms will have to suffice for the foreseeable future.
A review by Paul Dicken published online a week ago in The New Atlantis is representative of a certain kind of argument in contemporary social critiques of high tech. The piece discusses a book by Ben Schneiderman entitled Human-Centered AI, which came out earlier this year for Oxford UP, and mainly reads as an exposé of a benighted scientism that at best is hopelessly naïve about its potential to effect meaningful emancipatory social change and at worst is disingenuous about the extractive and exploitative agendas that underwrite its deployment.
One would not wish to deny that Schneiderman makes for a good target: computer scientists as a sociological class are hardly more self-reflexive or engagé than any other similarly-defined professional group, and divulgative AI-and-management texts seldom present incisive and counterintuitive social commentary. Nonetheless, it is hard to miss a certain symmetry between the attacks on the political self-awareness of the author in question (how could he have missed the damning social implications??) and the peans to progress through techno-solutionism which characterized public debate on Web2.0 before the techlash.
The fact itself that Dicken refers back to Charles Babbage as a precursor of contemporary AI research and its dark side should suggest that the entwinement of technological advancement with political economy might be a long-run phenomenon. What is different is that in the present conjuncture would-be social critics seem to harbor absolutely no faith that the political and social ills upstream from technological development can be righted, and no plan to do so. New technology changes affordances, and this shift makes certain social dynamics more visible. But in the absence of specifically political work, such visibility is ephemeral, irrelevant. Hence, the exposé of political cluelessness risks becoming the master trope of the techlash, essentially a declaration of social impotence.
An interesting, thoughtful article by Michelle Santiago Cortés in The Cut last week looks at affective relationships with algorithms and their role in shaping our identities.
Three parts of the analysis specifically stood out to me. The first revolves around our typical lack of knowledge of algorithms: Cortés’ story about
some YouTube alpha male […] out there uploading videos promising straight men advice on how to “hack” the Tinder algorithm to date like kingsis clearly only the tip of a gigantic societal iceberg, a cargo-culture-as-way-of-life involving pretty much everyone in the remotest, most diverse walks of life. The ever-evolving nature of these algorithms compounds the obfuscation effect, making end-users’ strategic attempts, whether exploitation- or resistance-focused, generally appear puerile.
Second, the clarity with which Cortés encapsulated the main tradeoff in the relationship was truly apt:
[w]e are, to varying degrees, okay with being surveilled as long as we get to feel seen.
The assertion of visibility and assurance of recognition are two of the key assets algorithmic systems offer their users, and their value can hardly be minimized as mere late-consumerist narcissism.
Finally, the comparison between algorithmic portraits of personality and astrology was extremely telling: closing the behavioral loop from algorithmic interaction to the redefinition of one’s own identity on the basis of the algorithm’s inevitably distorting mirror is still a matter of choice, or rather, a sensibility that can be honed and socialized regarding the most empowering and nurturing use of what is ultimately a hermeneutic tool. Of course, such a benign conclusion rests on the ambit of application of such technologies: music videos, entertainment, dating. As soon as our contemporary astrological devices are put in charge of directing outcomes in the field of political economy and public policy, the moral calculus rapidly shifts.
A long blog post on Olivier Ertzscheid’s personal website [in French] tackles the ideological orientation of the major social media platforms from a variety of points of view (the political leanings of software developers, of bosses, of companies, the politics of content moderation, political correctness, the revolving door with government and political parties, the intrinsic suitability of different ideologies to algorithmic amplification, and so forth).
The conclusions are quite provocative: although algorithms and social media platforms are both demonstrably biased and possessed of independent causal agency, amplifying, steering, and coarsening our public debate, in the end it is simply those with greater resources, material, social, cultural, etc., whose voices are amplified. Algorithms skew to the right because so does our society.
An extremely interesting series of talks hosted by the Digital Freedom Fund: the automation of welfare system decisons is where the neoliberal agenda and digitalization intersect in the most socially explosive fashion. All six events look good, but I am particularly looking forward to the discussion of the Dutch System Risk Indication (SyRI) scandal on Oct. 27th. More info and free registration on the DFF’s website.