Activation Layer Logo
← Back to blog

Published on Sun Feb 19 2023 18:37:00 GMT+0000 (Coordinated Universal Time) by Ryan Williams

Information Disorder and Generative AI - An Evolving Reading List

Last updated: Jul 11, 2023

Introduction

We intend for this list to function as a living document that evolves alongside the rapidly changing landscape of generative AI, disinformation, and misinformation.

The list is structured into three main categories, each offering a unique perspective.

First, we provide accessible explanations of core technologies and concepts like neural networks and transformer models. We also discuss the capabilities and limitations of large language models.

The second category focuses on the intersection of Generative AI and Information Disorder. Here, we explore the potential misuse of language models for disinformation campaigns, the psychological impact of an AI-saturated ‘post-truth world’, and challenges for privacy, democracy, and national security. We also link to frameworks for understanding the participatory nature of strategic information operations and the market dynamics incentivizing the creation of disinformation.

Finally, we turn our attention to AI Governance and Policy, offering both US and global perspectives on the matter.

Feel free to navigate through the sections at your own pace as per your interest. Happy reading!

Generative AI – Technology

  • But what is a neural network? – (19m) This video explains the basics of neural networks. No technical skills required.

  • What is a Transformer Model – NVIDIA Blog. The Transformer model architecture is one of the enabling innovations behind the recent explosion of AI applications.

  • The Text-to-Image Revolution, Explained – (13m) Vox

  • GPT-4 System Card (60pgs) – OpenAI authored paper that characterizes the capabilities of GPT-4.

  • Talking about Large language Models (11pgs) This influential paper by Murray Shanahan aims to inject philosophical nuance in the discourse around artificial intelligence by clarifying how large language models work and steering away from anthropomorphic analogies.

  • On the Opportunities and Risks of Foundation Models (161 pgs) – Large Language Models are sometimes called “foundation models” because they enable a diverse set of capabilities through fine-tuning and prompt engineering. This comprehensive report was written by the Center for Research on Foundation Models (CRFM) and Stanford Institute for Human-Centered Artificial Intelligence. It can help you develop an intuition for what these models are “good for”.

  • Chain-of-thought Reasoning – (31pgs) One feature of all state-of-the-art models is that the quality of generation is dependent on the user’s prompting strategy. In other words, the instructions you provide the model can dramatically affect what the model produces. This paper demonstrates the chain-of-thought prompting technique.

  • Retrieval Augmented Generation – (25m) Why are tech executives so confident that LLMs will transform search experiences? This interview explains the concepts behind Retrieval Augmented Generation.

  • The Impact of AI on Developer Productivity: Evidence from GitHub Copilot – (19pgs) While not directly relevant to information disorder, generative AI is increasingly capable of producing useful code from natural language instructions. It is easy to imagine how these tools will make inauthentic automated amplification even more accessible to bad actors.

Generative AI and Information Disorder

General

Frameworks for thinking about Information Disorder

  • Disinformation as Collaborative Work: Surfacing the Participatory Nature of Strategic Information Operations (20pgs) – This article by Kate Starbird, Ahmer Arif, and Tom Wilson foregrounds the concept of computer-mediated collaborative work in disinformation campaigns. Information operations don’t “just happen”. Thinking about the sociotechnical systems that enable disinformation campaigns can help us anticipate how generative AI will serve sophisticated threat actors beyond simply “making more content”.

  • The Marketplace for Rationalizations (22pgs) – This innovative paper by Daniel Williams explores how market dynamics incentivize the creation of disinformation and misinformation. As generative AI changes the marginal cost of producing content, these market dynamics may usher in sophisticated infrastructure for delivering misinformation.

AI Governance and Policy

US Perspectives

Global Perspectives

Theories of Governance and Disinformation Policy

These ideas and frameworks will sensitize you to the questions facing governments and corporations around the world.

Written by Ryan Williams

← Back to blog