Project Demonstrates Potential of New Transparency Standard for Synthetic Media

Justin Hendrix is CEO and Editor of Tech Policy Press. The views expressed here are his own. With the proliferation of tools to generate synthetic media, including images and video, there is a great deal of interest in how to mark content artifacts to prove their provenance and disclose other information about how they were generated and edited. This week, Truepic, a firm that aims to provide authenticity infrastructure for the Internet, and Revel.ai, a creative studio that bills itself as a leader in the ethical production of synthetic content, released a “deepfake” video “signed” with such a marking to disclose its origin and source. The experiment could signal how standards adopted by content creators, publishers and platforms might permit the more responsible use of synthetic media by providing viewers with signals that demonstrate transparency. The video features a message delivered by a synthetic representation of Nina Schick, the creator of ‘The Era of Generative AI’ online community and author of the book ‘DEEPFAKES.’ The project follows years of effort by a wide variety of actors, including tech and media firms as well as nonprofit organizations and NGOs, to create the conditions for such signals to meet an interoperable standard. The video is compliant with the open content provenance standard developed by the Coalition for Content Provenance and Authenticity (C2PA), an alliance between Adobe, Intel, Microsoft, Truepic, and a British semiconductor and software design company called Arm. A joint development foundation intended to produce such a standard, the C2PA itself emerged…Project Demonstrates Potential of New Transparency Standard for Synthetic Media

Can Piaget Explain Jair Bolsonaro?

Paulo Blikstein is an Associate Professor at Teachers College, Columbia University, an Affiliate Associate Professor in the Department of Computer Science at Columbia University, and Director of the Transformative Learning Technologies Lab and of the Lemann Center for Brazilian Studies. Renato Russo is a doctoral student at Teachers College and a researcher at the Transformative Learning Technologies Lab. Swiss cognitive scientist Jean Piaget demonstrated that there is nothing more resilient than a theory we create on our own. Narratives and stories are powerful, but they lack one crucial property, in comparison: they don’t make us feel as clever and intellectually capable. We propose that this pleasure and feeling of self-efficacy in theorizing – also proven by decades of neuroscience research – is closely related to current political communication and democracy-threatening events that took place in Brazil last January. Starting with the election of Lula in late October, thousands of Bolsonaro supporters spent as much as two months camping in front of military facilities, mobilized around the claim of rigged elections, culminating in the siege of the Brazilian capital on January 8th, 2023.  “Fake news” explains part of a campaign that elected Brazil’s Jair Bolsonaro four years ago and that kept part of his constituency mobilized. But it is only part of the story. Research by media studies and communication scholars, such as Francesca Tripodi, Alice Marwick, and Ethan Zuckerman, has shown how extremists resort to epistemological practices that in a way resemble those of scientific communities. Drawing on this scholarship,…Can Piaget Explain Jair Bolsonaro?

What Generative AI Reveals About the Limits of Technological Innovation

Dr. Joe Bak-Coleman is an associate research scientist at the Craig Newmark Center for Journalism Ethics and Security at Columbia University and an RSM assembly fellow at the Berkman Klein Center’s Institute for Rebooting Social Media. March 1940 meeting of scientists developing the atomic bomb in the Radiation Laboratory at Berkeley, California: Ernest O. Lawrence, Arthur H. Compton, Vannevar Bush, James B. Conant, Karl T. Compton, and Alfred L. Loomis. Wikimedia Over the past month, generative AI has ignited a flurry of discussion about the implications of software that can generate everything from photorealistic images to academic papers and functioning code. During that time period, mass adoption has begun in earnest, with generative AI integrated into everything from Photoshop and search engines to software development tools. Microsoft’s Bing has integrated a large language model (LLM) into its search feature, complete with hallucinations of basic fact, oddly manipulative expressions of love, and the occasional “Heil Hitler.” Google’s Bard has fared similarly– getting textbook facts about planetary discovery wrong in its demo. A viral image of the pope in “immaculate drip” created by Midjourney even befuddled experts and celebrities alike who, embracing their inner Fox Mulder, just wanted to believe. Even in the wake of Silicon Valley Bank’s collapse and slowdown in the tech industry, the funding, adoption, and embrace of these technologies appears to have occurred before their human counterparts could generate– much less agree on– a complete list of things to be concerned about. Academics have raised the alarm about…What Generative AI Reveals About the Limits of Technological Innovation

Evaluating New Technology for Equitable and Secure Voter Verification

Dr. Juan E. Gilbert is the Andrew Banks Family Preeminence Endowed Professor and Chair of the Computer & Information Science & Engineering Department at the University of Florida. He leads the Computing for Social Good Lab, where Jasmine McKenzie, Alaina Smith and London Thompson are PhD students. Shutterstock Elections are the bedrock of democracy. As such, access to voting is essential; however, there have been severe challenges over the decades to voting access for people of color, those with disabilities and other marginalized groups in the United States. One of those challenges revolves around the verification of voter eligibility. New technologies may present solutions to this problem, but substantial research is necessary to verify the efficacy and address the downsides of any new tools and techniques that determine who has access to the franchise. Essentially, voter verification determines who has access to vote. Voter verification methods vary across the U.S. by state. Each state requires some form of identification to register and vote. These requirements have often served as tools to disenfranchise communities of color.  For example, in Texas, a pistol license granted by the Department of Public Safety is an acceptable form of voter identification (ID); however, a student ID from a Texas public university is not. A driver’s license is the primary form of voter identification in most states; however, voters of color and the elderly may use public transportation and may not have a state-issued driver’s license. These disparities in state criteria have the effect of disenfranchising…Evaluating New Technology for Equitable and Secure Voter Verification

NPR is Not RT: Twitter’s New State-Affiliated Media Policy is Misleading

Joseph Bodnar is a research analyst at the Alliance for Securing Democracy at the German Marshall Fund, where he tracks Russian propaganda and disinformation. On April 4, Twitter placed a state-affiliated media label on NPR’s account. The label is meant to provide users with context when they see a media account that is under a state’s editorial control, like Russia’s RT and China’s People’s Daily that lack organizational and financial firewalls to insulate their coverage from government interference. NPR doesn’t fit that description. The outlet gets less than 1% of its funding from the federal government. The other 99% comes largely from corporate sponsorships, membership drives, and fees from affiliate radio stations. This arrangement ensures that NPR remains free from state control. Twitter’s move to add a state media label to NPR’s account therefore equates editorially independent media with propaganda outlets used by autocratic regimes to do things like cover up war crimes and cultural genocide.   At the time Twitter labeled NPR, the platform’s own policy explicitly named the public broadcaster as an example of media that receives state funding but maintains its editorial freedom. NPR did nothing to force Twitter’s policy change. The way the platform makes content moderation decisions changed. A team that understood state-backed media and information campaigns used to oversee those policies. Now, rules are being dictated by a person whose ideas often seem to reflect advice given by trolls.   Regardless, Twitter’s labeling of NPR does not appear to be part of any broader policy change—at…NPR is Not RT: Twitter’s New State-Affiliated Media Policy is Misleading

Autocrats Will Benefit Most from Twitter’s New Approach to State-Affiliated Media

E. Rosalie Li is a researcher and recent interdisciplinary graduate from the Johns Hopkins Bloomberg School of Public Health. On April 4, Elon Musk’s Twitter added a new label to the account for National Public Radio (NPR). The decades-old independent broadcaster is now classified as “state-affiliated.”  The move came even though Twitter’s policy said that funding from a government source alone does not warrant the label of “state-affiliated media,” provided the outlet has editorial independence, as in the case of the BBC in the UK. Contrast that with outlets such as Russia’s RT and Sputnik, where even employees have highlighted the lack of editorial freedom in these Kremlin-backed organizations, bolstering the case for their classification as “state-affiliated.”  Twitter’s (or perhaps more accurately, Musk’s) seemingly arbitrary decision risks undermining the distinction between ethical, state-financed journalism and propaganda arms controlled by autocratic regimes. A gift to dictatorial leaders who hide behind claims of persecution, these corrosive changes occur in an increasingly hostile online environment, where censorship and suppression of free speech are major concerns. And, it raises serious questions about whether Musk is a responsible steward of such an important platform. NPR and RT: Apples and Oranges  As of April 5, 2023, Twitter’s policy on “Government and state-affiliated” outlets still seemed to exclude an outlet such as NPR, which receives only a small portion of its budget from government sources: “State-financed media organizations with editorial independence, like the BBC in the U.K. for example, are not defined as state-affiliated media for the…Autocrats Will Benefit Most from Twitter’s New Approach to State-Affiliated Media

AI Will Break Online Search

AI is going to give super powers to Blackhat SEO, breaking online search as we know it. RIP magic box at the top of the browser where we all reflexively type “how to [blank]”, “[blank] near me”, “best [blank] Reddit”, and anything else we seek a semi-reliable answer to at any given moment. Online search results are about to get flooded with Astroturf and spam content at a level never before possible. People have discussed Google Search results going down hill for years. A lot of factors influence why Google search lost some of its magic, I’ve written thousands of words on the topic. Experts have universal consensus of two factors for search quality decline, the volume of things on the internet (quantity) and the lower average caliber of all those things (quality). Not the only factors, but everyone from former Google execs to an ad agency demon like myself agree those have major impact on Google’s search product. If this was only a deluge of content from real human people posting on Twitter, TiKToK and poorly edited blogs the companies in the business of sorting an parsing could handle it. The problem is attempted manipulation of ranking within those companies indexes of the web. AI is about to scale those blackhat systems. AI Will Clog The Internet’s Toilets “Typeface, a startup developing an AI-powered dashboard for drafting marketing copy and images, emerged from stealth this week with $65 million in venture equity backing”, so begins the TechCrunch article about…AI Will Break Online Search

AI Will Break Online Search

AI is going to give super powers to Blackhat SEO, breaking online search as we know it. RIP magic box at the top of the browser where we all reflexively type “how to [blank]”, “[blank] near me”, “best [blank] Reddit”, and anything else we seek a semi-reliable answer to at any given moment. Online search results are about to get flooded with Astroturf and spam content at a level never before possible. People have discussed Google Search results going down hill for years. A lot of factors influence why Google search lost some of its magic, I’ve written thousands of words on the topic. Experts have universal consensus of two factors for search quality decline, the volume of things on the internet (quantity) and the lower average caliber of all those things (quality). Not the only factors, but everyone from former Google execs to an ad agency demon like myself agree those have major impact on Google’s search product. If this was only a deluge of content from real human people posting on Twitter, TiKToK and poorly edited blogs the companies in the business of sorting an parsing could handle it. The problem is attempted manipulation of ranking within those companies indexes of the web. AI is about to scale those blackhat systems. AI Will Clog The Internet’s Toilets “Typeface, a startup developing an AI-powered dashboard for drafting marketing copy and images, emerged from stealth this week with $65 million in venture equity backing”, so begins the TechCrunch article about…AI Will Break Online Search

How Forbes Monetizes The Frauds They Create

Sam Bankman-Fried, Elizabeth Holmes, and Charlie Javice share a few things in common. They all seem to have committed fraud; though a court of law convicted only Homes at the time of writing. The three are also Forbes ’30 Under 30′ Alumni. Forbes is the best in the business at monetizing frauds on the way up and later on the way down. Javice was the CEO of a startup called Frank that JPMorgan Chase acquired in large part because of the over 4 million users the company boasted. The best coverage of JPMorgan Chase’s lawsuit claiming Frank only had 300,000 real users and created 4.265 million fake customer records to satisfy due diligence is in Forbes. Holmes was found guilty on four of 11 fraud charges as the founder of the fake blood-testing startup Theranos. She received an 11-year prison sentence. Forbes has published many articles covering the trial, the sentencing, and the analysis of what went wrong. Just as they once happily propped Holmes up as some business and technical thought leader. Bankman-Fried, once propped up by fawning media coverage in Forbes, now graces the pages with words about his probable fraud with FTX and Alameda Research. These paragraphs are becoming monotonous. Forbes made money with articles, building up these fraudsters without critical analysis. The media juggernaut now makes money writing about the frauds, which were at least partly made possible by the endowment of trust from Forbes. In an article I wrote in 2020 about fake gurus, I talked about Sam Ovens, a man who…How Forbes Monetizes The Frauds They Create

TikTok, Drugs, Congress, and Monopoly

So ketamine is a dangerous drug. Alright, but what about alcohol? So TikTok is a national security threat. Alright, what about Meta? The declaration that one thing is unreconcilably bad and the other good or at least substantially less bad is based on little to no data. In the congressional hearings about TikTok, most complaints from the U.S. ruling class apply to all social media platforms. The main complaints also apply to all major internet companies monetized from collecting consumer data. The claims of TikTok being a national security threat are much the same arbitrary arguments as what drugs are made illegal. Ketamine, is an order of magnitude less harmful than alcohol. If you want to argue that fact, first check with any emergency medicine doctor or read the analyses of the Drug harms in the U.K paper, but I digress. My claim for TikTok is that it’s pretty much the same U.S. national security risk as Facebook and Google. That TikTok is a Chinese company may pose some security risks to U.S. interests not posed by other popular apps. Shoshana Wodinsky created a list of all the unique crimes the company was accused of committing during the, to use a line from Reason Magazine, masturbatory display of political theater. Very few accusations are unique to TikTok. But, to the extent that members of congress know about data brokers, I assume they feel the U.S. data brokers’ selling lists of all the closeted gay people, with cancer, taking anti depressants in Kentucky…TikTok, Drugs, Congress, and Monopoly