Greg Lukianoff, President of FIRE, gives his opening statement to Congress on AI and Freedom of Speech.
My name is Greg Lukianoff, and I am the CEO of the Foundation for Individual Rights and Expression, or “FIRE,” where I’ve worked for 23 years. FIRE is a nonpartisan, nonprofit that uses litigation, scholarship, and public outreach to defend and promote the value of free speech for all Americans. We proudly defend free speech regardless of a speaker’s viewpoint or identity, and we have represented people across the political spectrum.
I’m here to address the risk AI and AI regulation pose to freedom of speech and the creation of knowledge. We have good reason to be concerned. FIRE regularly fights government attempts to stifle speech on the internet.
FIRE is in federal court challenging a New York law that forces websites to “address” online speech that someone, somewhere finds humiliating or vilifying.
We’re challenging a new Utah law that requires age verification of all social media users.
We’ve raised concerns about the federal government funding development of AI tools to target speech including microaggressions.
And later this week, FIRE will file a brief with the Supreme Court explaining the danger of “jawboning” — the use of government pressure to force social media platforms to censor protected speech.
But the most chilling threat that the government poses in the context of emerging AI is regulatory overreach that limits its potential as a tool for contributing to human knowledge.
A regulatory panic could result in a small number of Americans deciding for everyone else what speech, ideas, and even questions are permitted in the name of “safety” or “alignment.”
I’ve dedicated my life to defending freedom of speech because it is an essential human right. However, free speech is more than that; it’s nothing less than essential to our ability to understand the world.
A giant step for human progress was the realization that, despite what our senses tell us, knowledge is hard to attain.
It's a never-ending, arduous, necessarily de-centralized process of testing and retesting, of chipping away at falsity to edge a bit closer to truth.
It’s not just about the proverbial “marketplace of ideas”; it’s about allowing information—independent of idea or argument—to flow freely so that we can hope to know the world as it really is. This means seeing value in expression even when it appears to be wrongheaded or useless.
This process has been aided by new technologies that have made communication easier. From the printing press, to the telegraph and radio, to phones and the internet: each one has accelerated the development of new knowledge by making it easier to share information.
But AI offers even greater liberating potential, empowered by First Amendment principles, including freedom to code, academic freedom, and freedom of inquiry.
We are on the threshold of a revolution in the creation and discovery of knowledge.
AI’s potential is humbling; indeed, even frightening.
But as the history of the printing press shows, attempts to put the genie back in the bottle will fail.
Despite the profound disruption the printing press caused in Europe in the short term, the long-term contribution to art, science, and again, knowledge was without equal.
Yes, we may have some fears about the proliferation of AI. But what those of us who care about civil liberties fear more is a government monopoly on advanced AI.
Or, more likely, regulatory capture and a government-empowered oligopoly that privileges a handful of existing players.
The end result of pushing too hard on AI regulation will be the concentration of AI influence in an even smaller number of hands.
Far from reining in the government’s misuse of AI to censor, we will have created the framework not only to censor but also to dominate and distort the production of knowledge itself.
“But why not just let OpenAI or a handful of existing AI engines dominate the space?” you may ask.
Trust in expertise and in higher education—another important developer of knowledge—has plummeted in recent years, due largely to self-inflicted wounds borne of the ideological biases shared by much of the expert class.
That same bias is often found baked into existing AI, and without competing AI models we may create a massive body of purported official facts that we can’t actually trust.
We’ve seen on campus that attempts to regulate hate speech have led to absurd results like punishing people for simply reading about controversial topics like racism; similarly, AI programs flag or refuse to answer questions about prohibited topics.
And, of course, the potential end result of America tying the hands of the greatest programmers in the world would be to lose our advantage to our most determined foreign adversaries.
But with decentralized development and use of AI, we have a better chance of defeating our staunchest rivals or even Skynet or Big Brother.
And it’s what gives us our best chance for understanding the world without being blinded by our current orthodoxies, superstitions, or darkest fears.
Thank you for the invitation to testify and I look forward to your questions.