Here’s how ChatGPT maker OpenAI plans to deter election misinformation in 2024

FILE - OpenAI CEO Sam Altman speaks before a Senate Judiciary Subcommittee on Privacy, Technology and the Law hearing on artificial intelligence, May 16, 2023, on Capitol Hill in Washington. ChatGPT maker OpenAI has outlined a plan, spelled out in a blog post on Monday, Jan. 15, 2024, to prevent its tools from being used to spread election misinformation as voters in more than 50 countries around the world prepare to vote in national elections in 2024. (AP Photo/Patrick Semansky, File)

NEW YORK — ChatGPT maker OpenAI has outlined a plan to prevent its tools from being used to spread election misinformation as voters in more than 50 countries prepare to cast their ballots in national elections this year.

The safeguards spelled out by the San Francisco-based artificial intelligence startup in a blog post this week include a mix of preexisting policies and newer initiatives to prevent the misuse of its wildly popular generative AI tools. They can create novel text and images in seconds but also be weaponized to concoct misleading messages or convincing fake photographs.


The steps will apply specifically to OpenAI, only one player in an expanding universe of companies developing advanced generative AI tools. The company, which announced the moves Monday, said it plans to “continue our platform safety work by elevating accurate voting information, enforcing measured policies, and improving transparency.”

It said it will ban people from using its technology to create chatbots that impersonate real candidates or governments, to misrepresent how voting works or to discourage people from voting. It said that until more research can be done on the persuasive power of its technology, it won’t allow its users to build applications for the purposes of political campaigning or lobbying.

Starting “early this year,” OpenAI said, it will digitally watermark AI images created using its DALL-E image generator. This will permanently mark the content with information about its origin, making it easier to identify whether an image that appears elsewhere on the web was created using the AI tool.

The company also said it is partnering with the National Association of Secretaries of State to steer ChatGPT users who ask logistical questions about voting to accurate information on that group’s nonpartisan website,

OpenAI CEO Sam Altman said that even with all of his company’s safeguards in place, his mind is not at ease.

“I think it’s good we have a lot of anxiety and are going to do everything we can to get it as right as we can,” he said during an interview Tuesday at a Bloomberg event during the World Economic Forum in Davos, Switzerland. “We’re going to have to watch this incredibly closely this year. Super tight monitoring. Super tight feedback loop.”

Leave a Reply

Your email address will not be published. Required fields are marked *


By participating in online discussions you acknowledge that you have agreed to the Star-Advertiser's TERMS OF SERVICE. An insightful discussion of ideas and viewpoints is encouraged, but comments must be civil and in good taste, with no personal attacks. If your comments are inappropriate, you may be banned from posting. To report comments that you believe do not follow our guidelines, email