NewsGuard Launches FAILSafe to Protect AI from Foreign Disinformation

NewsGuard Launches FAILSafe to Protect AI from Foreign Disinformation

NewsGuard has introduced the FAILSafe service to shield AI models from foreign influence operations, particularly targeting Russian, Chinese, and Iranian disinformation.

NewsGuard has launched a new service called the Foreign Adversary Infection of LLMs Safety Service (FAILSafe) to protect AI models from foreign influence operations, announced in a press release. This initiative comes in response to reports of a pro-Kremlin program that has infiltrated AI models with disinformation.

FAILSafe provides AI companies with real-time data verified by NewsGuard's disinformation researchers. The service includes a continuously updated feed of false narratives spread by Russian, Chinese, and Iranian influence operations, as well as a database of websites and accounts involved in these operations. This data helps AI companies prevent their systems from repeating these narratives.

Additionally, FAILSafe offers periodic stress-testing of AI products to identify the extent of disinformation infiltration and provides continuous monitoring and alerts about emerging disinformation risks. This comprehensive approach aims to safeguard AI models against the manipulation of large language models by foreign influence networks.

We hope you enjoyed this article.

Consider subscribing to one of several newsletters we publish like AI Policy Brief.

Also, consider following our LinkedIn page AI Safety & Regulation.

Subscribe to Daily AI Brief

Daily report covering major AI developments and industry news, with both top stories and complete market updates