New threats to machine learning models

Machine Learning


AI data contamination: A new threat to machine learning models

A bold new initiative has been launched in the artificial intelligence industry Intentionally distorting the data used to train a machine learning model to weaken it. The project, called “Poison Fountain,” was devised by people directly involved in the development of artificial intelligence and aims to highlight the vulnerabilities of these systems and draw attention to potential threats.

The project's authors suggest that website owners place links on their websites that lead to specially crafted pages containing false or harmful information, which are then automatically collected by AI-based search engines.

This data will eventually The training set is reduced and the accuracy and quality of the resulting model is reduced. These pages may contain: Defective program code that contains hard-to-detect logic errors that can damage language models trained on this content.

The idea is largely based on research conducted by Anthropic last October. At the time, researchers concluded that: This means that even a small number of malicious documents can have a significant impact on the behavior of the language model. Poison Fountain supporters say the discovery confirms that: How easy it is to destroy modern AI systems.

According to The Register, there are five people working on the project, some of whom work for major American AI companies. One organizer, who requested anonymity, said: The threat lies not in hypothetical scenarios, but in the AI-based technologies already in place . This, he said, was the motivation for launching the project. To demonstrate how easily trust in such systems can be undermined.

The Poison Fountain website contains two links. One is a link to a traditional website, and the other is only accessible through the Tor anonymous network. Visitors are encouraged to: store and distribute malicious data; In addition to facilitating the incorporation of this information, AI training dataset. The authors do not believe in the effectiveness of regulation and believe that: Technology is already so pervasive that it is believed that an effective response must be aggressive and disruptive.

Skepticism about regulation is fueled by the following facts: The largest AI companies are actively investing in lobbying and seeking to minimize government intervention. Therefore, Poison Spring participants believe that the only possible way to stop the development of AI is to: sabotage .

According to proponents of this idea, A large-scale data distortion campaign could accelerate the collapse of an entire industry. It is believed that they are already experiencing difficulties. The community has been discussing the following symptoms for a long time: So-called “model collapse” The process by which AI begins learning from synthetic data or data that has already been processed by proprietary algorithms The ability to accurately reproduce information is lost. In a rapidly polluted information environment, such a model would look like this: increasingly unreliable.

These efforts resonate in part with other efforts aimed at preventing the unauthorized use of content. For example, in the long-running Nightshade project, artists can counter the automatic collection of images from websites by introducing subtle distortions. The algorithm will not be able to recognize them correctly.

How effective intentional poisoning of training data is remains an open question. However, the very existence of such a project There is growing concern in the IT community about the further development of artificial intelligence and the implications of its uncontrolled use.

Follow us on Google News for daily updates on cybersecurity. If you would like to report news, insights or content for publication, please contact us.

Cropped RHC 3D Transp2 1766828557 300x300

Red Hot Cyber's editorial staff is comprised of IT and cybersecurity experts and is supported by a network of qualified sources who operate in strict confidence. The team works every day to analyze, verify, and publish news, insights, and reports on cybersecurity, technology, and digital threats, with a particular focus on information accuracy and protecting sources. The information published is based on first-hand research, field experience, and original contributions from national and international operational contexts.



Source link