AI

do we need a label for artificial intelligence? — RealnoeVremya.com


Russia’s State Duma is thinking about how to protect the country from malicious content created by neural networks

‘Made using AI’: do we need a label for artificial intelligence?
Photo: Радиф Кашапов

Does artificial intelligence pose a threat to the emergence of a huge array of unverified information, and how to filter the content created by AI, the deputies of the State Duma are thinking. There is no salvation from the potential threat yet. Perhaps, Russia will follow the path of Western countries, obliging technology giants to independently label content created by neural networks. In general, this should be done, experts agreed, but carefully so as not to slow down the development of advanced technologies. What the minister of digital development and his virtual assistant AiratAI, created on the basis of ChatGPT, think about the legislative initiative — in the material of Realnoe Vremya.

Artificial intelligence will be slightly limited

The State Duma has begun drafting a bill on labelling content created by a neural network. The idea was announced a year ago at the Russian University of Technology. According to Anton Nemkin, a member of the State Duma Committee on Information Policy, deputies are studying the legislative experience of other countries on labelling and consulting with experts. According to the parliamentarian, at the moment it is becoming difficult for a person to distinguish the real from the unreal.

“If we consider the long-term perspective, the development of AI services without the necessary control carries the danger of huge arrays of not only unverified texts, but even completely invented facts, figures and data,” RIA Novosti quotes him.

The deputy said that at first Russian AI services should be required to automatically label any generated texts. This is due to that it is not yet clear how to recognise the degree of participation of machines and artificial intelligence in the material.

Following the path of technology leaders

The USA has already followed this path. In July 2023, the White House announced an agreement with major American technology companies to implement additional security measures in the field of AI. Among the key measures is the development of a watermark system to identify AI-generated content, which is part of the fight against misinformation and other risks associated with this rapidly evolving technology.

Seven of the largest AI companies — Amazon, Anthropic, Google, Inflation, Meta (recognized as extremist and banned in the Russian Federation), Microsoft and OpenAI — voluntarily assumed additional obligations. As part of the agreement, American tech giants must develop and implement a watermark system for visual and audio content.

Facebook*, Threads, and Instagram* applications owned by American Meta Corporation will be labelled with AI content, the company already announced. Both users and the company itself will have the opportunity to mark “Made using AI” if they detect signs of neural networks working. Previously, Meta banned videos with generated people if they did something in the video that they did not do in life.

Since March 18, YouTube content authors are required to tag their videos created using artificial intelligence. The requirement only applies to realistic content that can be misleading. The platform does not require authors to tag obviously unrealistic content created using AI. This includes animations, colour adjustments or light filters, special effects, beauty filters, or other visual enhancements.

“It is important to work out the details carefully”

“In the context of the rapid development of IT technologies, including artificial intelligence and neural networks, the issue of labelling content created by neural networks is becoming increasingly relevant. On the one hand, labelling such content can promote transparency and increase user trust. It will allow them to easily determine whether the content was created by humans or artificial intelligence. This is especially important in the context of concerns about manipulation on social networks and other online platforms,” said Tatarstan Minister of Digital Development Airat Khairullin.

“On the other hand, it is also important to take into account the potential challenges and risks associated with labelling,” the minister adds. “For example, what will happen to the content created jointly by a person and a neural network? How will it be determined what is a sufficient human contribution so that the content is not considered to be created by a neural network? Will the labelling be applied to content created for personal purposes, or only to content related to commercial purposes?”

The head of the Ministry of Finance of Tatarstan believes that the draft law on labelling content created by neural networks can be useful, but it is important to carefully work out the details of its implementation so as not to put unnecessary barriers to innovation and scientific progress.

Airat Khairullin himself considers artificial intelligence technology to be an assistant rather than a threat to people. Last February, the minister introduced a free chatbot based on artificial intelligence called AiratAI. According to the minister, it was created to help students and schoolchildren — to get tips or search for theoretical information; journalists or content creators — to create a content plan, suggest ideas for articles or write press releases; programmers — to search for errors in the code; all those who need advice in finding an interesting book, film or exhibitions.

“I don’t have a personal opinion, but I can say”

We asked AiratAI what it thinks about labelling content created by neural networks.

“I do not have a personal opinion, but I can say that the law on labelling content created by neural networks can be an important step to regulate such content and protect the interests of consumers. This can help ensure transparency about the origin of the content and prevent its abuse. However, as with any legislative initiative, it is important to consider the potential consequences and balance the interests of different parties.”

Today, labelling of content created by neural networks is mainly carried out in countries where laws and regulations in the field of artificial intelligence and automated content have been adopted, AiratAI said. For example, such measures may be taken in the United States, the European Union, Japan, South Korea and other countries that have developed technological infrastructure and are working to regulate the use of artificial intelligence in various fields.

AiratAI assesses the penetration of neural networks in Russia as fast and successful. “Every year more and more companies in Russia are starting to use neural networks in various fields such as medicine, finance, technology, marketing and others. This allows them to automate processes, improve the efficiency and quality of their work,” said the bot.

Artificial intelligence can potentially pose a threat to people if it is used improperly or unethically, the virtual assistant to the minister agreed. “For example, autonomous military AI systems can lead to an increase in military conflicts and loss of lives. There is also a danger of using AI for mass control and surveillance, which can threaten the privacy and freedom of citizens,” it cited an example.

“However, when used correctly, artificial intelligence can bring huge benefits to people by helping them solve complex problems in medicine, science, business and other fields. It is important to develop and use AI in compliance with ethical norms and laws to minimise potential threats and maximise benefits for society,” he concluded.

Bulat Ganiev, a technology entrepreneur, investor and co-founder of Technocracy, believes that such labelling is the right decision.

“I believe the idea that sites should ‘tag’ content created by artificial intelligence is good. The origin of the content is a matter of principle. Content is now not only entertainment, but also a tool of influence. If in the case of a person there is an entity responsible, then the situation is more complicated with AI content. And ultimately, this is a more honest approach towards consumers. Perhaps, it will be fundamental for someone to consume content that is created by people,” Bulat Ganiev believes.

“Tempting, but hard to implement”

“Labelling the content created by the neural network looks tempting, but is hard to implement in practice,” says Vasil Zakiev, the co-founder of Sputnik, ex-project manager of the business incubator of the Kazan and Chelyabinsk IT parks. “Even now, professionally created AI content is indistinguishable from real content, and we are at the very early stage of technology development.”

“Mandatory labelling of all AI-created content can negatively affect its perception, implying its potential unreliability and reducing the integration of AI into creative and scientific fields. This can slow down the development and implementation of advanced technologies,” Vasil Zakiev believes.

“Instead of expensive and frontal restrictions, it is better for the government to focus on raising public awareness of the opportunities and risks of AI, promoting the responsibility of AI technology developers. This will help to achieve a balanced approach to innovation and security, avoiding a negative impact on technological development,” the entrepreneur suggests.

His comment, in particular, was created using AI, Vasil added. “Did it make it worse? I don’t think so,” Zakiev noted. “The illustration on Instagram (recognised as extremist and banned in the Russian Federation) does not get worse from using filters.”

*The activities of the resources are recognised as extremist and banned in Russia.

Yulia Garaeva


Tatarstan





Source

Related Articles

Back to top button