AI

Stanford University team apologises over claims they copied Chinese project for AI model


“We want to sincerely apologise to the original authors,” Stanford computer science undergraduates Aksh Garg and Siddharth Sharma said in a statement posted on X on Monday.

They said the original model had been taken down.

05:03

How does China’s AI stack up against ChatGPT?

How does China’s AI stack up against ChatGPT?

Llama 3-V was developed by Garg, Sharma and another researcher, Mustafa Aljadery, who is not from Stanford. The three researchers did not immediately respond to requests for comment.

Launching Llama 3-V last week, they claimed it could be trained to rival the performance of cutting-edge AI models such as GPT4-V, Gemini Ultra and Claude Opus at a cost of just under US$500.

Soon after its release, Llama 3-V made it into the top five trending list on Hugging Face, a popular artificial intelligence platform.

But questions were raised within the AI community over whether a large part of the new model might have been stolen from MiniCPM-Llama3-V 2.5. That model was jointly developed by Tsinghua University’s Natural Language Processing Lab and ModelBest, a Beijing-based AI start-up founded in 2022.

Content posted by one whistle-blower on open-source platform GitHub suggests the model structure and code of the two projects are almost identical.

Liu Zhiyuan, co-founder of ModelBest, said in a WeChat post on Monday that he was “relatively sure” that the new model had stolen from their project.

He said MiniCPM-Llama3-V2.5 had an embedded feature – it can identify bamboo slips from the Warring States Period (about 475-221 BC).

In 2008, Tsinghua University acquired 2,500 bamboo slips – Chinese texts written on strips of bamboo – from this period.

Liu’s team scanned and annotated the texts verbatim to create a dataset for training. That dataset is not publicly available, but the Llama3-V model showed the same recognition ability, according to Liu.

“Even the wrong cases are the same,” he said.

Liu said rapid development of AI could not be achieved without global open-source sharing of algorithms, data and models. He noted that their model had used the latest open-source Llama 3 from Meta as a base.

But he said the cornerstones of open-source sharing were adhering to protocols, trusting other contributors, and respecting and acknowledging the work of pioneers, which the Stanford team had “seriously undermined”.

08:54

Is China’s technology falling behind in the race for its own ChatGPT?

Is China’s technology falling behind in the race for its own ChatGPT?

In Monday’s statement, Garg and Sharma, the two Stanford students, said the third team member, Aljadery, had written all the code for the project.

“We apologise to the authors and take full responsibility for not doing the diligence to verify the originality of this work,” they said.

In a post on X on Tuesday, Christopher Manning, a professor of computer science and linguistics at Stanford University and director of the Stanford Artificial Intelligence Laboratory, said he did not have any knowledge of the case. “‘Fake it before you make it’ is an ignoble product of Silicon Valley,” he added.

The case has caused a stir on social media, particularly in China where it topped the list of hottest topics on Weibo on Tuesday. It has also prompted broader discussion of China’s progress in artificial intelligence.

Lucas Beyer, a researcher at AI research lab Google DeepMind, commented in a post on X that “such a good model” already existed – MiniCPM-Llama3-V 2.5 – but had received a lot less attention because it was not from an Ivy League university, but from a Chinese lab.

In his WeChat post, Liu from ModelBest acknowledged the “significant” gap between China’s generative AI models and top-tier Western projects such as Sora and GPT-4. But he said China had rapidly gone “from a nobody more than a decade ago to a key driver of AI technology innovation”.



Source

Related Articles

Back to top button