kNaru LogokNaru
⌘K
전체K-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining
Issue#Humor#K-pop#Humor#Meme#Viral#Korea#Sports#Gaming

Feeds

All FeedsIssuesK-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining

Feeds

All FeedsIssuesK-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining
⌘K
Home/Update on Plagiarism Allegations Against Domestic AI Model
dogdrip|General•Recently

Update on Plagiarism Allegations Against Domestic AI Model

378
0
0
Post image
AI Translated Image

Upstage's Solar Open 100B model, which recently gained a lot of attention for its impressive performance in a domestic AI development project, was released on Hugging Face. However...

Post image
AI Translated Image

Sionic AI, another domestic AI firm, has raised allegations that the model is plagiarized. They claim it's actually just China's Z.ai open model, GLM-4.5-Air, with some additional fine-tuning. Here is their evidence: 1. Similarity in Model Weights In an AI model, weights are essentially the brain's information itself. The data accumulated during AI training is stored in the form of weights. However, analysis results have been released showing that the weights of Upstage's Solar Open model and GLM-4.5-Air are almost identical.

Post image
AI Translated Image

This graph compares the similarity between different layers within the GLM and Solar models versus the similarity between the same layers in both models. Generally, if they were completely different models, you wouldn't expect high similarity between the same layers. Yet, Solar and GLM showed nearly 99% cosine similarity in identical layers. In human terms, it's like having almost the exact same information stored in neurons located in the same parts of the brain. 2. Model Architecture Similarity Mixture-of-Experts (MoE), a widely used architecture nowadays, essentially means there are multiple models contained within one. Just as the brain is divided into regions for hearing, vision, and higher reasoning, an MoE model uses different internal 'experts' for fields like coding, math, conversation, or search to boost efficiency. (Note to angry experts: I've sacrificed some accuracy for the sake of a simple explanation. In reality, experts are assigned per token, which is a different concept from being split by field.) Details like the total number of internal models and how many are used at once often vary between models. However, the Solar model and the GLM model have the exact same total number of internal models and the same number of models used simultaneously.

Post image
AI Translated Image

This similarity in internal architecture suggests that the Solar model may have been derived from the GLM model. 3. Speculation on the Training Process

Post image
AI Translated Image

The company raising these doubts speculates that the training was done in this manner: They preserved the weights representing the base intelligence of the model and only retrained other components like the attention layers or the MoE router. 4. Significance The company raised concerns based on architecture and weights. In fact, referencing an existing model's architecture when developing a new one isn't really a huge deal. In the open-source ecosystem, 'yoinking' architecture ideas happens all the time and is even encouraged for the sake of AI advancement. However, the core issue is the weights. Using an existing model's weights as a starting point for further training is called fine-tuning. This is also a very common technique and isn't problematic in itself. But Upstage publicly committed to developing a 'Foundation Model' from scratch. Given that fine-tuning is inherently limited in its training scale, it doesn't fit a project aimed at securing national foundation model capabilities. This part needs to be clearly explained. 5. Upstage's Response Upstage has acknowledged these concerns and stated they will undergo a verification process of the model's training.

Post image
AI Translated Image

It looks like we'll have to wait and see the results.

AITechControversyUpstageLLMKoreaTechMachineLearning
🧐

"Users are skeptical about the 'domestic' AI model, suspecting it's just a rebranded Chinese model used to snag government subsidies. Many are calling out the hypocrisy in the tech community and demanding the money be returned if the fraud is proven."

#Mixed

Best Comments (0)

Log in to join the discussion

#KoreaTech

Continue Browsing

Community Vibe

🎉
🎉 Party Mode!82°

Based on 115 posts in 6h

😄 69🌈 35😠 6💪 3

Weekly Best Gallery

1화만에 난리난 네이버 웸툰 ㅋㅋㅋㅋ이타치가 마이트 가이를 보고 튄 이유.JPG요리사가 항상 화가 나 있는 이유올해들어서 한국진출 한다는 미국 패스트푸드점[카연] 카르마이야기 #1~3랄프 로렌이 디자인한 2026 밀라노 코르티나 동계올림픽 미국팀 유니폼

Live Activity

New PostCurrent status of the '86-liners this year

Community Stats

dogdrip
33%
fmkorea
23%
ruliweb
16%
ppomppu
15%
natepann
3%
dcinside
3%
clien
3%
mlbpark
1%
instiz
1%
todayhumor
1%
fmkorea_star
0%
fmkorea_movietv
0%
theqoo
0%

Trending Tags

View all →
HumorK-popHumorMemeViralKoreaSportsGamingFootballAnimeK-VarietyFood

Real-time Search

    No data available
AboutPrivacy PolicyTerms of Service

© 2024 kNaru. All rights reserved.

Related Posts

The latest work from the Chinese CG master who's currently wrecking live-action movies..mp4
dogdrip

The latest work from the Chinese CG master who's currently wrecking live-action movies..mp4

1h ago
View 87
[NSFL] I've started hating AI since the New Year
dogdrip

[NSFL] I've started hating AI since the New Year

5h ago
View 211
Today's AI 'Makgora' (Deathmatch) post vibe.jpg
dogdrip

Today's AI 'Makgora' (Deathmatch) post vibe.jpg

9h ago
View 126
A Truly High-Value AI Business
dogdrip

A Truly High-Value AI Business

9h ago
View 153
Recap of the 'K-Sovereign AI Dev's Career-Ending Duel' Livestream.jpg
dogdrip

Recap of the 'K-Sovereign AI Dev's Career-Ending Duel' Livestream.jpg

10h ago
View 179