kNaru LogokNaru
⌘K
전체K-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining
Issue#Humor#K-pop#Humor#Meme#Viral#Korea#Sports#Gaming

Feeds

All FeedsIssuesK-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining

Feeds

All FeedsIssuesK-Pop & EntertainmentK-Drama & ShowsDaily Life & CultureGaming & EsportsViral & MemesNews & TrendingFood & Dining
⌘K
Home/Update on Plagiarism Allegations Against Domestic AI Model
dogdrip|General•Recently

Update on Plagiarism Allegations Against Domestic AI Model

366
0
0
Post image
AI Translated Image

Upstage's Solar Open 100B model, which recently gained a lot of attention for its impressive performance in a domestic AI development project, was released on Hugging Face. However...

Post image
AI Translated Image

Sionic AI, another domestic AI firm, has raised allegations that the model is plagiarized. They claim it's actually just China's Z.ai open model, GLM-4.5-Air, with some additional fine-tuning. Here is their evidence: 1. Similarity in Model Weights In an AI model, weights are essentially the brain's information itself. The data accumulated during AI training is stored in the form of weights. However, analysis results have been released showing that the weights of Upstage's Solar Open model and GLM-4.5-Air are almost identical.

Post image
AI Translated Image

This graph compares the similarity between different layers within the GLM and Solar models versus the similarity between the same layers in both models. Generally, if they were completely different models, you wouldn't expect high similarity between the same layers. Yet, Solar and GLM showed nearly 99% cosine similarity in identical layers. In human terms, it's like having almost the exact same information stored in neurons located in the same parts of the brain. 2. Model Architecture Similarity Mixture-of-Experts (MoE), a widely used architecture nowadays, essentially means there are multiple models contained within one. Just as the brain is divided into regions for hearing, vision, and higher reasoning, an MoE model uses different internal 'experts' for fields like coding, math, conversation, or search to boost efficiency. (Note to angry experts: I've sacrificed some accuracy for the sake of a simple explanation. In reality, experts are assigned per token, which is a different concept from being split by field.) Details like the total number of internal models and how many are used at once often vary between models. However, the Solar model and the GLM model have the exact same total number of internal models and the same number of models used simultaneously.

Post image
AI Translated Image

This similarity in internal architecture suggests that the Solar model may have been derived from the GLM model. 3. Speculation on the Training Process

Post image
AI Translated Image

The company raising these doubts speculates that the training was done in this manner: They preserved the weights representing the base intelligence of the model and only retrained other components like the attention layers or the MoE router. 4. Significance The company raised concerns based on architecture and weights. In fact, referencing an existing model's architecture when developing a new one isn't really a huge deal. In the open-source ecosystem, 'yoinking' architecture ideas happens all the time and is even encouraged for the sake of AI advancement. However, the core issue is the weights. Using an existing model's weights as a starting point for further training is called fine-tuning. This is also a very common technique and isn't problematic in itself. But Upstage publicly committed to developing a 'Foundation Model' from scratch. Given that fine-tuning is inherently limited in its training scale, it doesn't fit a project aimed at securing national foundation model capabilities. This part needs to be clearly explained. 5. Upstage's Response Upstage has acknowledged these concerns and stated they will undergo a verification process of the model's training.

Post image
AI Translated Image

It looks like we'll have to wait and see the results.

AITechControversyUpstageLLMKoreaTechMachineLearning
🧐

"Users are skeptical about the 'domestic' AI model, suspecting it's just a rebranded Chinese model used to snag government subsidies. Many are calling out the hypocrisy in the tech community and demanding the money be returned if the fraud is proven."

#Mixed

Best Comments (0)

Log in to join the discussion

Knaru - Korean Community Aggregated Feed

Latest Updates

Continue Browsing

Community Vibe

🎉
🎉 Party Mode!82°

Based on 160 posts in 6h

😄 92🌈 50💪 8😠 8

Weekly Best Gallery

1화만에 난리난 네이버 웸툰 ㅋㅋㅋㅋ"최애의 아이"에서 충격적인 장면.manhwa올해들어서 한국진출 한다는 미국 패스트푸드점랄프 로렌이 디자인한 2026 밀라노 코르티나 동계올림픽 미국팀 유니폼박은영 셰프가 중식을 선택한 이유.jpgAI)여행 브이로그

Live Activity

New PostCharacteristics of those born from 2025 onwards

Community Stats

dogdrip
33%
fmkorea
23%
ruliweb
16%
ppomppu
15%
natepann
3%
dcinside
3%
clien
3%
mlbpark
1%
instiz
1%
todayhumor
1%
fmkorea_star
0%
fmkorea_movietv
0%
theqoo
0%

Trending Tags

View all →
HumorK-popHumorMemeViralKoreaSportsGamingFootballAnimeK-VarietyFood

Real-time Search

    No data available
AboutPrivacy PolicyTerms of Service

© 2024 kNaru. All rights reserved.

Related Posts

[NSFL] I've started hating AI since the New Year
dogdrip

[NSFL] I've started hating AI since the New Year

3h ago
View 177
Today's AI 'Makgora' (Deathmatch) post vibe.jpg
dogdrip

Today's AI 'Makgora' (Deathmatch) post vibe.jpg

7h ago
View 116
A Truly High-Value AI Business
dogdrip

A Truly High-Value AI Business

7h ago
View 138
Recap of the 'K-Sovereign AI Dev's Career-Ending Duel' Livestream.jpg
dogdrip

Recap of the 'K-Sovereign AI Dev's Career-Ending Duel' Livestream.jpg

8h ago
View 163
[AI] Travel Vlog
dogdrip

[AI] Travel Vlog

9h ago
View 178