zack.1989@outlook.com

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO

The remarkable success of OpenAI’s o1 series and DeepSeek-R1 has unequivocally demonstrated the power of large-scale reinforcement learning (RL) in eliciting sophisticated reasoning behaviors and significantly enhancing the capabilities of large language models (LLMs). However, the core training methodologies behind these groundbreaking reasoning models often remain veiled in their technical reports. Recent community efforts have […]

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO Read More »

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO

The remarkable success of OpenAI’s o1 series and DeepSeek-R1 has unequivocally demonstrated the power of large-scale reinforcement learning (RL) in eliciting sophisticated reasoning behaviors and significantly enhancing the capabilities of large language models (LLMs). However, the core training methodologies behind these groundbreaking reasoning models often remain veiled in their technical reports. Recent community efforts have

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO Read More »

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO

The remarkable success of OpenAI’s o1 series and DeepSeek-R1 has unequivocally demonstrated the power of large-scale reinforcement learning (RL) in eliciting sophisticated reasoning behaviors and significantly enhancing the capabilities of large language models (LLMs). However, the core training methodologies behind these groundbreaking reasoning models often remain veiled in their technical reports. Recent community efforts have

Can GRPO be 10x Efficient? Kwai AI’s SRPO Suggests Yes with SRPO Read More »

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO

Beijing, China – April 15, 2025– In a strategic move that underscores its technological prowess and global ambitions, potentially paving the way for a future IPO, Chinese AI company Zhipu.AI has announced the comprehensive open-sourcing of its next-generation General Language Models (GLM). This release includes the advanced GLM-4 series and the groundbreaking GLM-Z1 inference models,

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO Read More »

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO

Beijing, China – April 15, 2025– In a strategic move that underscores its technological prowess and global ambitions, potentially paving the way for a future IPO, Chinese AI company Zhipu.AI has announced the comprehensive open-sourcing of its next-generation General Language Models (GLM). This release includes the advanced GLM-4 series and the groundbreaking GLM-Z1 inference models,

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO Read More »

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO

Beijing, China – April 15, 2025– In a strategic move that underscores its technological prowess and global ambitions, potentially paving the way for a future IPO, Chinese AI company Zhipu.AI has announced the comprehensive open-sourcing of its next-generation General Language Models (GLM). This release includes the advanced GLM-4 series and the groundbreaking GLM-Z1 inference models,

Zhipu.AI’s Open-Source Power Play: Blazing-Fast GLM Models & Global Expansion Ahead of Potential IPO Read More »

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT

DeepSeek AI, a prominent player in the large language model arena, has recently published a research paper detailing a new technique aimed at enhancing the scalability of general reward models (GRMs) during the inference phase. Simultaneously, the company has hinted at the imminent arrival of its next-generation model, R2, building anticipation within the AI community.

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT Read More »

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT

DeepSeek AI, a prominent player in the large language model arena, has recently published a research paper detailing a new technique aimed at enhancing the scalability of general reward models (GRMs) during the inference phase. Simultaneously, the company has hinted at the imminent arrival of its next-generation model, R2, building anticipation within the AI community.

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT Read More »

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT

DeepSeek AI, a prominent player in the large language model arena, has recently published a research paper detailing a new technique aimed at enhancing the scalability of general reward models (GRMs) during the inference phase. Simultaneously, the company has hinted at the imminent arrival of its next-generation model, R2, building anticipation within the AI community.

DeepSeek Signals Next-Gen R2 Model, Unveils Novel Approach to Scaling Inference with SPCT Read More »

AI Video Generation Race Shifts from Capability to Profitability, Challenging Sora’s Dominance

Just a year after the initial explosion of interest in AI video generation, the competitive landscape is reportedly undergoing a significant transformation. The focus is shifting from simply achieving video generation capabilities to the critical challenge of demonstrating profitability. This evolution appears to be eroding the once seemingly unassailable dominant status of OpenAI’s Sora, as

AI Video Generation Race Shifts from Capability to Profitability, Challenging Sora’s Dominance Read More »

Scroll to Top