Back to feeds

Parampreet Singh

Stealth • 2d

3B LLM outperforms 405B LLM 🤯 Similarly, a 7B LLM outperforms OpenAI o1 & DeepSeek-R1 🤯 🤯 LLM: llama 3 Datasets: MATH-500 & AIME-2024 This has done on research with compute optimal Test-Time Scaling (TTS). Recently, OpenAI o1 shows that Test-Time Scaling (TTS) can enhance the reasoning capabilities of LLMs by allocating additional computation at inference time, which improves LLM performance. In Simple terms: "think slowly with long Chain-of-Thought." But, By generating multiple outputs on a sample and picking the best one and training model again, which eventually leads to perform 0.5B LLM better than GPT-4o. But more computation. To make it efficient, they've used search based methods with the reward-aware Compute-optimal TTS. CC Paper: Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling 🔗 https://arxiv.org/abs/2502.06703 #Openai #LLM #GPT #GPTo1 #deepseek #llama3

1 replies5 likes
Replies (1)

More like this

Recommendations from Medial

Image Description

Chirotpal Das

Stealth • 21d

I want a list of reasoning questions that openai o1 and/or deepseek r1 is failing to answer correctly. Quick help is much appreciated. Working on something and want to test it for reasoning capabilities.

1 replies5 likes

Payal Manghnani

Stealth • 5m

Meet latest game-changer OpenAI o1 – an AI model built for deep thinking and unmatched performance - Takes extra time to think before responding for more accurate answers - Ranks in the 89th percentile in competitive coding (Codeforces) - Secure

See More
0 replies3 likes
2
Image Description
Image Description

Vishu Bheda

 • 

Medial • 3m

Jensen Huang, the CEO of NVIDIA, describes how AI is advancing in three key dimensions: 1. Pre-training: This is like getting a college degree. AI models are trained on massive datasets to develop broad, general knowledge about the world. 2. Post-

See More
5 replies13 likes
2

Download the medial app to read full posts, comements and news.