AI Specialist | Rese... • 9m
Revolutionizing AI with Inference-Time Scaling: OpenAI's o1 Model" Inference-time Scaling: Focuses on improving performance during inference (when the model is used) rather than just training. Reasoning through Search: The o1 model enhances reasoning by using search methods during inference. Shift in Approach: This marks a significant shift in how AI models are designed, moving beyond learning from data to active reasoning at runtime. Goal: Improve complex problem-solving by searching for relevant information during the model's use, leading to better answers.
Python Developer 💻 ... • 4m
3B LLM outperforms 405B LLM 🤯 Similarly, a 7B LLM outperforms OpenAI o1 & DeepSeek-R1 🤯 🤯 LLM: llama 3 Datasets: MATH-500 & AIME-2024 This has done on research with compute optimal Test-Time Scaling (TTS). Recently, OpenAI o1 shows that Test-
See MoreCTO @OctranTechnolog... • 3m
OpenAI has recently released their latest reasoning model, "o1-pro", in their developer APIs. It is estimated to be the most expensive AI model yet. OpenAI is charging $150 per million tokens (~750,000 words) fed into the model and $600 per million
See MoreAI Deep Explorer | f... • 2m
LLM Post-Training: A Deep Dive into Reasoning LLMs This survey paper provides an in-depth examination of post-training methodologies in Large Language Models (LLMs) focusing on improving reasoning capabilities. While LLMs achieve strong performance
See MorePassionate about Pos... • 1m
Xiaomi has introduced MiMo, its first open-source large language model, developed by the newly formed Big Model Core Team. With 7 billion parameters, MiMo excels in mathematical reasoning and code generation, matching the performance of significantly
See MoreI'm just a normal gu... • 2m
Sentient, a San Francisco-based AI development lab backed by Peter Thiel's Founder’s Fund, has unveiled its open-source AI search framework, positioning its work as a response to China's DeepSeek. Sentient released its Open Deep Search (ODS) framewo
See MoreBuilding-HatchUp.ai • 2d
OpenAI has begun using Google's tensor processing units (TPUs) to power ChatGPT and other products, marking a significant shift away from its reliance on Nvidia chips and Microsoft's data centers. This strategic move aims to reduce the high costs ass
See MoreDownload the medial app to read full posts, comements and news.