Back

🤔 𝐎𝐩𝐞𝐧𝐀𝐈 𝐨𝟏 - 𝐢𝐬 𝐢𝐭 𝐦𝐨𝐫𝐞 𝐛𝐢𝐠𝐠𝐞𝐫 𝐨𝐫 𝐦𝐨𝐫𝐞 𝐟𝐢𝐧𝐞-𝐭𝐮𝐧𝐞𝐝? We're all excited about OpenAI's o1 model and many other such bigger models, but here's what keeps me up at night: Are we witnessing a genuinely larger, more a

See More

Tanuj Sharma

Python Developer | A... • 2m

Hi, It is trained on more bigger dataset and with more gpu hours so its not fine tuned on earlier versions, its trained with more parameters than the earliest versions, if you have heard its training cost is around 50-100M $, while if you see thier latest offering o3, its training cost till now around 1-1.4B $, while training cost for chat gpt initial version was only around ~10M $, but as per the latest rumours if you heard they already have exhausted all available data on internet in training thier models till now, so future mai its going to be very hard for them to train their models on new data, then it would be mostly fine tuning on older versions with synthetic datasets.

0 replies

More like this

Recommendations from Medial

Image Description

Varun reddy

 • 

GITAM • 10m

Fine-Tuning: The Secret Sauce of AI Magic! Ever wonder how AI gets so smart? It’s all about fine-tuning! Imagine a pre-trained model as a genius with general knowledge. 🧠✨ Fine-tuning takes that genius and hones its skills for a specific task, li

See More
1 replies4 likes
Image Description
Image Description

Harsh Dwivedi

 • 

Medial • 6m

LIVE on MEDIAL - APPLE yearly mega launch is here. Tune to find out everything about the latest of new versions of iPhone MacBooks watches and more. Get the experience of an Apple launch event.

9 replies42 likes
2

Ayush Tripathi

 • 

Medial • 1m

one of the best articles that I read on deepseek and it's affect on the nvidia stock, it explains in details of how the models is trained and fine tuned to have such strong logical thinking.

0 replies8 likes

Rohan Saha

complicated Investor • 16d

Grok 3 has rapidly advanced its training on X, positioning itself as one of the most sophisticated AI tools available today.

0 replies3 likes

Bhoop singh Gurjar

AI Deep Explorer | f... • 2d

"A Survey on Post-Training of Large Language Models" This paper systematically categorizes post-training into five major paradigms: 1. Fine-Tuning 2. Alignment 3. Reasoning Enhancement 4. Efficiency Optimization 5. Integration & Adaptation 1️⃣ Fin

See More
0 replies7 likes
1
Anonymous
Image Description
Image Description

I have an idea to build Indian alternative to Scale ai, mechanical turk for outsourcing data labeling work for AI training. Focusing on dataset creation on Indian context. IMO this is a necessity rn as most AI models are trained on western data sou

See More
3 replies3 likes
2
Anonymous
Image Description
Image Description

What is the cheap shipping method from china to india (alibaba) I orderd a product worth 16k and its cost around 5k for DHL Express delivery I’m looking for a solution that’s very cheap

3 replies5 likes
Image Description
Image Description

Vivek kumar

On medial • 3m

Perplexity AI vs. ChatGPT 1. Purpose: Perplexity AI: A conversational search engine designed to answer questions directly and generate relevant search results using AI. ChatGPT: A general-purpose AI language model designed for interactive conversa

See More
7 replies12 likes
3
Image Description
Image Description

Ripudaman Singh

Trying to build big,... • 7m

Recently had an idea about training a model on memes and making a simple classification model. Memes can be dark, intelligent, puns, and all sorts of things. If a model is trained on such a labelled dataset of memes, it can better classify a future m

See More
9 replies8 likes
Image Description
Image Description

Vishu Bheda

 • 

Medial • 4m

Jensen Huang, the CEO of NVIDIA, describes how AI is advancing in three key dimensions: 1. Pre-training: This is like getting a college degree. AI models are trained on massive datasets to develop broad, general knowledge about the world. 2. Post-

See More
5 replies13 likes
2

Download the medial app to read full posts, comements and news.