Recently had an idea about training a model on memes and making a simple classification model. Memes can be dark, intelligent, puns, and all sorts of things. If a model is trained on such a labelled dataset of memes, it can better classify a future m
See More
Ripudaman Singh
Trying to build big,... • 7m
Yes it’s entirely possible right. That’s the whole point of training a model. Once it is trained, it can be fine tuned to accomodate a person’s interests,opinions and preferences
0 replies
More like this
Recommendations from Medial
Ayush Tripathi
•
Medial • 2m
one of the best articles that I read on deepseek and it's affect on the nvidia stock, it explains in details of how the models is trained and fine tuned to have such strong logical thinking.
0 replies8 likes
Aditya Karnam
Hey I am on Medial • 12d
"Just fine-tuned LLaMA 3.2 using Apple's MLX framework and it was a breeze! The speed and simplicity were unmatched. Here's the LoRA command I used to kick off training:
```
python lora.py \
--train \
--model 'mistralai/Mistral-7B-Instruct-v0.2' \
-
Fine-Tuning: The Secret Sauce of AI Magic!
Ever wonder how AI gets so smart? It’s all about fine-tuning!
Imagine a pre-trained model as a genius with general knowledge. 🧠✨ Fine-tuning takes that genius and hones its skills for a specific task, li
See More
1 replies4 likes
Bhoop singh Gurjar
AI Deep Explorer | f... • 4d
"A Survey on Post-Training of Large Language Models"
This paper systematically categorizes post-training into five major paradigms:
1. Fine-Tuning
2. Alignment
3. Reasoning Enhancement
4. Efficiency Optimization
5. Integration & Adaptation
1️⃣ Fin
🌐 Russia Hits Google with a $2.5 Decillion Fine! 🌐
Russia has slapped Google with a jaw-dropping fine of $2.5 decillion for not restoring 17 Russian TV channels on YouTube. To put it simply, that’s a number so astronomical, it surpasses the total
See More
0 replies3 likes
Vishu Bheda
•
Medial • 4m
Jensen Huang, the CEO of NVIDIA, describes how AI is advancing in three key dimensions:
1. Pre-training: This is like getting a college degree. AI models are trained on massive datasets to develop broad, general knowledge about the world.
2. Post-
🤔 𝐎𝐩𝐞𝐧𝐀𝐈 𝐨𝟏 - 𝐢𝐬 𝐢𝐭 𝐦𝐨𝐫𝐞 𝐛𝐢𝐠𝐠𝐞𝐫 𝐨𝐫 𝐦𝐨𝐫𝐞 𝐟𝐢𝐧𝐞-𝐭𝐮𝐧𝐞𝐝?
We're all excited about OpenAI's o1 model and many other such bigger models, but here's what keeps me up at night: Are we witnessing a genuinely larger, more a
See More
4 replies13 likes
Pink Cakes
Deliver tasty browni... • 2m
Exposed: The Truth Behind Volkai – Is It Truly India Made AI?
So, I recently dug into the buzz around Volkai, the AI company gaining fame as India’s very own innovation in the AI space. Everyone's been hyping it up for being completely built from
Huge announcement from Meta. Welcome Llama 3.1🔥
This is all you need to know about it:
The new models:
- The Meta Llama 3.1 family of multilingual large language models (LLMs) is a collection of pre-trained and instruction-tuned generative models
Microsoft recently revealed VASA-1, an impressive generative AI model that can turn a single still photo into a believable video. That's fuckin scary.
Here is simple explanation of how does it work:
Essentially, VASA-1 examines a still image and us