Back

Subhajit Mandal

Software Developer • 6m

1. LoRA on a reasonably small open model (best balance for small compute): apply low-rank adapters (PEFT/LoRA). Requires less GPU memory and works well for 700–3000 rows. 2. Full fine-tune (costly / heavy): only if you have >A100 GPU or cloud paid GPU. Not recommended for early MVP. 3. No-fine-tune alternative (fast & free): use retrieval + prompting (RAG) — keep base LLM and add context from your 3k+ rows. Great when compute is limited.

Reply

More like this

Recommendations from Medial

Image Description

Sandeep Prasad

Business Coach • 6m

šŸ”„ Government set to name ~8 Indian teams for foundational model incentives next week – second-round beneficiaries may include BharatGen; GPU access remains tight as only ~17,374 of planned 34,333 GPUs are installed so far. šŸ¤” Why It Matters – More

See More
Reply
5
1
Image Description
Image Description

Narendra

Willing to contribut... • 4m

I fine-tuned 3 models this week to understand why people fail. Used LLaMA-2-7B, Mistral-7B, and Phi-2. Different datasets. Different methods (full tuning vs LoRA vs QLoRA). Here's what I learned that nobody talks about: 1. Data quality > Data quan

See More
2 Replies
10
1
Image Description

Mridul Das

Introvert! • 29d

This is crazy šŸ¤ÆšŸ”„! Mumbai-based AI startup Neysa has raises $1.2 billion funding, around ₹10,897 crore($600M equity + $600M debt) led by Blackstone at the valuation of $1.4 billion šŸ’µ. The Startup founded in 2023 by Sharad Sanghi and Anindya Das.

See More
Reply
5
1

Bhoop Singh Gurjar

AI Deep Explorer | f... • 11m

"A Survey on Post-Training of Large Language Models" This paper systematically categorizes post-training into five major paradigms: 1. Fine-Tuning 2. Alignment 3. Reasoning Enhancement 4. Efficiency Optimization 5. Integration & Adaptation 1ļøāƒ£ Fin

See More
Reply
1
8

Swamy Gadila

Founder of Friday AI • 8m

🚨 Open AI is an WrapperšŸ‘€šŸ¤Æ Hot take, but let’s break it down logically: OpenAI is not a full-stack AI company — it’s a high-level wrapper over Azure and NVIDIA. Here’s why that matters šŸ‘‡ šŸ”¹ 1. Infra Backbone = Microsoft Azure Almost 90%+ of Op

See More
Reply
2
4

Rahul Agarwal

Founder | Agentic AI... • 3m

SLM vs LLM — which AI model is best for you? I’ve explained both in simple steps below. š—¦š—Ÿš—  (š—¦š—ŗš—®š—¹š—¹ š—Ÿš—®š—»š—“š˜‚š—®š—“š—² š— š—¼š—±š—²š—¹) (š˜“š˜µš˜¦š˜±-š˜£š˜ŗ-š˜“š˜µš˜¦š˜±) Lightweight AI models built for speed, focus, and on-device execution. 1. š——š—²š—³š—¶š—»š—²

See More
Reply
2
12

Swamy Gadila

Founder of Friday AI • 3m

From Emotional AI to Enterprise Infrastructure: The Story of Friday AI I don’t know whether to call it a phobia or obsession, but when it comes to my work, I chase correctness relentlessly. I’m not perfect, but I’m consistent. Where It Started Fri

See More
Reply
3

Bhoop Singh Gurjar

AI Deep Explorer | f... • 11m

LLM Post-Training: A Deep Dive into Reasoning LLMs This survey paper provides an in-depth examination of post-training methodologies in Large Language Models (LLMs) focusing on improving reasoning capabilities. While LLMs achieve strong performance

See More
Reply
2
Image Description

Soumya

Developer • 1y

šŸ’”An Idea to Change the Game for AI Startups: Making AI Processing Faster, Cheaper, and Effortless Running AI models like ChatGPT, DALLĀ·E, or AlphaCode is a computing monster—they need massive power to function, which makes them expensive to operate

See More
2 Replies
4

Download the medial app to read full posts, comements and news.