Back

Subhajit Mandal

Software Developerย โ€ขย 3m

1. LoRA on a reasonably small open model (best balance for small compute): apply low-rank adapters (PEFT/LoRA). Requires less GPU memory and works well for 700โ€“3000 rows. 2. Full fine-tune (costly / heavy): only if you have >A100 GPU or cloud paid GPU. Not recommended for early MVP. 3. No-fine-tune alternative (fast & free): use retrieval + prompting (RAG) โ€” keep base LLM and add context from your 3k+ rows. Great when compute is limited.

Reply

More like this

Recommendations from Medial

Image Description

Sandeep Prasad

Business Coachย โ€ขย 3m

๐Ÿ”ฅ Government set to name ~8 Indian teams for foundational model incentives next week โ€“ second-round beneficiaries may include BharatGen; GPU access remains tight as only ~17,374 of planned 34,333 GPUs are installed so far. ๐Ÿค” Why It Matters โ€“ More

See More
Reply
2
1
Image Description
Image Description

Narendra

Willing to contribut...ย โ€ขย 1m

I fine-tuned 3 models this week to understand why people fail. Used LLaMA-2-7B, Mistral-7B, and Phi-2. Different datasets. Different methods (full tuning vs LoRA vs QLoRA). Here's what I learned that nobody talks about: 1. Data quality > Data quan

See More
2 Replies
10
1

Swamy Gadila

Founder of Friday AIย โ€ขย 6m

๐Ÿšจ Open AI is an Wrapper๐Ÿ‘€๐Ÿคฏ Hot take, but letโ€™s break it down logically: OpenAI is not a full-stack AI company โ€” itโ€™s a high-level wrapper over Azure and NVIDIA. Hereโ€™s why that matters ๐Ÿ‘‡ ๐Ÿ”น 1. Infra Backbone = Microsoft Azure Almost 90%+ of Op

See More
Reply
2
4

AI Engineer

AI Deep Explorer | f...ย โ€ขย 9m

"A Survey on Post-Training of Large Language Models" This paper systematically categorizes post-training into five major paradigms: 1. Fine-Tuning 2. Alignment 3. Reasoning Enhancement 4. Efficiency Optimization 5. Integration & Adaptation 1๏ธโƒฃ Fin

See More
Reply
1
8

Rahul Agarwal

Founder | Agentic AI...ย โ€ขย 29d

SLM vs LLM โ€” which AI model is best for you? Iโ€™ve explained both in simple steps below. ๐—ฆ๐—Ÿ๐—  (๐—ฆ๐—บ๐—ฎ๐—น๐—น ๐—Ÿ๐—ฎ๐—ป๐—ด๐˜‚๐—ฎ๐—ด๐—ฒ ๐— ๐—ผ๐—ฑ๐—ฒ๐—น) (๐˜ด๐˜ต๐˜ฆ๐˜ฑ-๐˜ฃ๐˜บ-๐˜ด๐˜ต๐˜ฆ๐˜ฑ) Lightweight AI models built for speed, focus, and on-device execution. 1. ๐——๐—ฒ๐—ณ๐—ถ๐—ป๐—ฒ

See More
Reply
1
12

Swamy Gadila

Founder of Friday AIย โ€ขย 9d

From Emotional AI to Enterprise Infrastructure: The Story of Friday AI I donโ€™t know whether to call it a phobia or obsession, but when it comes to my work, I chase correctness relentlessly. Iโ€™m not perfect, but Iโ€™m consistent. Where It Started Fri

See More
Reply
3
Image Description

Soumya

Developerย โ€ขย 1y

๐Ÿ’กAn Idea to Change the Game for AI Startups: Making AI Processing Faster, Cheaper, and Effortless Running AI models like ChatGPT, DALLยทE, or AlphaCode is a computing monsterโ€”they need massive power to function, which makes them expensive to operate

See More
2 Replies
4

AI Engineer

AI Deep Explorer | f...ย โ€ขย 8m

LLM Post-Training: A Deep Dive into Reasoning LLMs This survey paper provides an in-depth examination of post-training methodologies in Large Language Models (LLMs) focusing on improving reasoning capabilities. While LLMs achieve strong performance

See More
Reply
2

Rahul Agarwal

Founder | Agentic AI...ย โ€ขย 1m

Steps to building real-world AI systems. I've given a simple detailed explanation below. ๐—ฆ๐˜๐—ฒ๐—ฝ 1 โ€“ ๐——๐—ฒ๐—ฝ๐—น๐—ผ๐˜†๐—บ๐—ฒ๐—ป๐˜ & ๐—–๐—ผ๐—บ๐—ฝ๐˜‚๐˜๐—ฒ ๐—Ÿ๐—ฎ๐˜†๐—ฒ๐—ฟ โ€ข This is where all the ๐—ต๐—ฒ๐—ฎ๐˜ƒ๐˜† ๐—ฝ๐—ฟ๐—ผ๐—ฐ๐—ฒ๐˜€๐˜€๐—ถ๐—ป๐—ด ๐—ต๐—ฎ๐—ฝ๐—ฝ๐—ฒ๐—ป๐˜€. โ€ข It provides the ๐—ต๐—ฎ๐—ฟ๏ฟฝ

See More
Reply
1
1

Download the medial app to read full posts, comements and news.