Back to feeds

Chirotpal Das

Stealthย โ€ขย 2d

๐Ÿค” ๐Ž๐ฉ๐ž๐ง๐€๐ˆ ๐จ๐Ÿ - ๐ข๐ฌ ๐ข๐ญ ๐ฆ๐จ๐ซ๐ž ๐›๐ข๐ ๐ ๐ž๐ซ ๐จ๐ซ ๐ฆ๐จ๐ซ๐ž ๐Ÿ๐ข๐ง๐ž-๐ญ๐ฎ๐ง๐ž๐? We're all excited about OpenAI's o1 model and many other such bigger models, but here's what keeps me up at night: Are we witnessing a genuinely larger, more advanced LLM, or is this the result of brilliant engineering and fine-tuning of existing architectures? ๐“๐ก๐ž ๐ซ๐ž๐š๐ฅ ๐ช๐ฎ๐ž๐ฌ๐ญ๐ข๐จ๐ง ๐ข๐ฌ: Can we, as users and developers, ever truly distinguish between a massive pre-trained model and an expertly fine-tuned one by ourselves? It's like trying to tell if a master chef created a new recipe or perfectly refined an existing one. The taste might be extraordinary either way. What do you think? ๐Ÿง 

4 replies13 likes
Replies (4)

Download the medial app to read full posts, comements and news.