Huge announcement from Meta. Welcome Llama 3.1🔥 This is all you need to know about it: The new models: - The Meta Llama 3.1 family of multilingual large language models (LLMs) is a collection of pre-trained and instruction-tuned generative models in 8B, 70B, and 405B sizes (text in/text out) - All models support long context length (128k) and are optimized for inference with support for grouped query attention (GQA) - Optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks -Llama 3.1 is an auto-regressive language model with an optimized transformer architecture, using SFT and RLHF for alignment. Its core LLM architecture is the same dense structure as Llama 3 for text input and output - Tool use, Llama 3.1 Instruct Model (Text) is fine-tuned for tool use, enabling it to generate tool calls for search, image generation, code execution, and mathematical reasoning, and also supports zero-shot tool use
Download the medial app to read full posts, comements and news.