Видео недоступно для предпросмотра
Смотреть в Telegram
Huge announcement from Meta. Welcome Llama 3.1!
This is all you need to know about it:
The new models:
- The Meta Llama 3.1 family of multilingual large language models (LLMs) is a collection of pre-trained and instruction-tuned generative models in 8B, 70B, and 405B sizes (text in/text out).
- All models support long context length (128k) and are optimized for inference with support for grouped query attention (GQA).
- Optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.
- Llama 3.1 is an auto-regressive language model with an optimized transformer architecture, using SFT and RLHF for alignment. Its core LLM architecture is the same dense structure as Llama 3 for text input and output.
- Tool use, Llama 3.1 Instruct Model (Text) is fine-tuned for tool use, enabling it to generate tool calls for search, image generation, code execution, and mathematical reasoning, and also supports zero-shot tool use.
This is all you need to know about it:
The new models:
- The Meta Llama 3.1 family of multilingual large language models (LLMs) is a collection of pre-trained and instruction-tuned generative models in 8B, 70B, and 405B sizes (text in/text out).
- All models support long context length (128k) and are optimized for inference with support for grouped query attention (GQA).
- Optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.
- Llama 3.1 is an auto-regressive language model with an optimized transformer architecture, using SFT and RLHF for alignment. Its core LLM architecture is the same dense structure as Llama 3 for text input and output.
- Tool use, Llama 3.1 Instruct Model (Text) is fine-tuned for tool use, enabling it to generate tool calls for search, image generation, code execution, and mathematical reasoning, and also supports zero-shot tool use.