Back

Llama-3.2-3B-Instruct

The Llama 3.2 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 1B and 3B sizes (text in/text out).

Sentiment analysis

In this example, we are using Llama 3.2 3B model for doing sentiment analysis for the input text. For explainability, we are using DLBacktrace.

Input

Prompt* String

how many s in mississippi. think step by step
Input text for the model
temperature

*

number

(minimum: 0, maximum: 1)

0.7
Controls randomness. Lower values make the model more deterministic, higher values make it more random. 

Default: 0.7
top_p

*

number

(minimum: 0, maximum: 1)

0.95
Controls randomness. Lower values make the model more deterministic, higher values make it more random. 

Default: 0.7
max_tokens

*

integer

(maximum: 1)

512
Maximum number of tokens to generate

Default: 0.7

Input

Prompt* String
Classify the text into neutral, negative, or positive.

Text: The product mentioned full coverage, but my claim wasn’t accepted, and the experience was not something I would recommend to others.

Sentiment:
Input text for the model
temperature
(minimum: 0, maximum: 1)
1
Controls randomness. Lower values make the model more deterministic, higher values make it more random.
top_p
(minimum: 0, maximum: 1)
1
Controls randomness. Lower values make the model more deterministic, higher values make it more random.
max_tokens
(minimum: 1, maximum: 128,000)
50
Maximum number of tokens to generate

Output

Sentiment analysis model output (Llama 3.2-3B)

GPU: A1 100
1.5mins

Response

Neutral. 

Reason: The text mentions a negative experience, but it does not express strong emotions or use language that is overly critical. The text simply states the facts of the situation, without making a value judgment or expressing a personal opinion. 

Note

Explainability: (DL Backtrace)

The Llama 3.2-3B model is part of Meta's Llama 3.2 series, which features advanced large language models (LLMs) optimized for a variety of natural language processing tasks. With approximately 3.21 billion parameters, this model is designed to handle multilingual dialogue and excels in tasks such as text generation, summarization, and agentic retrieval. It represents a significant improvement over its predecessors, offering enhanced performance across multiple industry benchmarks.Model ArchitectureLlama 3.2 employs an auto-regressive transformer architecture, which is a common framework for modern language models. This architecture allows the model to generate text by predicting the next token in a sequence based on the preceding context. The 3B variant is characterized by:

  • Parameter Count: Approximately 3.21 billion parameters, striking a balance between performance and resource efficiency.
  • Context Length: Supports a context length of up to 128k tokens, enabling it to process long sequences of text effectively.
  • Tokenization: Utilizes an improved tokenizer with a vocabulary size of 128K tokens, enhancing token efficiency and reducing the number of tokens processed compared to previous models.

Training Methodology

The training process for Llama 3.2-3B incorporates several advanced techniques:

  • Pretraining: The model was pretrained on up to 9 trillion tokens from diverse publicly available datasets, allowing it to learn a wide range of linguistic patterns and knowledge.
  • Instruction Tuning: The model undergoes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to better align its outputs with user expectations for helpfulness and safety.
  • Knowledge Distillation: Outputs from larger models (such as Llama 3.1's 8B and 70B variants) were used as targets during the pretraining phase, enhancing the learning process through knowledge distillation techniques.

Key Features

  1. Multilingual Capabilities: Llama 3.2-3B supports multiple languages, making it suitable for global applications in various linguistic contexts.
  2. High Performance: The model has demonstrated superior performance on benchmarks like MMLU (Massive Multitask Language Understanding), outperforming many existing open-source and closed chat models.
  3. Enhanced Instruction Following: With improvements in alignment and response diversity, the model exhibits better instruction-following capabilities compared to earlier versions.
  4. Grouped Query Attention (GQA): This technique has been implemented to improve inference efficiency, allowing the model to process queries more effectively.

Use Cases

Llama 3.2-3B is designed for a variety of applications:

  • Text Generation: Capable of producing coherent and contextually relevant text based on user prompts.
  • Summarization: Efficiently condenses information from longer texts into concise summaries.
  • Conversational Agents: Facilitates the development of chatbots that can engage in meaningful dialogues across different languages.

Conclusion

The Llama 3.2-3B model represents a significant advancement in the development of large language models, combining high performance with efficient resource usage. Its multilingual capabilities and enhanced instruction-following features make it an ideal choice for developers looking to build sophisticated natural language processing applications. With its robust architecture and comprehensive training methodology, Llama 3.2-3B stands out as a powerful tool in the evolving landscape of AI technologies.

Is Explainability critical for your 'AI' solutions?

Schedule a demo with our team to understand how AryaXAI can make your mission-critical 'AI' acceptable and aligned with all your stakeholders.