Model Details

  • Base Model: meta-llama/Llama-3.1-8B-instruct

Merger Configuration

Source Adapters

All source adapters share the following configuration:

  • Rank (r): 16
  • Alpha: 16
  • Target Modules:
    • q_proj (Query projection)
    • k_proj (Key projection)
    • v_proj (Value projection)
    • o_proj (Output projection)
    • up_proj (Upsampling projection)
    • down_proj (Downsampling projection)
    • gate_proj (Gate projection)
  • The order of loading adapters may affect the final result
  • Equal weights were chosen to maintain balanced influence from each adapter
  • The merged adapter maintains the same architecture and rank as the original adapters
  • While this adapter merges multiple fine-tunes, each component was developed as part of independent research efforts to explore and language model capabilities as part of R&D process.
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.