type ModelInfo = { name: string; params: number; hidden_size: number; intermediate_size: number; num_hidden_layers: number; }; export const MODELS: ModelInfo[] = [ // LLaMA 3.1 Models { name: 'LLaMA 3.1 (70B)', params: 70, hidden_size: 8192, intermediate_size: 28672, num_hidden_layers: 80, }, { name: 'LLaMA 3.1 (8B)', params: 8, hidden_size: 4096, intermediate_size: 14336, num_hidden_layers: 32, }, // LLaMA 3 Models { name: 'LLaMA 3 (70B)', params: 70, hidden_size: 8192, intermediate_size: 28672, num_hidden_layers: 80, }, { name: 'LLaMA 3 (8B)', params: 8, hidden_size: 4096, intermediate_size: 14336, num_hidden_layers: 32, }, // LLaMA 2 Models { name: 'LLaMA 2 (70B)', params: 70, hidden_size: 8192, intermediate_size: 28672, num_hidden_layers: 80, }, { name: 'LLaMA 2 (13B)', params: 13, hidden_size: 5120, intermediate_size: 13824, num_hidden_layers: 40, }, { name: 'LLaMA 2 (7B)', params: 7, hidden_size: 4096, intermediate_size: 11008, num_hidden_layers: 32, }, // Mistral Models { name: 'Mistral (13B NeuralPivot)', params: 13, hidden_size: 4096, intermediate_size: 14336, num_hidden_layers: 60, }, { name: 'Mistral (7B)', params: 7, hidden_size: 4096, intermediate_size: 14336, num_hidden_layers: 32, }, { name: 'Mistral (13B Amethyst)', params: 13, hidden_size: 5120, intermediate_size: 13824, num_hidden_layers: 40, }, // Qwen Models { name: 'Qwen (7B)', params: 7, hidden_size: 4096, intermediate_size: 22016, num_hidden_layers: 32, }, { name: 'Qwen (1.5 7B)', params: 7, hidden_size: 4096, intermediate_size: 11008, num_hidden_layers: 32, }, // Llava Models { name: 'Llava (1.6 34B)', params: 34, hidden_size: 7168, intermediate_size: 20480, num_hidden_layers: 60, }, { name: 'Llava (1.5 13B)', params: 13, hidden_size: 5120, intermediate_size: 13824, num_hidden_layers: 40, }, { name: 'Llava (7B)', params: 7, hidden_size: 4096, intermediate_size: 11008, num_hidden_layers: 32, }, // Gemma Models { name: 'Gemma (27B)', params: 27, hidden_size: 4608, intermediate_size: 36864, num_hidden_layers: 46, }, { name: 'Gemma (2.9B)', params: 2.9, hidden_size: 3584, intermediate_size: 14336, num_hidden_layers: 42, }, { name: 'Gemma (2B)', params: 2, hidden_size: 2048, intermediate_size: 16384, num_hidden_layers: 18, }, // Mixtral Models { name: 'Mixtral (46B)', params: 46, hidden_size: 4096, intermediate_size: 14336, num_hidden_layers: 32, }, ]; export default MODELS;