<script lang="ts">
	import type { Conversation } from "$lib/components/InferencePlayground/types";

	import { GENERATION_CONFIG_KEYS, GENERATION_CONFIG_SETTINGS } from "./generationConfigSettings";

	export let conversation: Conversation;
	export let classNames = "";

	const customMaxTokens: { [key: string]: number } = {
		"01-ai/Yi-1.5-34B-Chat": 2048,
		"HuggingFaceM4/idefics-9b-instruct": 2048,
		"deepseek-ai/DeepSeek-Coder-V2-Instruct": 16384,
		"bigcode/starcoder": 8192,
		"bigcode/starcoderplus": 8192,
		"HuggingFaceH4/starcoderbase-finetuned-oasst1": 8192,
		"google/gemma-7b": 8192,
		"google/gemma-1.1-7b-it": 8192,
		"google/gemma-2b": 8192,
		"google/gemma-1.1-2b-it": 8192,
		"google/gemma-2-27b-it": 8192,
		"google/gemma-2-9b-it": 4096,
		"google/gemma-2-2b-it": 8192,
		"tiiuae/falcon-7b": 8192,
		"tiiuae/falcon-7b-instruct": 8192,
		"timdettmers/guanaco-33b-merged": 2048,
		"mistralai/Mixtral-8x7B-Instruct-v0.1": 32768,
		"Qwen/Qwen2.5-72B-Instruct": 32768,
		"meta-llama/Meta-Llama-3-70B-Instruct": 8192,
		"CohereForAI/c4ai-command-r-plus-08-2024": 32768,
		"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": 32768,
		"meta-llama/Llama-2-70b-chat-hf": 8192,
		"HuggingFaceH4/zephyr-7b-alpha": 17432,
		"HuggingFaceH4/zephyr-7b-beta": 32768,
		"mistralai/Mistral-7B-Instruct-v0.1": 32768,
		"mistralai/Mistral-7B-Instruct-v0.2": 32768,
		"mistralai/Mistral-7B-Instruct-v0.3": 32768,
		"mistralai/Mistral-Nemo-Instruct-2407": 32768,
		"meta-llama/Meta-Llama-3-8B-Instruct": 8192,
		"mistralai/Mistral-7B-v0.1": 32768,
		"bigcode/starcoder2-3b": 16384,
		"bigcode/starcoder2-15b": 16384,
		"HuggingFaceH4/starchat2-15b-v0.1": 16384,
		"codellama/CodeLlama-7b-hf": 8192,
		"codellama/CodeLlama-13b-hf": 8192,
		"codellama/CodeLlama-34b-Instruct-hf": 8192,
		"meta-llama/Llama-2-7b-chat-hf": 8192,
		"meta-llama/Llama-2-13b-chat-hf": 8192,
		"OpenAssistant/oasst-sft-6-llama-30b": 2048,
		"TheBloke/vicuna-7B-v1.5-GPTQ": 2048,
		"HuggingFaceH4/starchat-beta": 8192,
		"bigcode/octocoder": 8192,
		"vwxyzjn/starcoderbase-triviaqa": 8192,
		"lvwerra/starcoderbase-gsm8k": 8192,
	} as const;

	$: modelMaxLength = customMaxTokens[conversation.model.id] ?? conversation.model.tokenizerConfig.model_max_length;
	$: maxTokens = Math.min(modelMaxLength ?? GENERATION_CONFIG_SETTINGS["max_tokens"].max, 64_000);
</script>

<div class="flex flex-col gap-y-7 {classNames}">
	{#each GENERATION_CONFIG_KEYS as key}
		{@const { label, min, step } = GENERATION_CONFIG_SETTINGS[key]}
		{@const max = key === "max_tokens" ? maxTokens : GENERATION_CONFIG_SETTINGS[key].max}
		<div>
			<div class="flex items-center justify-between">
				<label for="temperature-range" class="mb-2 block text-sm font-medium text-gray-900 dark:text-white"
					>{label}</label
				>
				<input
					type="number"
					class="w-18 rounded border bg-transparent px-1 py-0.5 text-right text-sm dark:border-gray-700"
					{min}
					{max}
					{step}
					bind:value={conversation.config[key]}
				/>
			</div>
			<input
				id="temperature-range"
				type="range"
				{min}
				{max}
				{step}
				bind:value={conversation.config[key]}
				class="h-2 w-full cursor-pointer appearance-none rounded-lg bg-gray-200 accent-black dark:bg-gray-700 dark:accent-blue-500"
			/>
		</div>
	{/each}

	<div class="mt-2">
		<label class="flex cursor-pointer items-center justify-between">
			<input type="checkbox" bind:checked={conversation.streaming} class="peer sr-only" />
			<span class="text-sm font-medium text-gray-900 dark:text-gray-300">Streaming</span>
			<div
				class="peer relative h-5 w-9 rounded-full bg-gray-200 after:absolute after:start-[2px] after:top-[2px] after:h-4 after:w-4 after:rounded-full after:border after:border-gray-300 after:bg-white after:transition-all after:content-[''] peer-checked:bg-black peer-checked:after:translate-x-full peer-checked:after:border-white peer-focus:outline-none dark:border-gray-600 dark:bg-gray-700 dark:peer-checked:bg-blue-600"
			></div>
		</label>
	</div>
</div>