Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -4,7 +4,6 @@ from transformers import (
|
|
4 |
pipeline,
|
5 |
AutoTokenizer,
|
6 |
AutoModelForCausalLM,
|
7 |
-
GenerationConfig,
|
8 |
)
|
9 |
from langchain_huggingface import HuggingFacePipeline
|
10 |
from langchain_core.prompts import PromptTemplate
|
@@ -15,19 +14,16 @@ MODEL_ID = "bigcode/starcoder2-3b"
|
|
15 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
16 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, trust_remote_code=True)
|
17 |
|
18 |
-
# β
|
19 |
-
gen_config = GenerationConfig(
|
20 |
-
max_new_tokens=64,
|
21 |
-
temperature=0.2,
|
22 |
-
top_p=0.95,
|
23 |
-
do_sample=False,
|
24 |
-
)
|
25 |
pipe = pipeline(
|
26 |
"text-generation",
|
27 |
model=model,
|
28 |
tokenizer=tokenizer,
|
29 |
device_map="auto",
|
30 |
-
|
|
|
|
|
|
|
31 |
)
|
32 |
llm = HuggingFacePipeline(pipeline=pipe)
|
33 |
|
|
|
4 |
pipeline,
|
5 |
AutoTokenizer,
|
6 |
AutoModelForCausalLM,
|
|
|
7 |
)
|
8 |
from langchain_huggingface import HuggingFacePipeline
|
9 |
from langchain_core.prompts import PromptTemplate
|
|
|
14 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
15 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, trust_remote_code=True)
|
16 |
|
17 |
+
# β Pipeline setup (pass generation parameters directly) β
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
pipe = pipeline(
|
19 |
"text-generation",
|
20 |
model=model,
|
21 |
tokenizer=tokenizer,
|
22 |
device_map="auto",
|
23 |
+
max_new_tokens=64,
|
24 |
+
temperature=0.2,
|
25 |
+
top_p=0.95,
|
26 |
+
do_sample=False,
|
27 |
)
|
28 |
llm = HuggingFacePipeline(pipeline=pipe)
|
29 |
|