victor-johnson commited on
Commit
253aa25
Β·
verified Β·
1 Parent(s): 1643b9a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -9
app.py CHANGED
@@ -4,7 +4,6 @@ from transformers import (
4
  pipeline,
5
  AutoTokenizer,
6
  AutoModelForCausalLM,
7
- GenerationConfig,
8
  )
9
  from langchain_huggingface import HuggingFacePipeline
10
  from langchain_core.prompts import PromptTemplate
@@ -15,19 +14,16 @@ MODEL_ID = "bigcode/starcoder2-3b"
15
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
16
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, trust_remote_code=True)
17
 
18
- # β€” Generation config & pipeline (new API) β€”
19
- gen_config = GenerationConfig(
20
- max_new_tokens=64,
21
- temperature=0.2,
22
- top_p=0.95,
23
- do_sample=False,
24
- )
25
  pipe = pipeline(
26
  "text-generation",
27
  model=model,
28
  tokenizer=tokenizer,
29
  device_map="auto",
30
- generation_config=gen_config,
 
 
 
31
  )
32
  llm = HuggingFacePipeline(pipeline=pipe)
33
 
 
4
  pipeline,
5
  AutoTokenizer,
6
  AutoModelForCausalLM,
 
7
  )
8
  from langchain_huggingface import HuggingFacePipeline
9
  from langchain_core.prompts import PromptTemplate
 
14
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
15
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, trust_remote_code=True)
16
 
17
+ # β€” Pipeline setup (pass generation parameters directly) β€”
 
 
 
 
 
 
18
  pipe = pipeline(
19
  "text-generation",
20
  model=model,
21
  tokenizer=tokenizer,
22
  device_map="auto",
23
+ max_new_tokens=64,
24
+ temperature=0.2,
25
+ top_p=0.95,
26
+ do_sample=False,
27
  )
28
  llm = HuggingFacePipeline(pipeline=pipe)
29