Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -113,15 +113,7 @@ db = Chroma(persist_directory=CHROMA_PATH, embedding_function=embeddings)
|
|
113 |
|
114 |
|
115 |
|
116 |
-
|
117 |
-
generation_config.temperature = 0
|
118 |
-
generation_config.num_return_sequences = 1
|
119 |
-
generation_config.max_new_tokens = 256
|
120 |
-
generation_config.use_cache = False
|
121 |
-
generation_config.repetition_penalty = 1.7
|
122 |
-
generation_config.pad_token_id = tokenizer.eos_token_id
|
123 |
-
generation_config.eos_token_id = tokenizer.eos_token_id
|
124 |
-
generation_config
|
125 |
|
126 |
|
127 |
prompt = """
|
@@ -160,16 +152,6 @@ stopping_criteria = StoppingCriteriaList(
|
|
160 |
[StopGenerationCriteria(stop_tokens, tokenizer, model.device)]
|
161 |
)
|
162 |
|
163 |
-
generation_pipeline = pipeline(
|
164 |
-
model=model,
|
165 |
-
tokenizer=tokenizer,
|
166 |
-
return_full_text=True,
|
167 |
-
task="text-generation",
|
168 |
-
stopping_criteria=stopping_criteria,
|
169 |
-
generation_config=generation_config,
|
170 |
-
)
|
171 |
-
|
172 |
-
llm = HuggingFacePipeline(pipeline=generation_pipeline)
|
173 |
|
174 |
|
175 |
class CleanupOutputParser(BaseOutputParser):
|
|
|
113 |
|
114 |
|
115 |
|
116 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
117 |
|
118 |
|
119 |
prompt = """
|
|
|
152 |
[StopGenerationCriteria(stop_tokens, tokenizer, model.device)]
|
153 |
)
|
154 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
155 |
|
156 |
|
157 |
class CleanupOutputParser(BaseOutputParser):
|