Spaces:
Running
Running
Update app[tapex].py
Browse files- app[tapex].py +1 -7
app[tapex].py
CHANGED
@@ -32,13 +32,7 @@ def response(user_question, table_data):
|
|
32 |
|
33 |
# Experiment with generation parameters
|
34 |
outputs = model.generate(
|
35 |
-
**encoding
|
36 |
-
#num_beams=5, # Beam search to generate more diverse responses
|
37 |
-
#top_k=50, # Top-k sampling for diversity
|
38 |
-
#top_p=0.95, # Nucleus sampling
|
39 |
-
#temperature=0.7, # Temperature scaling (if supported by the model)
|
40 |
-
#max_length=50, # Limit the length of the generated response
|
41 |
-
#early_stopping=True # Stop generation when an end token is generated
|
42 |
)
|
43 |
|
44 |
ans = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|
|
|
32 |
|
33 |
# Experiment with generation parameters
|
34 |
outputs = model.generate(
|
35 |
+
**encoding
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
)
|
37 |
|
38 |
ans = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|