Update app.py
Browse files
app.py
CHANGED
@@ -6,8 +6,6 @@ from dotenv import load_dotenv
|
|
6 |
from langchain_groq import ChatGroq
|
7 |
from langchain_core.output_parsers import StrOutputParser
|
8 |
from langchain_core.prompts import ChatPromptTemplate
|
9 |
-
from transformers import BarkModel, AutoProcessor, AutoModelForTextToWaveform, pipeline
|
10 |
-
import scipy
|
11 |
|
12 |
load_dotenv()
|
13 |
client = Groq(api_key=os.getenv('GROQ_API_KEY'))
|
@@ -70,9 +68,4 @@ def answer(user_question):
|
|
70 |
answer = chain.invoke({'question': user_question})
|
71 |
return answer
|
72 |
|
73 |
-
synthesiser = pipeline("text-to-speech", "suno/bark")
|
74 |
-
|
75 |
-
speech = synthesiser("Hello, my dog is cooler than you!", forward_params={"do_sample": True})
|
76 |
-
|
77 |
-
scipy.io.wavfile.write("bark_out.wav", rate=speech["sampling_rate"], data=speech["audio"])
|
78 |
frontend()
|
|
|
6 |
from langchain_groq import ChatGroq
|
7 |
from langchain_core.output_parsers import StrOutputParser
|
8 |
from langchain_core.prompts import ChatPromptTemplate
|
|
|
|
|
9 |
|
10 |
load_dotenv()
|
11 |
client = Groq(api_key=os.getenv('GROQ_API_KEY'))
|
|
|
68 |
answer = chain.invoke({'question': user_question})
|
69 |
return answer
|
70 |
|
|
|
|
|
|
|
|
|
|
|
71 |
frontend()
|