Spaces:
Runtime error
Runtime error
Update README.md
Browse files
README.md
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
---
|
2 |
title: Financial Bot
|
3 |
-
emoji:
|
4 |
colorFrom: red
|
5 |
-
colorTo:
|
6 |
sdk: gradio
|
7 |
sdk_version: 4.16.0
|
8 |
app_file: app.py
|
@@ -10,4 +10,16 @@ pinned: false
|
|
10 |
license: mit
|
11 |
---
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
title: Financial Bot
|
3 |
+
emoji: π
|
4 |
colorFrom: red
|
5 |
+
colorTo: green
|
6 |
sdk: gradio
|
7 |
sdk_version: 4.16.0
|
8 |
app_file: app.py
|
|
|
10 |
license: mit
|
11 |
---
|
12 |
|
13 |
+
This is the Inference module of a 3-part FTI feature-training-inference RAG-framework LLMOps course. \
|
14 |
+
In this iteration, I've replaced Falcon 7B Instruct with the currently-SoTa (Jan '24) Mistral-7B-Instruct-v0.2, \
|
15 |
+
fine-tuned using Unsloth on financial questions and answers generated with the help of GPT-4, quantized \
|
16 |
+
and augmented with a 4bit QLoRa. \
|
17 |
+
\
|
18 |
+
Prompt analysis and model registry is handled by Comet LLM, and finance news is pulled via an Alpaca API, processed \
|
19 |
+
by Bytewax, and then sent as a vector embedding to Qdrant's serverless vector store. LangChain chains the prompt and \
|
20 |
+
most relevant news article to provide answers with real-time finance information embedded within the output. \
|
21 |
+
\
|
22 |
+
#TODO: Add citations to output to show end-user which article has been used to generate the output.
|
23 |
+
|
24 |
+
I have contributed to the original MIT licensed (ka-ching!) course which can be found here:
|
25 |
+
https://medium.com/decoding-ml/the-llms-kit-build-a-production-ready-real-time-financial-advisor-system-using-streaming-ffdcb2b50714
|