7jimmy commited on
Commit
b76f032
·
1 Parent(s): 1c0296c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -78
README.md CHANGED
@@ -1,78 +1,10 @@
1
- # ask-question-image-web-app-streamlit-langchain
2
-
3
-
4
- <p align="center">
5
- <a href="https://www.youtube.com/watch?v=71EOM5__vkI">
6
- <img width="600" src="https://utils-computervisiondeveloper.s3.amazonaws.com/thumbnails/with_play_button/ask_question_image.jpg" alt="Watch the video">
7
- </br>Watch on YouTube: Ask questions to an image using Python, Streamlit and Langchain !
8
- </a>
9
- </p>
10
-
11
- This is a Streamlit application that allows users to ask questions about an uploaded image and receive responses from a conversational AI agent. The agent uses the OpenAI GPT-3.5 Turbo model to generate answers based on the provided image and user input.
12
-
13
- ## installation
14
-
15
- 1. Clone the repository:
16
-
17
- git clone https://github.com/your-username/image-question-answering.git
18
-
19
- 2. Change to the project directory:
20
-
21
- cd ask-question-image-web-app-streamlit-langchain
22
-
23
- 3. Install the required dependencies:
24
-
25
- pip install -r requirements.txt
26
-
27
- 4. Obtain an **OpenAI API key**. You can sign up for an API key at [OpenAI](https://platform.openai.com).
28
-
29
- 5. Replace the placeholder API key in the main.py file with your actual OpenAI API key:
30
-
31
- llm = ChatOpenAI(
32
- openai_api_key='YOUR_API_KEY',
33
- temperature=0,
34
- model_name="gpt-3.5-turbo"
35
- )
36
-
37
- 6. Run the Streamlit application:
38
-
39
- streamlit run main.py
40
-
41
- 7. Open your web browser and go to http://localhost:8501 to access the application.
42
-
43
- ## usage
44
-
45
- 1. Upload an image by clicking the file upload button.
46
-
47
- 2. The uploaded image will be displayed.
48
-
49
- 3. Enter a question about the image in the text input field.
50
-
51
- 4. The conversational AI agent will generate a response based on the provided question and image.
52
-
53
- 5. The response will be displayed below the question input.
54
-
55
- ## tools
56
-
57
- The application utilizes the following custom tools:
58
-
59
- - **ImageCaptionTool**: Generates a textual caption for the uploaded image.
60
- - **ObjectDetectionTool**: Performs object detection on the uploaded image and identifies the objects present.
61
-
62
- ## contributing
63
-
64
- Contributions are welcome! If you have any ideas, improvements, or bug fixes, please submit a pull request.
65
-
66
- ## license
67
-
68
- This project is licensed under the MIT License.
69
-
70
- ## acknowledgements
71
-
72
- This project uses the OpenAI GPT-3.5 Turbo model. Visit [OpenAI](https://openai.com/) for more information.
73
-
74
- The Streamlit library is used for building the interactive user interface. Visit the [Streamlit documentation](https://docs.streamlit.io/) for more information.
75
-
76
- The LangChain library is used for managing the conversational AI agent and tools. Visit the [LangChain GitHub repository](https://github.com/hwchase17/langchain) for more information.
77
-
78
- The Transformers library is used to inference the AI features. Visit [this](https://huggingface.co/Salesforce/blip-image-captioning-large) and [this](https://huggingface.co/facebook/detr-resnet-50) pages for a more comprehensive description of the models used.
 
1
+ ---
2
+ title: Ask To Image
3
+ emoji: 🚀
4
+ colorFrom: 'green'
5
+ colorTo: 'blue'
6
+ sdk: streamlit
7
+ sdk_version: 4.11.3
8
+ app_file: my_app.py
9
+ pinned: true
10
+ ---