bniladridas commited on
Commit
3fd4e2d
verified
1 Parent(s): 404beb7

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +27 -77
  2. config.json +2 -2
  3. pytorch_model.bin +2 -2
README.md CHANGED
@@ -3,98 +3,48 @@ language: en
3
  license: mit
4
  tags:
5
  - conversational-ai
 
6
  - nlp
7
- - transformer
8
- - distilbert
9
  datasets:
10
- - conversational
11
  metrics:
12
- - perplexity
13
- - semantic-similarity
14
  model-index:
15
  - name: Conversational AI Base Model
16
- results: []
 
 
 
 
 
 
 
 
 
 
17
  ---
18
 
19
  # Conversational AI Base Model
20
 
21
- <p align="center">
22
- <a href="https://huggingface.co/bniladridas/conversational-ai-base-model">
23
- <img src="https://huggingface.co/front/assets/huggingface_logo-noborder.svg" width="200" alt="Hugging Face">
24
- </a>
25
- </p>
26
-
27
- ## Model Details
28
- - **Developed by:** Niladri Das
29
- - **Model type:** Conversational AI Transformer
30
- - **Language:** English
31
- - **License:** MIT
32
-
33
  ## Model Description
34
- A lightweight, flexible conversational AI model built on DistilBERT architecture, designed for advanced natural language understanding and generation.
35
-
36
- ### Model Specifications
37
- - **Base Architecture:** DistilBERT
38
- - **Model Size:** Reduced from full BERT
39
- - **Inference Speed:** Optimized for quick responses
40
- - **Context Window:** 512 tokens
41
 
42
- ## 馃殌 Quick Start
43
-
44
- ### Installation
45
- ```bash
46
- pip install transformers torch
47
- ```
48
 
49
- ### Python Example
50
  ```python
51
- from transformers import AutoModel, AutoTokenizer
52
 
53
- model = AutoModel.from_pretrained('bniladridas/conversational-ai-base-model')
54
  tokenizer = AutoTokenizer.from_pretrained('bniladridas/conversational-ai-base-model')
55
-
56
- # Tokenize input
57
- inputs = tokenizer("What is conversational AI?", return_tensors="pt")
58
-
59
- # Generate output
60
- outputs = model(**inputs)
61
  ```
62
 
63
- ## 馃 Model Capabilities
64
- - Context-aware response generation
65
- - Semantic understanding
66
- - Flexible transformer-based architecture
67
-
68
- ## 鈿狅笍 Limitations
69
  - Primarily trained on English text
70
- - May have inherent biases from training data
71
- - Requires careful evaluation for specific use cases
72
-
73
- ## 馃搳 Training Data
74
- Trained on a diverse corpus of conversational and informative text, focusing on natural language understanding.
75
-
76
- ## 馃 Ethical Considerations
77
- - Designed with fairness in mind
78
- - Transparent about model capabilities and limitations
79
- - Ongoing work to reduce potential biases
80
-
81
- ## 馃摎 Citation
82
- ```bibtex
83
- @misc{conversational-ai-model,
84
- title={Conversational AI Base Model},
85
- author={Niladri Das},
86
- year={2025},
87
- url={https://huggingface.co/bniladridas/conversational-ai-base-model}
88
- }
89
- ```
90
-
91
- ## 馃摓 Contact
92
- - GitHub: [bniladridas](https://github.com/bniladridas)
93
- - Hugging Face: [@bniladridas](https://huggingface.co/bniladridas)
94
-
95
- ## 馃敆 Links
96
- - [Personal Website](https://bniladridas.vercel.app)
97
-
98
- ---
99
-
100
- *Last Updated: February 2025*
 
3
  license: mit
4
  tags:
5
  - conversational-ai
6
+ - question-answering
7
  - nlp
8
+ - transformers
 
9
  datasets:
10
+ - squad
11
  metrics:
12
+ - exact_match
13
+ - f1_score
14
  model-index:
15
  - name: Conversational AI Base Model
16
+ results:
17
+ - task:
18
+ type: question-answering
19
+ dataset:
20
+ name: squad
21
+ type: question-answering
22
+ metrics:
23
+ - type: exact_match
24
+ value: 0.75
25
+ - type: f1_score
26
+ value: 0.85
27
  ---
28
 
29
  # Conversational AI Base Model
30
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  ## Model Description
32
+ A flexible, context-aware conversational AI model built on DistilBERT architecture.
 
 
 
 
 
 
33
 
34
+ ### Key Features
35
+ - Advanced response generation
36
+ - Context tracking
37
+ - Fallback mechanisms
38
+ - Supports multiple response strategies
 
39
 
40
+ ## Usage
41
  ```python
42
+ from transformers import AutoModelForQuestionAnswering, AutoTokenizer
43
 
44
+ model = AutoModelForQuestionAnswering.from_pretrained('bniladridas/conversational-ai-base-model')
45
  tokenizer = AutoTokenizer.from_pretrained('bniladridas/conversational-ai-base-model')
 
 
 
 
 
 
46
  ```
47
 
48
+ ## Limitations
 
 
 
 
 
49
  - Primarily trained on English text
50
+ - Requires domain-specific fine-tuning
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
- "DistilBertModel"
6
  ],
7
  "attention_dropout": 0.1,
8
  "dim": 768,
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased-distilled-squad",
3
  "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForQuestionAnswering"
6
  ],
7
  "attention_dropout": 0.1,
8
  "dim": 768,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1df15ef6ca103a52df977eec51dd1058d5f6a2fdf5b3ae5d2e7fc225e9801143
3
- size 265483293
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20e153a0baf9bf96c82063ef9232e6293d5d47f73b91c07d35535752b8b556cb
3
+ size 265491109