Commit
·
b5ce4fe
1
Parent(s):
35b83f9
README.md
Browse files
README.md
CHANGED
@@ -10,15 +10,22 @@ tags:
|
|
10 |
|
11 |
## Model description
|
12 |
|
13 |
-
|
14 |
|
15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
-
More information needed
|
18 |
|
19 |
## Training and evaluation data
|
20 |
|
21 |
-
|
22 |
|
23 |
## Training procedure
|
24 |
|
@@ -37,4 +44,8 @@ The following hyperparameters were used during training:
|
|
37 |
|
38 |

|
39 |
|
40 |
-
</details>
|
|
|
|
|
|
|
|
|
|
10 |
|
11 |
## Model description
|
12 |
|
13 |
+
Implementing RetinaNet: Focal Loss for Dense Object Detection.
|
14 |
|
15 |
+
This repo contains the model for the notebook [**Object Detection with RetinaNet**](https://keras.io/examples/vision/retinanet/)
|
16 |
+
|
17 |
+
Here the model is tasked with localizing the objects present in an image, and at the same time, classifying them into different categories. In this, RetinaNet has been implemented, a popular `single-stage detector`, which is accurate and runs fast. RetinaNet uses a `feature pyramid network` to efficiently detect objects at multiple scales and introduces a new loss, the `Focal loss function`, to alleviate the problem of the extreme foreground-background class imbalance.
|
18 |
+
|
19 |
+
Full credits go to [**Srihari Humbarwadi**](https://twitter.com/srihari_rh)
|
20 |
+
|
21 |
+
## References
|
22 |
+
* [RetinaNet Paper](https://arxiv.org/abs/1708.02002)
|
23 |
+
* [Feature Pyramid Network Paper](https://arxiv.org/abs/1612.03144)
|
24 |
|
|
|
25 |
|
26 |
## Training and evaluation data
|
27 |
|
28 |
+
The dataset used here is a [COCO2017 dataset](https://github.com/srihari-humbarwadi/datasets/releases/download/v0.1.0/data.zip)
|
29 |
|
30 |
## Training procedure
|
31 |
|
|
|
44 |
|
45 |

|
46 |
|
47 |
+
</details>
|
48 |
+
|
49 |
+
<center>
|
50 |
+
Model Reproduced By <u><a href="https://github.com/robotjellyzone"><b>Kavya Bisht</b></a></u>
|
51 |
+
</center>
|