You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* By default it will compute frame-level detections and store them as well as compute frame-mean-AP in models saved at 90k and 120k iteration.
102
-
* There is a log file file created for each iteration's frame-level evaluation.
102
+
* There is a log file created for each iteration's frame-level evaluation.
103
103
104
104
##### Build tubes
105
105
You will need frame-level detections and you will need to navigate to `online-tubes`
@@ -114,12 +114,12 @@ Results are saved in `save_root/results.mat`. Additionally,`action-path` and `ac
114
114
##### frame-meanAP
115
115
To compute frame-mAP you can use `frameAP.m` script. You will need to specify `data_root`, `data_root`.
116
116
Use this script to produce results for your publication not the python one, both are almost identical,
117
-
but thier ap computation from precision and recall is slightly different.
117
+
but their ap computation from precision and recall is slightly different.
118
118
119
119
## Performance
120
120
##### UCF24 Test
121
-
Table below is similiar to [table 1 in our paper](https://arxiv.org/pdf/1611.08563.pdf). It contains more info than
122
-
that in paper, mostly about this implemenation.
121
+
The table below is similar to [table 1 in our paper](https://arxiv.org/pdf/1611.08563.pdf). It contains more info than
122
+
that in the paper, mostly about this implementation.
123
123
<tablestyle="width:100% th">
124
124
<tr>
125
125
<td>IoU Threshold = </td>
@@ -258,40 +258,37 @@ that in paper, mostly about this implemenation.
258
258
</tr>
259
259
</table>
260
260
261
-
##### Disscussion:
261
+
##### Discussion:
262
262
`Effect of training iterations:`
263
-
There is a effect of learing rate and number of itertaion
264
-
the model is trained.
265
-
If you train SSD on intial leanring rate for
263
+
There is an effect due to the choice of learning rate and the number of iterations the model is trained.
264
+
If you train the SSD network on initial learning rate for
266
265
many iterations then it performs is better on
267
266
lower IoU threshold, which is done in this case.
268
-
In orignal work using caffe implementation of SSD,
269
-
I trained SSD with 0.0005 learning rate for first 30K
270
-
iteration and dropped then learning rate by factor of 5
271
-
(divided by 5) and only trained for 45k itrations.
272
-
In this implementation all the models are trained for 120K
273
-
iterations, intial learninig rate is 0.0005 and learing is dropped by
274
-
the fastor of 5 after 70K and 90K iterations.
267
+
In original work using caffe implementation of SSD,
268
+
I trained the SSD network with 0.0005 learning rate for first 30K
269
+
iterations and dropped then learning rate by the factor of 5
270
+
(divided by 5) and further trained up to 45k iterations.
271
+
In this implementation, all the models are trained for 120K
272
+
iterations, the initial learning rate is set to 0.0005 and learning is dropped by the factor of 5 after 70K and 90K iterations.
275
273
276
274
`Kalogeiton et al. [5] ` make use mean fusion, so I thought we could try in our pipeline which was very easy to incorporate.
277
275
It is evident from above table that mean fusion performs better than other fusion techniques.
278
-
Also, their method rely on multiple frames as input in addition to post-processing of
279
-
bounding box coordinates at tubelet level.
276
+
Also, their method relies on multiple frames as input in addition to post-processing of bounding box coordinates at tubelet level.
280
277
281
278
##### Real-time aspect:
282
279
283
-
This implementation is mainly focused on producing the best number, it can be modified to tun fast.
280
+
This implementation is mainly focused on producing the best numbers (mAP) in the simplest manner, it can be modified to run faster.
284
281
There few aspect that would need changes:
285
-
- NMS is performed once in python then again in matlab; one has to do that on GPU in python
286
-
- Most of the time spent during tube generations is taken by disc operations; which can be elimnated completely.
282
+
- NMS is performed once in python then again in Matlab; one has to do that on GPU in python
283
+
- Most of the time spent during tube generations is taken by disc operations; which can be eliminated completely.
287
284
- IoU computation during action path is done multiple time just to keep the code clean that can be handled more smartly
288
285
289
-
Contact me if you want to implement real-time version.
290
-
Proper real-time version would require converting matlab part into python.
291
-
I presented the timing of indivual components in paper, which still holds.
286
+
Contact me if you want to implement the real-time version.
287
+
The Proper real-time version would require converting Matlab part into python.
288
+
I presented the timing of individual components in the paper, which still holds true.
292
289
293
290
## Extras
294
-
To use pre-trained model download the pretrained weights from the links given below and make changes in `test-ucf24.py` to accept the downloaded weights.
291
+
To use pre-trained model download the pre-trained weights from the links given below and make changes in `test-ucf24.py` to accept the downloaded weights.
295
292
296
293
##### Download pre-trained networks
297
294
- Currently, we provide the following PyTorch models:
0 commit comments