Table 7: Peliculas Results.
Model
Accuracy
(%)
Inference
Time(ms)
Benchmark 97.7 552
EMI-LSTM 99.2 412
EMI-GRU 99.0 413
EMI-FastGRNN 93.1 413
Table 8: Test Conditions.
Model Benchmark Ours
Processor Intel Xeon ARM v8
Number of cores 12 4
Processor Speed(GHz) 2.9 1.4
LSTM and EMI-GRU are very similar in their per-
formances but LSTM indicates an improved accuracy
score.
The test conditions of the benchmark and our
model are compared in Table 8. We implement our
solution with significantly fewer resources than the
benchmark.
8 CONCLUSION
Multiple Instance Learning and Early Stopping con-
cepts (EMI) were applied on two real-world crime
detection datasets and the feature extraction for the
same was optimized to have a faster extraction time
by sampling videos at a Sub-Nyquist rate. The
proposed models surpassed existing benchmarks and
have proven capable of being deployable on resource-
constrained technologies connected to surveillance
cameras. We achieved a maximum accuracy of
91.3%, inference time of 1.3s and a minimum model
size of 0.334MB in the UCF-Crime dataset. As far
as the Peliculas dataset is concerned, we achieved an
accuracy of 99.2% and an inference time of 412 ms.
ACKNOWLEDGEMENTS
The authors would like to acknowledge Solarillion
Foundation for its support and resources provided for
the research work carried out.
REFERENCES
Agarwal, U. and Sabitha, A. S. (2016). Time series forecast-
ing of stock market index. In 2016 1st India Interna-
tional Conference on Information Processing (IICIP).
Alexandrie, G. (2017). Surveillance cameras and crime:
a review of randomized and natural experiments.
Journal of Scandinavian Studies in Criminology and
Crime Prevention, 18(2):210–222.
Anguita, D., Ghio, A., Oneto, L., Parra, X., and Reyes-
Ortiz, J. L. (2012). Human activity recognition
on smartphones using a multiclass hardware-friendly
support vector machine. In International workshop on
ambient assisted living, pages 216–223. Springer.
Anguita, D., Ghio, A., Oneto, L., Parra, X., and Reyes-
Ortiz, J. L. (2013). A public domain dataset for human
activity recognition using smartphones. In Esann.
Brendel, W. and Todorovic, S. (2010). Activities as time
series of human postures. In European conference on
computer vision, pages 721–734. Springer.
Carreira, J. and Zisserman, A. (2017). Quo vadis, action
recognition? a new model and the kinetics dataset.
In proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pages 6299–6308.
Cui, Y., Li, Q., Nutanong, S., and Xue, C. J. (2019). Online
rare category detection for edge computing. In 2019
Design, Automation & Test in Europe Conference &
Exhibition (DATE), pages 1269–1272. IEEE.
Dalal, N. and Triggs, B. (2005). Histograms of oriented
gradients for human detection.
Dennis, D., Pabbaraju, C., Simhadri, H. V., and Jain, P.
(2018). Multiple instance learning for efficient se-
quential data classification on resource-constrained
devices. In Advances in Neural Information Process-
ing Systems, pages 10953–10964.
Farneb
¨
ack, G. (2003). Two-frame motion estimation based
on polynomial expansion. In Scandinavian conference
on Image analysis, pages 363–370. Springer.
Girdhar, R., Carreira, J., Doersch, C., and Zisserman, A.
(2019). Video action transformer network. In Pro-
ceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, pages 244–253.
Gracia, I. S., Suarez, O. D., Garcia, G. B., and Kim,
T.-K. (2015). Fast fight detection. PloS one,
10(4):e0120448.
Gupta, C., Suggala, A. S., Goyal, A., Simhadri, H. V.,
Paranjape, B., Kumar, A., Goyal, S., Udupa, R.,
Varma, M., and Jain, P. (2017). Protonn: compressed
and accurate knn for resource-scarce devices. In Pro-
ceedings of the 34th International Conference on Ma-
chine Learning-Volume 70, pages 1331–1340. JMLR.
org.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In The IEEE Con-
ference on Computer Vision and Pattern Recognition
(CVPR).
Kumar, A., Goyal, S., and Varma, M. (2017). Resource-
efficient machine learning in 2 kb ram for the internet
of things. In Proceedings of the 34th International
Conference on Machine Learning-Volume 70, pages
1935–1944. JMLR. org.
Kusupati, A., Singh, M., Bhatia, K., Kumar, A., Jain, P.,
and Varma, M. (2018). Fastgrnn: A fast, accurate, sta-
ble and tiny kilobyte sized gated recurrent neural net-
work. In Advances in Neural Information Processing
Systems, pages 9017–9028.