temporal segmentation of egocentric videos to highlight ...10.1007/978-3-319-46604... · temporal...

13
Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary Material Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato Department of Mathematics and Computer Science University of Catania {furnari,gfarinella,battiato}@dmi.unict.it This document is intended for the convenience of the reader and contains sup- plementary material which could not be included in the submitted paper due to space limits. Specifically, Section 1 reports and discusses additional results to allow a better assessment of the performances of the proposed method. 1 Additional Results Fig. 1 to Fig.11 report a graphical representation of the segmentation results related to the proposed method (method [c] in Table 1). Each image reports the results for the multi-class classifier alone (first row), the rejection mechanism (second row) and the HMM (third row). The last row reports the ground truth results. As outlined in the manuscript, simple class discrimination (top row) yields noisy predictions when ground truth frames are negative. The rejection mechanism successfully detects negative segments. The use of a HMM finally helps reducing sudden changes in the predicted labels. In Fig.s 12 - 21, we report the segmentation diagrams for the methods com- pared in Table1. Fig. 22 reports the segmentation diagram for the concatenation of all sequences. The diagrams are intended for qualitative and comparative as- sessment of the results. Each diagram reports the ground truth labels as well as the predicted labels for each of the considered methods. The depicted seg- mentations are the output of the HMM applied to each of the methods. The proposed methods ([c] and [f ] in Table1) in average outperform the competitors and reach remarkable performances in some cases (e.g., Fig.s 15, 16, 18, 20). As can be noted, the localization accuracy of the segmentation boundaries is slightly affected by the proposed negative rejection method. References 1. Gal, Y., Ghahramani, Z.: Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. arXiv preprint arXiv:1506.02142 (2015) 2. Torralba, A., Murphy, K.P., Freeman, W.T., Rubin, M.A.: Context-based vision system for place and object recognition. International Conference on Computer Vision (2003)

Upload: letruc

Post on 19-Mar-2018

215 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Temporal Segmentation of Egocentric Videos toHighlight Personal Locations of Interest

Supplementary Material

Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Department of Mathematics and Computer ScienceUniversity of Catania

{furnari,gfarinella,battiato}@dmi.unict.it

This document is intended for the convenience of the reader and contains sup-plementary material which could not be included in the submitted paper dueto space limits. Specifically, Section 1 reports and discusses additional results toallow a better assessment of the performances of the proposed method.

1 Additional Results

Fig. 1 to Fig.11 report a graphical representation of the segmentation resultsrelated to the proposed method (method [c] in Table 1). Each image reports theresults for the multi-class classifier alone (first row), the rejection mechanism(second row) and the HMM (third row). The last row reports the ground truthresults. As outlined in the manuscript, simple class discrimination (top row)yields noisy predictions when ground truth frames are negative. The rejectionmechanism successfully detects negative segments. The use of a HMM finallyhelps reducing sudden changes in the predicted labels.

In Fig.s 12 - 21, we report the segmentation diagrams for the methods com-pared in Table1. Fig. 22 reports the segmentation diagram for the concatenationof all sequences. The diagrams are intended for qualitative and comparative as-sessment of the results. Each diagram reports the ground truth labels as wellas the predicted labels for each of the considered methods. The depicted seg-mentations are the output of the HMM applied to each of the methods. Theproposed methods ([c] and [f ] in Table1) in average outperform the competitorsand reach remarkable performances in some cases (e.g., Fig.s 15, 16, 18, 20).As can be noted, the localization accuracy of the segmentation boundaries isslightly affected by the proposed negative rejection method.

References

1. Gal, Y., Ghahramani, Z.: Dropout as a Bayesian Approximation: RepresentingModel Uncertainty in Deep Learning. arXiv preprint arXiv:1506.02142 (2015)

2. Torralba, A., Murphy, K.P., Freeman, W.T., Rubin, M.A.: Context-based visionsystem for place and object recognition. International Conference on ComputerVision (2003)

Page 2: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

2 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Accuracy Comp. PerformancesId Settings Discrim. +Rejection +HMM Dimensions Time

[c] L ND 94.53 85.00 88.63 378 MB 13.10 ms[f ] L LR 92.31 81.00 85.37 26 MB 10.23 ms

[g] SIFT 34.64 33.16 – 71 MB 5170.1 ms[h] L ND NE 73.84 76.42 79.69 378 MB 12.82 ms[i] SVM [5] 87.76 74.14 79.64 423 MB 97.83 ms

Table 1. Comparisons with the state of the art. Architectural settings: L the con-volutional layers are locked, ND dropout is disabled, LR fully connected layers arereplaced by a single logistic regression layer, SIFT the SIFT feature matching baseline,NE the model is trained on both positive and negative samples, SVM classificationbased on one-class and multiclass SVM classifiers.

Sequence 1

Discr.

+Rej.

+HMM

GT

Fig. 1. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence1.

3. Templeman, R., Korayem, M., Crandall, D., Apu, K.: PlaceAvoider: Steering First-Person Cameras away from Sensitive Spaces. In: Annual Network and DistributedSystem Security Symposium. (2014) 23–26

4. Bishop, C.M.: Pattern recognition and Machine Learning. Springer (2006)5. Furnari, A., Farinella, G.M., Battiato, S.: Recognizing personal contexts from ego-

centric images. In: Workshop on Assistive Computer Vision and Robotics (ACVR)in conjunction with the IEEE International Conference on Computer Vision. (2015)

Page 3: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 3

Sequence 2

Discr.

+Rej.

+HMM

GT

Fig. 2. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 2.

Sequence 3

Discr.

+Rej.

+HMM

GT

Fig. 3. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 3.

Page 4: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

4 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Sequence 4

Discr.

+Rej.

+HMM

GT

Fig. 4. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 4.

Sequence 5

Discr.

+Rej.

+HMM

GT

Fig. 5. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 5.

Page 5: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 5

Sequence 6

Discr.

+Rej.

+HMM

GT

Fig. 6. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 6.

Sequence 7

Discr.

+Rej.

+HMM

GT

Fig. 7. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 7.

Page 6: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

6 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Sequence 8

Discr.

+Rej.

+HMM

GT

Fig. 8. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 8.

Sequence 9

Discr.

+Rej.

+HMM

GT

Fig. 9. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 9.

Page 7: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 7

Sequence 10

Discr.

+Rej.

+HMM

GT

Fig. 10. Segmentation results obtained with the proposed method [c] in Table 1 relatedto Sequence 10.

All sequences

S1 S2 S3 S4 S5 S6 S7 S8 S9 S10

Discr.

+Rej.

+HMM

GT

Fig. 11. Segmentation results obtained with the proposed method [c] in Table 1 relatedthe concatenation of all sequences.

Page 8: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

8 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Sequence 1

[c]

[f]

[g]

[h]

[i]

GT

Fig. 12. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 1.

Sequence 2

[c]

[f]

[g]

[h]

[i]

GT

Fig. 13. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 2.

Page 9: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 9

Sequence 3

[c]

[f]

[g]

[h]

[i]

GT

Fig. 14. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 3.

Sequence 4

[c]

[f]

[g]

[h]

[i]

GT

Fig. 15. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 4.

Page 10: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

10 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Sequence 5

[c]

[f]

[g]

[h]

[i]

GT

Fig. 16. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 5.

Sequence 6

[c]

[f]

[g]

[h]

[i]

GT

Fig. 17. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 6.

Page 11: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 11

Sequence 7

[c]

[f]

[g]

[h]

[i]

GT

Fig. 18. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 7.

Sequence 8

[c]

[f]

[g]

[h]

[i]

GT

Fig. 19. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 8.

Page 12: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

12 Antonino Furnari, Giovanni Maria Farinella, Sebastiano Battiato

Sequence 9

[c]

[f]

[g]

[h]

[i]

GT

Fig. 20. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 9.

Sequence 10

[c]

[f]

[g]

[h]

[i]

GT

Fig. 21. Comparative segmentation results of the methods reported in Table 1 relatedto Sequence 10.

Page 13: Temporal Segmentation of Egocentric Videos to Highlight ...10.1007/978-3-319-46604... · Temporal Segmentation of Egocentric Videos to Highlight Personal Locations of Interest Supplementary

Highlighting Personal Location of Interest in Egocentric Video 13

All sequences

S1 S2 S3 S4 S5 S6 S7 S8 S9 S10

[c]

[f]

[g]

[h]

[i]

GT

Fig. 22. Comparative segmentation results of the methods reported in Table 1 relatedto all sequences.