skip to main content
research-article

Context Sensing Attention Network for Video-based Person Re-identification

Authors Info & Claims
Published:27 February 2023Publication History
Skip Abstract Section

Abstract

Video-based person re-identification (ReID) is challenging due to the presence of various interferences in video frames. Recent approaches handle this problem using temporal aggregation strategies. In this work, we propose a novel Context Sensing Attention Network (CSA-Net), which improves both the frame feature extraction and temporal aggregation steps. First, we introduce the Context Sensing Channel Attention (CSCA) module, which emphasizes responses from informative channels for each frame. These informative channels are identified with reference not only to each individual frame, but also to the content of the entire sequence. Therefore, CSCA explores both the individuality of each frame and the global context of the sequence. Second, we propose the Contrastive Feature Aggregation (CFA) module, which predicts frame weights for temporal aggregation. Here, the weight for each frame is determined in a contrastive manner: i.e., not only by the quality of each individual frame, but also by the average quality of the other frames in a sequence. Therefore, it effectively promotes the contribution of relatively good frames. Extensive experimental results on four datasets show that CSA-Net consistently achieves state-of-the-art performance.

REFERENCES

  1. [1] Aich Abhishek, Zheng Meng, Karanam Srikrishna, Chen Terrence, Roy-Chowdhury Amit K., and Wu Ziyan. 2021. Spatio-temporal representation factorization for video-based person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 152162.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Chen Cuiqun, Ye Mang, Qi Meibin, Wu Jingjing, Liu Yimin, and Jiang Jianguo. 2022. Saliency and granularity: Discovering temporal coherence for video-based person re-identification. IEEE Transactions on Circuits and Systems for Video Technology 32, 9 (2022), 61006112. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Chen Dapeng, Li Hongsheng, Xiao Tong, Yi Shuai, and Wang Xiaogang. 2018. Video person re-identification with competitive snippet-similarity aggregation and co-attentive snippet embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 11691178.Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Chen Guangyi, Rao Yongming, Lu Jiwen, and Zhou Jie. 2020. Temporal coherence or temporal motion: Which is more critical for video-based person re-identification? In Proceedings of the European Conference on Computer Vision. Springer, 660676.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Chen Zengqun, Zhou Zhiheng, Huang Junchu, Zhang Pengyu, and Li Bo. 2020. Frame-guided region-aligned representation for video person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence. 1059110598.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Ding Changxing, Wang Kan, Wang Pengfei, and Tao Dacheng. 2022. Multi-task learning with coarse priors for robust part-aware person re-identification. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 3 (2022), 14741488.Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Eom Chanho, Lee Geon, Lee Junghyup, and Ham Bumsub. 2021. Video-based person re-identification with spatial and temporal memory networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 1203612045.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Fan Hehe, Zheng Liang, Yan Chenggang, and Yang Yi. 2018. Unsupervised person re-identification: Clustering and fine-tuning. ACM Transactions on Multimedia Computing, Communications, and Applications 14, 4 (2018), 118.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Fang Pengfei, Ji Pan, Petersson Lars, and Harandi Mehrtash. 2021. Set augmented triplet loss for video person re-identification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 464473.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Felzenszwalb Pedro F., Girshick Ross B., McAllester David, and Ramanan Deva. 2009. Object detection with discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and Machine Intelligence 32, 9 (2009), 16271645.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. [11] Fu Yang, Wang Xiaoyang, Wei Yunchao, and Huang Thomas. 2019. Sta: Spatial-temporal attention for large-scale video-based person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence. 82878294.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. [12] Gao Yajun, Liang Tengfei, Jin Yi, Gu Xiaoyan, Liu Wu, Li Yidong, and Lang Congyan. 2021. MSO: Multi-feature space joint optimization network for RGB-infrared person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia. 52575265.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Ge Wenhang, Pan Chunyan, Wu Ancong, Zheng Hongwei, and Zheng Wei-Shi. 2021. Cross-camera feature prediction for intra-camera supervised person re-identification across distant scenes. In Proceedings of the 29th ACM International Conference on Multimedia. 36443653.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Gu Xinqian, Chang Hong, Ma Bingpeng, and Shan Shiguang. 2022. Motion feature aggregation for video-based person re-identification. IEEE Transactions on Image Processing 31 (2022), 39083919. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Gu Xinqian, Chang Hong, Ma Bingpeng, Zhang Hongkai, and Chen Xilin. 2020. Appearance-preserving 3d convolution for video-based person re-identification. In Proceedings of the European Conference on Computer Vision. Springer, 228243.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. [16] Gu Xinqian, Ma Bingpeng, Chang Hong, Shan Shiguang, and Chen Xilin. 2019. Temporal knowledge propagation for image-to-video person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 96479656.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] He Kaiming, Zhang Xiangyu, Ren Shaoqing, and Sun Jian. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 770778.Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Hou Ruibing, Chang Hong, Ma Bingpeng, Huang Rui, and Shan Shiguang. 2021. BiCnet-TKS: Learning efficient spatial-temporal representation for video person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 20142023.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Hou Ruibing, Chang Hong, Ma Bingpeng, Shan Shiguang, and Chen Xilin. 2020. Temporal complementary learning for video person re-identification. In Proceedings of the European Conference on Computer Vision. Springer, 388405.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Hou Ruibing, Ma Bingpeng, Chang Hong, Gu Xinqian, Shan Shiguang, and Chen Xilin. 2019. Interaction-and-aggregation network for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 93179326.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Hou Ruibing, Ma Bingpeng, Chang Hong, Gu Xinqian, Shan Shiguang, and Chen Xilin. 2019. VRSTC: Occlusion-free video person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 71837192.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Hu Jie, Shen Li, and Sun Gang. 2018. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 71327141.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Hu Shuping, Wang Kan, Cheng Jun, Tan Huan, and Pang Jianxin. 2022. Triplet ratio loss for robust person re-identification. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision. Springer, 4254.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. [24] Li Jianing, Wang Jingdong, Tian Qi, Gao Wen, and Zhang Shiliang. 2019. Global-local temporal representations for video person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 39583967.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Li Jianing, Zhang Shiliang, and Huang Tiejun. 2019. Multi-scale 3d convolution network for video based person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 86188625.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Li Jianing, Zhang Shiliang, and Huang Tiejun. 2020. Multi-scale temporal cues learning for video person re-identification. IEEE Transactions on Image Processing 29 (2020), 44614473. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Li Mengliu, Xu Han, Wang Jinjun, Li Wenpeng, and Sun Yongli. 2020. Temporal aggregation with clip-level attention for video-based person re-identification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Li Shuang, Bak Slawomir, Carr Peter, and Wang Xiaogang. 2018. Diversity regularized spatiotemporal attention for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 369378.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Li Zhaoju, Zhou Zongwei, Jiang Nan, Han Zhenjun, Xing Junliang, and Jiao Jianbin. 2020. Spatial preserved graph convolution networks for person re-identification. ACM Transactions on Multimedia Computing, Communications, and Applications 16, 1s (2020), 114.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Liu Hao, Jie Zequn, Jayashree Karlekar, Qi Meibin, Jiang Jianguo, Yan Shuicheng, and Feng Jiashi. 2018. Video-based person re-identification with accumulative motion context. IEEE Transactions on Circuits and Systems for Video Technology 28, 10 (2018), 27882802.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. [31] Liu Jiawei, Zha Zheng-Jun, Chen Xuejin, Wang Zilei, and Zhang Yongdong. 2019. Dense 3D-convolutional neural network for person re-identification in videos. ACM Transactions on Multimedia Computing, Communications, and Applications 15, 1s (2019), 119.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. [32] Liu Liangchen, Yang Xi, Wang Nannan, and Gao Xinbo. 2021. Viewing from frequency domain: A DCT-based information enhancement network for video person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia. 227235.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. [33] Liu Xuehu, Zhang Pingping, Yu Chenyang, Lu Huchuan, and Yang Xiaoyun. 2021. Watching you: Global-guided reciprocal learning for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1333413343.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Liu Yu, Yan Junjie, and Ouyang Wanli. 2017. Quality aware network for set to set recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 57905799.Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Liu Yiheng, Yuan Zhenxun, Zhou Wengang, and Li Houqiang. 2019. Spatial and temporal mutual promotion for video-based person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 87868793.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Matiyali Neeraj and Sharma Gaurav. 2020. Video person re-identification using learned clip similarity aggregation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 26552664.Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] McLaughlin Niall, Rincon Jesus Martinez Del, and Miller Paul. 2016. Recurrent convolutional network for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 13251334.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Pang Bo, Zhai Deming, Jiang Junjun, and Liu Xianming. 2022. Fully unsupervised person re-identification via selective contrastive learning. ACM Transactions on Multimedia Computing, Communications, and Applications 18, 2 (2022), 115.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. [39] Qin Zequn, Zhang Pengyi, Wu Fei, and Li Xi. 2021. Fcanet: Frequency channel attention networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 783792.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Raychaudhuri Dripta S. and Roy-Chowdhury Amit K.. 2020. Exploiting temporal coherence for self-supervised one-shot video re-identification. In Proceedings of the European Conference on Computer Vision. Springer, 258274.Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. [41] Ristani Ergys, Solera Francesco, Zou Roger, Cucchiara Rita, and Tomasi Carlo. 2016. Performance measures and a data set for multi-target, multi-camera tracking. In Proceedings of the European Conference on Computer Vision. Springer, 1735.Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Ruan Weijian, Liang Chao, Yu Yi, Wang Zheng, Liu Wu, Chen Jun, and Ma Jiayi. 2020. Correlation discrepancy insight network for video re-identification. ACM Transactions on Multimedia Computing, Communications, and Applications 16, 4 (2020), 121.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Schroff Florian, Kalenichenko Dmitry, and Philbin James. 2015. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 815823.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Shen Chen, Jin Zhongming, Chu Wenqing, Jiang Rongxin, Chen Yaowu, Qi Guo-Jun, and Hua Xian-Sheng. 2019. Multi-level similarity perception network for person re-identification. ACM Transactions on Multimedia Computing, Communications, and Applications 15, 2 (2019), 119.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. [45] Song Guanglu, Leng Biao, Liu Yu, Hetang Congrui, and Cai Shaofan. 2018. Region-based quality estimation network for large-scale person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Subramaniam Arulkumar, Nambiar Athira, and Mittal Anurag. 2019. Co-segmentation inspired attention networks for video-based person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 562572.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Sun Yifan, Zheng Liang, Yang Yi, Tian Qi, and Wang Shengjin. 2018. Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline). In Proceedings of the European Conference on Computer Vision. 480496.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. [48] Sutskever Ilya, Martens James, Dahl George, and Hinton Geoffrey. 2013. On the importance of initialization and momentum in deep learning. In Proceedings of the International Conference on Machine Learning. PMLR, 11391147.Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. [49] Tang Zengming and Huang Jun. 2022. Harmonious multi-branch network for person re-identification with harder triplet loss. ACM Transactions on Multimedia Computing, Communications, and Applications 18, 4 (2022), 121.Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. [50] Wang Haoran, Jiao Licheng, Liu Fang, Li Lingling, Liu Xu, Ji Deyi, and Gan Weihao. 2021. IPGN: Interactiveness proposal graph network for human-object interaction detection. IEEE Transactions on Image Processing 30 (2021), 65836593. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  51. [51] Wang Haoran, Jiao Licheng, Yang Shuyuan, Li Lingling, and Wang Zexin. 2020. Simple and effective: Spatial rescaling for person reidentification. IEEE Transactions on Neural Networks and Learning Systems 33, 1 (2020), 145156. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  52. [52] Wang Hanzheng, Zhao Jiaqi, Zhou Yong, Yao Rui, Chen Ying, and Chen Silin. 2021. AMC-net: Attentive modality-consistent network for visible-infrared person re-identification. Neurocomputing 463 (2021), 226236. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  53. [53] Wang Kan, Ding Changxing, Maybank Stephen J., and Tao Dacheng. 2020. CDPM: Convolutional deformable part models for semantically aligned person re-identification. IEEE Transactions on Image Processing 29 (2020), 34163428. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. [54] Wang Kan, Hu Shuping, Cheng Jun, Pang Jianxin, and Tan Huan. 2022. RA loss: Relation-aware loss for robust person re-identification. In Proceedings of the Asian Conference on Computer Vision. 177194.Google ScholarGoogle Scholar
  55. [55] Wang Kan, Wang Pengfei, Ding Changxing, and Tao Dacheng. 2021. Batch coherence-driven network for part-aware person re-identification. IEEE Transactions on Image Processing 30 (2021), 34053418. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. [56] Wang Pengfei, Ding Changxing, Shao Zhiyin, Hong Zhibin, Zhang Shengli, and Tao Dacheng. 2022. Quality-aware part models for occluded person re-identification. IEEE Transactions on Multimedia (2022). DOI:Google ScholarGoogle ScholarCross RefCross Ref
  57. [57] Wang Taiqing, Gong Shaogang, Zhu Xiatian, and Wang Shengjin. 2014. Person re-identification by video ranking. In Proceedings of the European Conference on Computer Vision. Springer, 688703.Google ScholarGoogle ScholarCross RefCross Ref
  58. [58] Wang Xiaolong, Girshick Ross, Gupta Abhinav, and He Kaiming. 2017. Non-local neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 77947803.Google ScholarGoogle Scholar
  59. [59] Wu Yu, Lin Yutian, Dong Xuanyi, Yan Yan, Ouyang Wanli, and Yang Yi. 2018. Exploit the unknown gradually: One-shot video-based person re-identification by stepwise learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 51775186.Google ScholarGoogle ScholarCross RefCross Ref
  60. [60] Xu Shuangjie, Cheng Yu, Gu Kang, Yang Yang, Chang Shiyu, and Zhou Pan. 2017. Jointly attentive spatial-temporal pooling networks for video-based person re-identification. In Proceedings of the IEEE International Conference on Computer Vision. 47334742.Google ScholarGoogle ScholarCross RefCross Ref
  61. [61] Xu Sheng, Liu Chang, Zhang Baochang, Lü Jinhu, Guo Guodong, and Doermann David. 2022. BiRe-ID: Binary neural network for efficient person re-ID. ACM Transactions on Multimedia Computing, Communications, and Applications 18, 1s (2022), 122.Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. [62] Yan Yichao, Qin Jie, Chen Jiaxin, Liu Li, Zhu Fan, Tai Ying, and Shao Ling. 2020. Learning multi-granular hypergraphs for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 28992908.Google ScholarGoogle ScholarCross RefCross Ref
  63. [63] Yang Jinrui, Zheng Wei-Shi, Yang Qize, Chen Yingcong, and Tian Qi. 2020. Spatial-temporal graph convolutional network for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 32893299.Google ScholarGoogle ScholarCross RefCross Ref
  64. [64] Yang Xun, Wang Meng, Hong Richang, Tian Qi, and Rui Yong. 2017. Enhancing person re-identification in a self-trained subspace. ACM Transactions on Multimedia Computing, Communications, and Applications 13, 3 (2017), 123.Google ScholarGoogle ScholarDigital LibraryDigital Library
  65. [65] Zhang Guowen, Zhang Pingping, Qi Jinqing, and Lu Huchuan. 2021. Hat: Hierarchical aggregation transformers for person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia. 516525.Google ScholarGoogle ScholarDigital LibraryDigital Library
  66. [66] Zhang Wenyu, Ding Qing, Hu Jian, Ma Yi, and Lu Mingzhe. 2021. Pixel-wise graph attention networks for person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia. 52315238.Google ScholarGoogle ScholarDigital LibraryDigital Library
  67. [67] Zhang Zhizheng, Lan Cuiling, Zeng Wenjun, and Chen Zhibo. 2020. Multi-granularity reference-aided attentive feature aggregation for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1040710416.Google ScholarGoogle ScholarCross RefCross Ref
  68. [68] Zhao Yiru, Shen Xu, Jin Zhongming, Lu Hongtao, and Hua Xian-sheng. 2019. Attribute-driven feature disentangling and temporal aggregation for video person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 49134922.Google ScholarGoogle ScholarCross RefCross Ref
  69. [69] Zheng Liang, Bie Zhi, Sun Yifan, Wang Jingdong, Su Chi, Wang Shengjin, and Tian Qi. 2016. Mars: A video benchmark for large-scale person re-identification. In Proceedings of the European Conference on Computer Vision. Springer, 868884.Google ScholarGoogle ScholarCross RefCross Ref
  70. [70] Zheng Liang, Shen Liyue, Tian Lu, Wang Shengjin, Wang Jingdong, and Tian Qi. 2015. Scalable person re-identification: A benchmark. In Proceedings of the IEEE International Conference on Computer Vision. 11161124.Google ScholarGoogle ScholarDigital LibraryDigital Library
  71. [71] Zheng Meng, Karanam Srikrishna, Wu Ziyan, and Radke Richard J.. 2019. Re-identification with consistent attentive siamese networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 57355744.Google ScholarGoogle ScholarDigital LibraryDigital Library
  72. [72] Zheng Zhedong, Zheng Liang, and Yang Yi. 2017. A discriminatively learned cnn embedding for person reidentification. ACM Transactions on Multimedia Computing, Communications, and Applications 14, 1 (2017), 120.Google ScholarGoogle ScholarDigital LibraryDigital Library
  73. [73] Zhong Zhun, Zheng Liang, Kang Guoliang, Li Shaozi, and Yang Yi. 2020. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence. 1300113008.Google ScholarGoogle ScholarCross RefCross Ref
  74. [74] Zhou Zhen, Huang Yan, Wang Wei, Wang Liang, and Tan Tieniu. 2017. See the forest for the trees: Joint spatial and temporal recurrent neural networks for video-based person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 47474756.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Context Sensing Attention Network for Video-based Person Re-identification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Multimedia Computing, Communications, and Applications
      ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 19, Issue 4
      July 2023
      263 pages
      ISSN:1551-6857
      EISSN:1551-6865
      DOI:10.1145/3582888
      • Editor:
      • Abdulmotaleb El Saddik
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 27 February 2023
      • Online AM: 1 December 2022
      • Accepted: 25 November 2022
      • Revised: 25 October 2022
      • Received: 14 June 2022
      Published in tomm Volume 19, Issue 4

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!