
In recent years, unsupervised person reidentification technology has made great strides. The technology retrieves images of interested persons under different cameras from massive repositories of unlabeled images. However, in the current research, there are some existing problems, such as the influence of pedestrians appearing across cameras and pseudo-label noise. To solve these problems, we conduct research in two ways: removing the camera bias and dynamically updating the memory model. In removing the camera bias, based on a learnable channel attention module, the features that are only related to cameras can be extracted from the feature map, thereby removing the camera bias in the global features and obtaining the features that can represent the pedestrians. In regards to dynamically updating the memory model, since the instance features do not necessarily belong to the identity represented by the pseudo-label, we adopt a method to update the memory dynamically according to the distance between the instance features and the category features so that the category features tend to be true. We combine the removal of the camera bias and the dynamic updating of the memory model to better solve problems in this field. Extensive experimentation demonstrates the superiority of our method over the state-of-the-art approaches on fully unsupervised Re-ID tasks.
This paper improves unsupervised person reidentification by removing the camera bias and dynamically updating the memory model.
Figure 1. Illustration of our pipeline. We first apply the channel attention module after the backbone network to explicitly separate the camera-related information from the feature maps. Then, we utilize a new mechanism of dynamically updating the memory dictionary according to the distance between the instance feature and the cluster feature.
[1] |
Huang Y, Peng P, Jin Y, et al. Domain adaptive attention learning for unsupervised person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 2020, 34: 11069–11076. DOI: 10.1609/aaai.v34i07.6762
|
[2] |
Ge Y, Zhu F, Chen D, et al. Self-paced contrastive learning with hybrid memory for domain adaptive object re-ID. In: NIPS’20: 34th International Conference on Neural Information Processing Systems. BC, Canada: Curran Associates Inc, 2020: 11309–11321.
|
[3] |
Chen H, Lagadec B, Brémond F. ICE: Inter-instance contrastive encoding for unsupervised person re-identification. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV). Montreal, Canada: IEEE, 2021: 14940–14949.
|
[4] |
Wang M, Lai B, Huang J, et al. Camera-aware proxies for unsupervised person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 2021, 35 (4): 2764–2772. DOI: 10.1609/aaai.v35i4.16381
|
[5] |
Ge Y, Chen D, Li H. Mutual mean-teaching: Pseudo label refinery for unsupervised domain adaptation on person re-identification. 2020. https://doi.org/10.48550/arXiv.2001.01526. Accessed December 3, 2021
|
[6] |
Zheng K, Lan C, Zeng W, et al. Exploiting sample uncertainty for domain adaptive person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 2021, 35 (4): 3538–3546. DOI: 10.1609/aaai.v35i4.16468
|
[7] |
Lin Y, Xie L, Wu Y, et al. Unsupervised person re-identification via softened similarity learning. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, USA: IEEE, 2020: 3387–3396.
|
[8] |
Tian J, Tang Q, Li R, et al. A camera identity-guided distribution consistency method for unsupervised multi-target domain person re-identification. ACM Transactions on Intelligent Systems and Technology, 2021, 12 (4): 1–18. DOI: 10.1145/3454130
|
[9] |
Xuan S, Zhang S. Intra-inter camera similarity for unsupervised person re-identification. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Nashville, USA: IEEE, 2021: 11921–11930.
|
[10] |
Zheng L, Shen L, Tian L, et al. Scalable person re-identification: A benchmark. In: 2015 IEEE International Conference on Computer Vision (ICCV). Santiago, Chile: IEEE, 2015: 1116–1124.
|
[11] |
Ristani E, Solera F, Zou R, et al. Performance measures and a data set for multi-target, multi-camera tracking. In: Hua G, Jégou H, editors. Computer Vision–ECCV 2016 Workshops. Cham: Springer International Publishing, 2016: 17–35.
|
[12] |
Song L, Wang C, Zhang L, et al. Unsupervised domain adaptive re-identification: Theory and practice. Pattern Recognition, 2020, 102: 107173. DOI: 10.1016/j.patcog.2019.107173
|
[13] |
Fu Y, Wei Y, Wang G, et al. Self-similarity grouping: A simple unsupervised cross domain adaptation approach for person re-identification. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV). Seoul, South Korea: IEEE, 2019: 6111–6120.
|
[14] |
Lin Y, Dong X, Zheng L, et al. A bottom-up clustering approach to unsupervised person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 2019, 33 (1): 8738–8745. DOI: 10.1609/aaai.v33i01.33018738
|
[15] |
Zeng K, Ning M, Wang Y, et al. Hierarchical clustering with hard-batch triplet loss for person re-identification. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, USA: IEEE, 2020: 13654–13662.
|
[16] |
Zhang X, Cao J, Shen C, et al. Self-training with progressive augmentation for unsupervised cross-domain person re-identification. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV). Seoul, South Korea: IEEE, 2019: 8221–8230.
|
[17] |
Yu H X, Zheng W S, Wu A, et al. Unsupervised person re-identification by soft multilabel learning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach, USA: IEEE, 2019: 2143–2152.
|
[18] |
Wang D, Zhang S. Unsupervised person re-identification via multi-label classification. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, USA: IEEE, 2020: 10978–10987.
|
[19] |
Wei L, Zhang S, Gao W, et al. Person transfer GAN to bridge domain gap for person re-identification. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE, 2018: 79–88.
|
[20] |
Zhao L, Li X, Zhuang Y, et al. Deeply-learned part-aligned representations for person re-identification. In: 2017 IEEE International Conference on Computer Vision (ICCV). Venice, Italy: IEEE, 2017: 3239–3248.
|
[21] |
Zou Y, Yang X, Yu Z, et al. Joint disentangling and adaptation for cross-domain person re-identification. In: Vedaldi A, Bischof H, Brox T, editors. Computer Vision-ECCV 2020. Cham: Springer International Publishing, 2020: 87–104.
|
[22] |
Choi Y, Choi M, Kim M, et al. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE, 2018: 8789–8797.
|
[23] |
Yang F, Zhong Z, Luo Z, et al. Joint noise-tolerant learning and meta camera shift adaptation for unsupervised person re-identification. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Nashville, USA: IEEE, 2021: 4853–4862.
|
[24] |
van den Oord A, Li Y, Vinyals O. Representation learning with contrastive predictive coding. 2018. https://doi.org/10.48550/arXiv.1807.03748. Accessed December 12, 2021.
|
[25] |
Zheng Y, Tang S, Teng G, et al. Online pseudo label generation by hierarchical cluster dynamics for adaptive person re-identification. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV). Montreal, Canada: IEEE, 2021: 8351–8361.
|
[26] |
Wu Y, Huang T, Yao H, et al. Multi-centroid representation network for domain adaptive person re-ID. Proceedings of the AAAI Conference on Artificial Intelligence, 2022, 36 (5): 2750–2758. DOI: 10.1609/aaai.v36i3.20178
|
[27] |
Wu Y, Wu X, Li X, et al. MGH: metadata guided hypergraph modeling for unsupervised person re-identification. In: MM '21: Proceedings of the 29th ACM International Conference on Multimedia. New York: Association for Computing Machinery, 2021: 1571–1580.
|
Algorithm 2.1 CEEMDAN_Multi_Att_RL algorithm. |
Require: Technical indicators, including the Open, High, Low, Close, Volume, Amount, Open_MA_2, Open_MA_5, and Open_MA_10 (All technical indicators are at the minute level); |
Ensure: Sequence value processed by the CEEMDAN method; |
1: Perform CEEMDAN so that \overline{{\rm{IMF}}_{1} },\cdots,\overline{{\rm{IMF}}_{n} } and R_{n}(t) of each factor sequence are obtained; |
2: For the j th factor of the i th stock, add the values of the corresponding positions of arrays \overline{{\rm{IMF}}_{2} },\cdots,\overline{{\rm{IMF}}_{n-1} } to obtain a new sequence value D_{i,j} ; |
Require: Sequence value D of technical indicators processed by the CEEMDAN method, Account information (balance and position information of each stock), Q network architecture and \epsilon ; |
3: Initialize all parameters \theta of the Q network randomly; |
4: Initialize the action-value function Q corresponding to all states and actions based on \theta ; |
5: Initialize replay memory {\cal{D}} ; |
6: for i = 1,\cdots , t do |
7: Initialize state S to obtain s_1 ; |
8: Initialize a random process \epsilon for action; |
9: Take s_t as the input of Q network to obtain the Q value outputs corresponding to all actions; |
10: Select a_t = {{\rm{argmax}}}_{a_t} Q(s_t,a_t,\theta) ; |
11: Execute the action a_t in the state s_t to obtain the new state s_{t+1} and reward r_t ; |
12: Decide whether to terminate the states (\text{is}\_\text{end} = \text{true}/\text{false}); |
13: Save ( s_t , a_t , r_t , s_{t+1} , is_end) to replay memory {\cal{D}} ; |
14: S = s_{t+1}; |
15: M samples ( s_k , a_k , r_k , s_{k+1} , is_end) are sampled from replay memory {\cal{D}} , and calculate the current target Q value y_k ; |
16: y_k=\left\{\begin{aligned} r_k, \qquad \qquad && {\rm{is} }\_{\rm{end} } &={\rm{true} }; \\ r_k +\gamma \max_{a'} Q(s',a'; \theta_{k-1}), && {\rm{is} }\_{\rm{end} } &={\rm{false} }; \end{aligned}\right. |
17: Use the mean squaresp loss function: |
L_k(\theta_k) = \mathbb{E}_{\pi}[(r_k+\gamma \max\limits_{a'} Q(s', a';\theta_{k-1})-Q(s,a;\theta_k))^2]; |
18: The gradient back propagation of the neural network is used to update all the parameters \theta of the Q network; |
19: If s_{t+1} is in the termination state, the current round of iteration is completed; otherwise, continue to iterate; |
20: end for |
Abbreviation | Description |
CI | Compound interest value of structure |
SR | Sharpe ratio of structure |
MD | max drawdown of structure |
Mean of dr | Mean of simple interest of structure |
Std of dr | Standard deviation of simple interest of structure |
Min of dr | minimum of simple interest of structure |
Qn of dr | Quartile n of simple interest of structure |
Med of dr | Median of simple interest of structure |
Max of dr | median of simple interest of structure |
2018 | benchmark | CMAR_10 | MAR_10 | CDRL_10 | DRL_10 |
CI | 0.8538 | 1.0907 | 1.0676 | 1.0370 | 1.0000 |
SR | −0.5790 | 0.5609 | 0.4503 | 0.2900 | 0.1046 |
MD | −0.2294 | −0.1336 | −0.0957 | −0.1050 | −0.1256 |
Mean of dr | −0.0005 | 0.0004 | 0.0004 | 0.0002 | 0.0001 |
Std of dr | 0.0142 | 0.0134 | 0.0131 | 0.0140 | 0.0130 |
Min of dr | −0.0531 | −0.0407 | −0.0412 | −0.0498 | −0.0708 |
Q1 of dr | −0.0089 | −0.0079 | −0.0077 | −0.0079 | −0.0075 |
Med of dr | 0.0002 | 0.0006 | 0.0007 | 0.0005 | −0.0001 |
Q3 of dr | 0.0078 | 0.0084 | 0.0078 | 0.0086 | 0.0081 |
Max of dr | 0.0463 | 0.0457 | 0.0482 | 0.0437 | 0.0349 |
2018 | benchmark | CMAR_1_m | MAR_1_m | CDRL_1_m | DRL_1_m |
CI | 0.8538 | 1.1298 | 1.0937 | 1.0555 | 0.9706 |
SR | −0.5790 | 1.6010 | 1.0398 | 0.7624 | −0.2508 |
MD | -0.2294 | −0.0360 | −0.0612 | −0.0561 | −0.1032 |
Mean of dr | −0.0005 | 0.0005 | 0.0004 | 0.0002 | −0.0001 |
Std of dr | 0.0142 | 0.0055 | 0.0063 | 0.0051 | 0.0064 |
Min of dr | −0.0531 | −0.0144 | −0.0240 | −0.0147 | −0.0250 |
Q1 of dr | −0.0089 | −0.0029 | −0.0032 | −0.0029 | −0.0033 |
Med of dr | 0.0002 | 0.0001 | 0.0001 | 0.0004 | −0.0002 |
Q3 of dr | 0.0078 | 0.0033 | 0.0042 | 0.0030 | 0.0036 |
Max of dr | 0.0463 | 0.0199 | 0.0200 | 0.0154 | 0.0201 |
2019 | benchmark | CMAR_10 | MAR_10 | CDRL_10 | DRL_10 |
CI | 1.3357 | 1.5413 | 1.4418 | 1.4116 | 1.3796 |
SR | 2.0358 | 3.1193 | 3.1507 | 2.6594 | 2.4129 |
MD | −0.1019 | −0.0827 | −0.0724 | −0.0702 | −0.1200 |
Mean of dr | 0.0013 | 0.0019 | 0.0016 | 0.0015 | 0.0014 |
Std of dr | 0.0115 | 0.0120 | 0.0096 | 0.0107 | 0.0109 |
Min of dr | −0.0346 | −0.0363 | −0.0282 | −0.0316 | −0.0298 |
Q1 of dr | −0.0051 | −0.0043 | −0.0033 | −0.0043 | −0.0053 |
Med of dr | 0.0013 | 0.0009 | 0.0011 | 0.0008 | 0.0020 |
Q3 of dr | 0.0076 | 0.0081 | 0.0072 | 0.0063 | 0.0073 |
Max of dr | 0.0450 | 0.0521 | 0.0358 | 0.0432 | 0.0392 |
2019 | benchmark | CMAR_1_m | MAR_1_m | CDRL_1_m | DRL_1_m |
CI | 1.3357 | 1.4118 | 1.3610 | 1.3411 | 1.3196 |
SR | 2.0358 | 4.2549 | 4.1825 | 3.2599 | 3.4278 |
MD | −0.1019 | −0.0295 | −0.0317 | −0.0421 | −0.0394 |
Mean of dr | 0.0013 | 0.0014 | 0.0013 | 0.0012 | 0.0012 |
Std of dr | 0.0115 | 0.0065 | 0.0058 | 0.0070 | 0.0062 |
Min of dr | −0.0346 | −0.0134 | −0.0132 | −0.0196 | −0.0151 |
Q1 of dr | −0.0051 | −0.0023 | −0.0019 | −0.0023 | −0.0027 |
Med of dr | 0.0013 | 0.0007 | 0.0009 | 0.0005 | 0.0005 |
Q3 of dr | 0.0076 | 0.0039 | 0.0041 | 0.0044 | 0.0042 |
Max of dr | 0.0450 | 0.0330 | 0.0282 | 0.0326 | 0.0231 |
2020 | benchmark | CMAR_10 | MAR_10 | CDRL_10 | DRL_10 |
CI | 1.1444 | 1.3836 | 1.2743 | 1.2397 | 1.1653 |
SR | 0.7393 | 1.7615 | 1.4447 | 1.2661 | 0.8979 |
MD | −0.1553 | −0.1300 | −0.1443 | −0.1286 | −0.1235 |
Mean of dr | 0.0007 | 0.0015 | 0.0011 | 0.0010 | 0.0007 |
Std of dr | 0.0161 | 0.0160 | 0.0139 | 0.0140 | 0.0143 |
Min of dr | −0.1138 | −0.0948 | −0.0930 | −0.0781 | −0.1025 |
Q1 of dr | −0.0065 | −0.0060 | −0.0057 | −0.0059 | −0.0058 |
Med of dr | −0.0005 | 0.0001 | −0.0001 | 0.0003 | −0.0005 |
Q3 of dr | 0.0083 | 0.0081 | 0.0074 | 0.0072 | 0.0070 |
Max of dr | 0.0662 | 0.0706 | 0.0511 | 0.0719 | 0.0744 |
2020 | benchmark | CMAR_1_m | MAR_1_m | CDRL_1_m | DRL_1_m |
CI | 1.1444 | 1.4176 | 1.2883 | 1.2313 | 1.1735 |
SR | 0.7393 | 3.6530 | 2.9738 | 2.1343 | 1.6955 |
MD | −0.1553 | −0.0413 | −0.0301 | −0.0461 | −0.0528 |
Mean of dr | 0.0007 | 0.0015 | 0.0011 | 0.0009 | 0.0007 |
Std of dr | 0.0161 | 0.0077 | 0.0065 | 0.0074 | 0.0070 |
Min of dr | −0.1138 | −0.0353 | −0.0169 | −0.0351 | −0.0394 |
Q1 of dr | −0.0065 | −0.0021 | −0.0022 | −0.0027 | −0.0030 |
Med of dr | −0.0005 | 0.0006 | 0.0002 | 0.0005 | −0.0001 |
Q3 of dr | 0.0083 | 0.0041 | 0.0036 | 0.0038 | 0.0033 |
Max of dr | 0.0662 | 0.0560 | 0.0375 | 0.0344 | 0.0384 |