[1] 袁柳, 李战怀, 陈世亮. 基于本体的Deep web数据标注[J]. 软件学报, 2008, 19(2):237-245. Yuan Liu, Li Zhanhuai, Chen Shiliang. Ontology-based annotation for deep web data[J]. Journal of Software, 2008, 19(2):237-245. [2] Deng Jia, Dong Wei, Socher R, et al. Imagenet:a large-scale hierarchical image database[C]//2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition(CVPR 2009). Florida:IEEE, 2009:20-25. [3] 周立君, 汪涛. 亚马逊土耳其机器人:科学研究的众包网络平台研究综述[J]. 科技进步与对策, 2014(8):162-166. Zhou Lijun, Wang Tao. Amazon mechanical turk:a survey on science research of crowdsourcing platform[J]. Science & Technology Progress and Policy, 2014(8):162-166. [4] 肖玲, 潘浩. 基于WiFi信号的人体动作识别系统[J]. 北京邮电大学学报, 2018, 41(3):119-124. Xiao Ling, Pan Hao. Human activity recognition system based on WiFi signal[J]. Journal of Beijing University of Posts and Telecommunications, 2018, 41(3):119-124. [5] 周全, 王磊, 周亮, 等. 基于多尺度上下文的图像标注算法[J]. 自动化学报, 2014(12):276-281. Zhou Quan, Wang Lie, Zhou Liang, et al. Multi-scale contextual image labeling[J]. ACTA Automatica Sinica, 2014(12):276-281. [6] Srihari R K, Zhang Zhongfei. Show&tell:a semi-automated image annotation system[J]. IEEE Multimedia, 2000, 7(3):61-71. [7] 税留成, 刘卫忠, 冯卓明. 基于生成式对抗网络的图像自动标注[J]. 计算机应用, 2019(7):2129-2133. Shui Liucheng, Liu Weizhong, Feng Zhuoming. Automatic image annotation based on generative adversarial network[J]. Journal of Computer Applications, 2019(7):2129-2133. [8] Lassner C, Romero J, Kiefel M, et al. Unite the people:closing the loop between 3D and 2D human representations[C]//2017 IEEE Conference on Computer Vision and Pattern Recognition(CVPR 2017). Hawaii:IEEE, 2017:6050-6059. [9] Kanazawa A, Black M J, Jacobs D W, et al. End-to-end recovery of human shape and pose[C]//2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR). Salt Lake City:IEEE, 2018:7122-7131. [10] Chen Yilun, Wang Zhicheng, Peng Yuxiang, et al. Cascaded pyramid network for multi-person pose estimation[C]//2018 IEEE Conference on Computer Vision and Pattern Recognition(CVPR 2018). Salt Lake City:IEEE, 2018:7103-7112. [11] He Kaiming, Gkioxari G, Dollar P, et al. Mask R-CNN[C]//2017 IEEE International Conference on Computer Vision(ICCV 2017). Vinice:IEEE, 2017:2961-2969. [12] Simonyan K, Zisserman A. Two-stream convolutional networks for action recognition in videos[J]. Advances in Neural Information Processing Systems, 2014(1):568-576. [13] Chéron G, Laptev I, Schmid C. P-CNN:pose-based CNN features for action recognition[C]//2015 IEEE International Conference on Computer Vision(ICCV 2015). Santiago:IEEE, 2015:3218-3226. [14] Brox T,Bruhn A,Papenberg N,et al.High accuracy optical flow estimation based on a theory for warping[C]//European Conference on Computer Vision.Heidelberg:Springer, 2004:25-36. [15] Qiu Zhaofan, Yao Ting, Mei Tao. Learning spatio-temporal representation with pseudo-3D residual networks[C]//2017 IEEE International Conference on Computer Vision(ICCV). Vinice:IEEE, 2017:5534-5542. |