Funder
National Natural Science Foundation of China
Reference27 articles.
1. P. Ji, M. Cao, Y. Zou, Visual Relation-Aware Unsupervised Video Captioning, in: International Conference on Artificial Neural Networks, 2022, pp. 495–507.
2. D. Nukrai, R. Mokady, A. Globerson, Text-Only Training for Image Captioning using Noise-Injected CLIP, in: Findings of the Association for Computational Linguistics: EMNLP 2022, 2022, pp. 4055–4063.
3. A. Radford, J.W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al., Learning transferable visual models from natural language supervision, in: International Conference on Machine Learning, 2021, pp. 8748–8763.
4. SNRCN2: Steganalysis noise residuals based CNN for source social network identification of digital images;Rana;Pattern Recognit. Lett.,2023
5. Sequence in sequence for video captioning;Wang;Pattern Recognit. Lett.,2020