Non-Stationary Transformer Architecture: A Versatile Framework for Recommendation Systems
-
Published:2024-05-27
Issue:11
Volume:13
Page:2075
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Liu Yuchen12ORCID, Li Gangmin3ORCID, Payne Terry R.2ORCID, Yue Yong12ORCID, Man Ka Lok12ORCID
Affiliation:
1. Department of Computing, Xi’an Jiaotong-Liverpool University, Suzhou 215123, China 2. Department of Computer Science, University of Liverpool, Liverpool L69 3BX, UK 3. HeXie Management Research Centre, College of Industry-Entrepreneurs (CIE), Xi’an Jiaotong-Liverpool University, Suzhou 215123, China
Abstract
Recommendation systems are crucial in navigating the vast digital market. However, user data’s dynamic and non-stationary nature often hinders their efficacy. Traditional models struggle to adapt to the evolving preferences and behaviours inherent in user interaction data, posing a significant challenge for accurate prediction and personalisation. Addressing this, we propose a novel theoretical framework, the non-stationary transformer, designed to effectively capture and leverage the temporal dynamics within data. This approach enhances the traditional transformer architecture by introducing mechanisms accounting for non-stationary elements, offering a robust and adaptable solution for multi-tasking recommendation systems. Our experimental analysis, encompassing deep learning (DL) and reinforcement learning (RL) paradigms, demonstrates the framework’s superiority over benchmark models. The empirical results confirm our proposed framework’s efficacy, which provides significant performance enhancements, approximately 8% in LogLoss reduction and up to 2% increase in F1 score with other attention-related models. It also underscores its potential applicability across accumulative reward scenarios with pure reinforcement learning models. These findings advocate adopting non-stationary transformer models to tackle the complexities of today’s recommendation tasks.
Funder
Suzhou Municipal Key Laboratory for Intelligent Virtual Engineering XJTLU Key Program Special Fund
Reference47 articles.
1. (2024, January 21). E-Commerce Sales by Country 2023. Available online: https://www.linkedin.com/pulse/ecommerce-sales-country-2023-julio-diaz-lg0be?trk=article-ssr-frontend-pulse_more-articles_related-content-card. 2. Liu, Y., Man, K., Li, G., Payne, T.R., and Yue, Y. (2022, January 19–21). Dynamic Pricing Strategies on the Internet. Proceedings of the International Conference on Digital Contents: AICo (AI, IoT, and Contents) Technology, Dehradun, India. 3. (2023, March 01). Global Short Video Platforms Market. Available online: https://www.grandviewresearch.com/press-release/global-short-video-platforms-market. 4. Guo, H., Tang, R., Ye, Y., Li, Z., and He, X. (2017, January 19–25). DeepFM: A factorization-machine based neural network for CTR prediction. Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, VIC, Australia. 5. Cheng, H.-T., Koc, L., Harmsen, J., Shaked, T., Chandra, T., Aradhye, H., Anderson, G., Corrado, G., Chai, W., and Ispir, M. (2016, January 15). Wide & deep learning for recommender systems. Proceedings of the 1st Workshop on Deep Learning for Recommender Systems, Boston, MA, USA.
|
|