TY - EJOU AU - Luo, Hongxin AU - Zhou, Xiaobing AU - Ding, Haiyan AU - Wang, Liqing TI - Multi-Task Deep Learning with Task Attention for Post-Click Conversion Rate Prediction T2 - Intelligent Automation \& Soft Computing PY - 2023 VL - 36 IS - 3 SN - 2326-005X AB - Online advertising has gained much attention on various platforms as a hugely lucrative market. In promoting content and advertisements in real life, the acquisition of user target actions is usually a multi-step process, such as impression→click→conversion, which means the process from the delivery of the recommended item to the user’s click to the final conversion. Due to data sparsity or sample selection bias, it is difficult for the trained model to achieve the business goal of the target campaign. Multi-task learning, a classical solution to this problem, aims to generalize better on the original task given several related tasks by exploiting the knowledge between tasks to share the same feature and label space. Adaptively learned task relations bring better performance to make full use of the correlation between tasks. We train a general model capable of capturing the relationships between various tasks on all existing active tasks from a meta-learning perspective. In addition, this paper proposes a Multi-task Attention Network (MAN) to identify commonalities and differences between tasks in the feature space. The model performance is improved by explicitly learning the stacking of task relationships in the label space. To illustrate the effectiveness of our method, experiments are conducted on Alibaba Click and Conversion Prediction (Ali-CCP) dataset. Experimental results show that the method outperforms the state-of-the-art multi-task learning methods. KW - Multi-task learning; recommend system; attention; meta-learning DO - 10.32604/iasc.2023.036622