Personalized Title Generation and Automatic Cover Image Synthesis for Information‑Flow Scenarios
This article presents technical approaches for generating personalized article titles and automatically synthesizing cover images, covering keyword‑based, click‑sequence‑based, and author‑style‑based title models, as well as image restoration, key‑information extraction, object detection, and layout generation techniques to improve user engagement in recommendation and search systems.
In information‑flow scenarios such as QQ Browser, titles and cover images are critical cues that influence whether users click to view detailed content, making them essential for creators, platforms, and advertisers.
Personalized Title Generation aims to produce titles tailored to individual users or specific contexts. Three main use cases are described: recommendation systems, search engines, and creator platforms. Challenges include representing user interests, queries, or author style, and designing interaction mechanisms between scene representations and the article.
Three solution families are introduced:
Keyword‑based title generation: integrates keywords (e.g., tags, interests, queries) into a Transformer model by adding a dedicated keyword representation layer or by using the keywords as queries in a multi‑head attention module, resulting in titles that better match user interests.
Click‑sequence‑based title generation: combines a Transformer encoder for article content with an LSTM decoder guided by a user embedding derived from historical click sequences. The user embedding can initialize the LSTM hidden state, participate in attention distribution, or be incorporated into the gating network, producing titles aligned with the user’s past behavior.
Author‑style‑based title generation: incorporates an author’s historical titles into a Transformer model using contrastive learning to capture stylistic features, leading to titles that preserve the author’s brand while improving Rouge and BLEU scores.
Automatic Cover Image Synthesis addresses the need for visually appealing yet informative cover images. The pipeline includes:
Image restoration to remove watermarks, subtitles, or logos using techniques such as Faster R‑CNN and inpainting.
Seq2Seq (Pointer‑augmented T5) extraction of key information from titles and tags.
Object detection (e.g., Faster R‑CNN) to locate faces, objects, etc., ensuring they are not occluded.
Layout Generation to automatically place extracted textual information onto the cleaned image, producing a balanced and readable cover.
Experimental results show that both personalized title models and the cover‑image synthesis pipeline significantly improve click‑through rates and user satisfaction.
The presentation concludes with acknowledgments and a thank‑you to the audience.
DataFunTalk
Dedicated to sharing and discussing big data and AI technology applications, aiming to empower a million data scientists. Regularly hosts live tech talks and curates articles on big data, recommendation/search algorithms, advertising algorithms, NLP, intelligent risk control, autonomous driving, and machine learning/deep learning.
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.