Artificial Intelligence 20 min read

Causal Inference for Bias Mitigation in Kuaishou Recommendation Systems

This talk explains how recommendation bias arises from popularity and position effects, introduces causal inference concepts and three inference levels, reviews recent research such as DICE and Huawei’s causal embedding, and details Kuaishou’s practical applications—including popularity debias, causal representation decoupling, and video completion‑rate debias—along with experimental results and future challenges.

DataFunTalk
DataFunTalk
DataFunTalk
Causal Inference for Bias Mitigation in Kuaishou Recommendation Systems

Recommendation systems inevitably suffer from various biases, such as users preferring items at the top of the list or popular items, which leads to selection bias and feedback loops that degrade user experience and ecosystem health.

User interaction logs serve as training data, and directly using these biased logs introduces bias into the learned models, propagating biased recommendations.

To address this, causal inference is applied to identify and correct causal relationships between variables, distinguishing correlation from causation. Causal inference is described in three levels: (1) exploring correlations, (2) estimating the effect of interventions (e.g., uplift models), and (3) counterfactual reasoning to determine what would happen under alternative actions.

Typical causal methods include inverse propensity weighting (IPW), matching, and causal forest. Recent research highlighted includes:

DICE (WWW'21) – a causal embedding approach that separates user interest from conformity using triples.

Huawei’s Recsys'21 paper – splits feedback into biased and unbiased components and uses an information‑bottleneck loss.

KDD'21 work – removes popularity bias by treating it as a direct effect and subtracting the direct effect from the total effect.

SIGIR'21 work – applies a back‑door adjustment to eliminate popularity influence on item exposure.

In Kuaishou, three concrete applications were deployed:

Popularity debias – a back‑door operator removes the edge from popularity (Z) to item relevance (I) during training, while retaining the beneficial edge from popularity to content quality (C) during inference. The model predicts P(C|U,I,Z) as a product of a matching score and a popularity factor, with a controllable γ parameter and pair loss.

Causal representation decoupling – user behavior is split into interest and conformity embeddings. Positive/negative samples are constructed using reward (watch time + interactions) for interest and like count for conformity. Multi‑task training jointly optimizes feedback loss (main task) and the two embedding losses.

Video completion‑rate debias – a length‑aware threshold derived from the bimodal distribution of completion rates defines positive samples. IPW weights based on video length are applied to the loss to balance short and long videos.

Experiments show increased exposure and play counts for medium‑tail items, reduced exposure of low‑quality click‑bait, and higher completion rates across video lengths. However, challenges remain: determining the appropriate debiasing strength, building a unified debiasing system, and addressing information‑filter bubbles by shifting from purely satisficing recommendations to guiding user exploration.

The session concluded with a Q&A covering validation of interest vs. conformity embeddings, potential negative impacts of over‑debiasing, and strategies for long‑tail item correction.

machine learningrecommendation systemscausal inferenceBias MitigationKuaishou
DataFunTalk
Written by

DataFunTalk

Dedicated to sharing and discussing big data and AI technology applications, aiming to empower a million data scientists. Regularly hosts live tech talks and curates articles on big data, recommendation/search algorithms, advertising algorithms, NLP, intelligent risk control, autonomous driving, and machine learning/deep learning.

0 followers
Reader feedback

How this landed with the community

login Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.