This article details the architectural evolution of Pinterest's shopping conversion candidate generation model. It focuses on addressing challenges like data sparsity and noise in offsite conversion events through innovative training data design, feature engineering, and a two-tower model architecture utilizing parallel DCN v2 and a unified multi-task learning approach to optimize for both engagement and conversion.
Read original on Pinterest EngineeringPinterest's engineering team tackled the complex problem of optimizing shopping ads for offsite conversions, which are sparse, noisy, and delayed compared to onsite engagement signals. Their journey involved moving from an engagement-focused system to a dedicated conversion candidate generation model, highlighting key architectural and machine learning design decisions to improve advertiser value and user experience.
To counteract the sparsity and noise of offsite conversion data, several strategic design choices were made:
Key Learning: Handling Data Sparsity and Noise
Balancing high-value, sparse signals (conversions) with abundant, noisier signals (engagement) is a common challenge in real-world ML systems. Techniques like multi-task learning with weighted losses or auxiliary tasks are crucial for stable training and avoiding signal dilution.
The core of the system is a two-tower retrieval model, encoding user and Pin features separately using DCN v2 for cross-feature interactions. Two significant architectural evolutions improved performance:
The architectural and modeling advancements led to significant improvements, including a 2.3% increase in shopping conversion volume, a 2.7% lift in impression to conversion rate, and a 3.1% improvement in Return on Ad Spend (RoAS). This demonstrates the power of iterative architectural refinement and sophisticated machine learning techniques to solve complex business problems at scale.