A Simple Yet Powerful Baseline for Collaborative Filtering
We just learned: Matrix Factorization decomposes \(R \approx U \times V^T\)
But what if we could skip the factorization?
EASE (Steck, 2019): Learn item-item similarities directly
Classic autoencoder: Learn to reconstruct input through bottleneck
\[\text{Input} \rightarrow \text{Encoder} \rightarrow \text{Bottleneck} \rightarrow \text{Decoder} \rightarrow \text{Output}\]
EASE: Skip the bottleneck! Learn reconstruction weights directly
\[X \approx X \cdot B\]
where \(B\) is an item-item similarity matrix
Constraint: \(\text{diag}(B) = 0\) (can’t predict item from itself)
How does \(\hat{X} = X \cdot B\) make recommendations?
Users: Alice watched [Toy Story], Bob watched [Godfather, Die Hard]
Item-item similarity matrix B (learned from all users):
| Toy Story | Godfather | Die Hard | Inside Out | Heat | |
|---|---|---|---|---|---|
| Toy Story | 0.0 | 0.01 | 0.02 | 0.18 | 0.01 |
| Godfather | 0.01 | 0.0 | 0.08 | 0.02 | 0.15 |
| Die Hard | 0.02 | 0.08 | 0.0 | 0.01 | 0.12 |
Alice: Inside Out = 0.18 ✅ | Heat = 0.01 ❌
Bob: Heat = 0.15 + 0.12 = 0.27 ✅ | Inside Out = 0.02 + 0.01 = 0.03 ❌
Strengths ✅
Weaknesses ❌
Bottom line: Excellent baseline, but not the final answer!
When to use EASE:
Real-world impact:
Many production systems use EASE as: