A collection of sample research projects for inspiration and motivation.
How do you accurately predict if the user will click on an ad in less than a millisecond? Thankfully, you have billions of data points to help you.
A standard SVD works well. But what happens when you have to choose the top products amongst hundreds of thousands for every user, 2 billion times per day, in less than 50ms?
In a second-price auction, the theoretical optimal is to bid the expected value. But what happens when you run 15 billion auctions per day against the same competitors?
It’s easy, UCB and Thomson sampling have low regret. But what happens when new products come and go and when each ad displayed changes the reward of each arm?
Game theory/Reinforcement learning
How to find the optimal bidding strategy across multiple auctions? Can this be cast as a reinforcement learning problem in very high dimensions with unpredictable rewards.
You can always compute the classification error on model predicting the probability of a click. But is this really related to the online performance of a new model? What is the right offline metric that predicts online performance?
Stochastic gradient descent is great when you have lots of data. But what do you do when all data are not equal and you must distribute the learning over several hundred nodes?
- Criteo Labs Jobs [Link]