Kaggle Team

Recruit Ponpare is Japan’s leading joint coupon site, offering huge discounts on everything from hot yoga, to gourmet sushi, to a summer concert bonanza. The Recruit Coupon Purchase Prediction challenge asked the community to predict which coupons a customer would buy in a given period of time using past purchase and browsing behavior.

Image for post
Image for post

Halla Yang finished 2nd ahead of 1,191 other data scientists. His experience working with time series data helped him use unsupervised methods effectively in conjunction with gradient boosting. In this blog, Halla walks through his approach and shares key visualizations that helped him better understand and work with the dataset.

I’ve worked almost a decade in finance as a quantitative researcher and portfolio manager. I’ve also competed in several Kaggle contests, placing first in the Pfizer Volume Prediction Masters competition, sixth in the Merck Molecular Activity Challenge, and ninth in the Diabetic Retinopathy Detection.

Predicting prices for thousands of stocks and predicting purchases by thousands of Japanese internet users are loosely similar problems. You can forecast stock returns by looking at time series data such as past returns and cross-sectional data such as industry averages. You can forecast coupon purchases by looking at time series features based on past purchases and cross-sectional features based on peer group averages.

For each (user, coupon) pair, I calculated the probability that the user would purchase that coupon during the test period using a gradient boosting classifier. I sorted the coupons for each user by probability, composing the ten highest probability coupons into my submission.

To train my classifier, I constructed training data for 24 “train periods” that simulated the test period. Train period 1 is the week from 2012–01–08 through 2012–01–14, and includes all coupons with a DISPFROM date — the date on which they’re supposed to be first displayed — in that week. Train period 2 is the week from 2012–01–15 through 2012–01–21, and includes all coupons with a DISPFROM date in that week. Train period 24 is the week from 2012–06–17 through 2012–06–23, and includes all coupons with a DISPFROM date in that week.

For each of these training periods, I built a set of features for each (user, relevant coupon) pair. This set of features includes user-specific data, e.g. gender, days on site, and age; coupon-specific data, e.g. catalog price, genre, and price rate; as well as user-coupon interaction data, e.g. how often has the user viewed coupons of the same genre. The target for each observation is set to 1 if the user purchased that coupon during the training week, and 0 otherwise.

To calibrate the parameters of my model, I first trained a model on the first twenty-three weeks of data, and estimated my log loss and confusion matrix on the twenty-fourth week. I then trained a model on the full twenty-four weeks of data to generate my competition submission.

The only supervised learning method I used was gradient boosting, as implemented in the excellent xgboost package. I cycled through other algorithms at the start of my analysis to get a feel for their relative performance — logistic regressions, random forests, SVMs, as well as deep neural networks — but found that gradient boosting was the single best classifier for my approach.

First, many test set and training set coupons were viewed prior to their DISPFROM, the date on which they’re supposed to be first displayed, and so one could use direct views as a forecasting variable. The violin plot below shows the distribution of first view times relative to DISPFROM. A negative x-value indicates the coupon was viewed prior to its DISPFROM. Over a quarter of coupons are first viewed more than twelve hours before their DISPFROM, and five percent of coupons are first viewed more than ninety hours before their DISPFROM.

Image for post
Image for post


Simply counting the number of times a user has viewed a test set coupon is tremendously helpful in forecasting test set purchases. As shown in the left panel of the figure below, users are 2.5% likely to buy a coupon if they’ve viewed it exactly once prior to its DISPFROM, but that probability rises to 32% if they’ve viewed the coupon four or more times.

Image for post
Image for post

Second, users tend to buy the same coupons over and over. As shown in the middle panel of the above figure, a user who has purchased a coupon with a given prefecture, genre, and catalog price four or more times has a 38% chance of buying a matched coupon again in the next week if it is offered for sale.

Third, peer group averages can help forecast the behavior of users with little or no history. The right panel of the above figure shows that a user’s probability of buying a coupon increases from less than 0.1% to above 0.6% if more than ten percent of age, sex, and geography-matched peers have bought a coupon with the same characteristics.

Fourth, it’s important to consider the geographic coverage of each coupon. To be specific, a coupon is relevant for the multiple prefectures listed in coupon_area_train.csv, not just the single prefecture listed for that coupon in coupon_list_train.csv. In the kernel density plots below, I show the purchase intensity for users based in four prefectures: Tokyo, Kanagawa, Osaka, and Aichi, using the geographic data in coupon_list_train.csv. The purchases for Osaka and Aichi users appear strongly bimodal, with an unusually large number of purchases occurring in the Tokyo region.

Image for post
Image for post

On the other hand, if we look at all the prefectures that map to a given coupon, we find that Osaka users purchased Tokyo coupons not because they planned to travel to Tokyo, but because these coupons were also local to Osaka. If we plot the geographic intensity of “nearest-to-user” prefecture rather than a coupon’s primary listing prefecture, we see much more localized purchase behavior.

Focus on understanding the problem. Without understanding the problem, it’s impossible to develop a solution.

Start with simple approaches and models. A fast development cycle is key to testing out ideas and learning what works. Don’t start building computationally expensive ensembles until you have iterated through most of your best ideas.

Halla Yang has worked as a quantitative researcher, portfolio manager and trader at Goldman Sachs Asset Management, Jump Trading and Arrowstreet Capital. He holds a Ph.D. in Business Economics from Harvard, and a B.A. in Physics, summa cum laude, also from Harvard. He is about to start a new position as data scientist at a management consulting firm.



Source link

Leave a Reply

Your email address will not be published.