I came across this interesting set of blog posts by Sergei Feldman on the use of bandit approaches in online recommendation.
In particular, the one I really enjoyed reading was the comparison of the approaches needed to solve the multi armed bandit problem. Need to play around with his code someday
References: