<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>MAB on Sandeep Gangarapu</title>
    <link>https://sandeepgangarapu.com/tags/mab/</link>
    <description>Recent content in MAB on Sandeep Gangarapu</description>
    <image>
      <title>Sandeep Gangarapu</title>
      <url>https://sandeepgangarapu.com/images/profile/vivekkrishnanphotography-6_websize.jpg</url>
      <link>https://sandeepgangarapu.com/images/profile/vivekkrishnanphotography-6_websize.jpg</link>
    </image>
    <generator>Hugo</generator>
    <language>en-us</language>
    <lastBuildDate>Sun, 11 Oct 2020 23:13:00 -0600</lastBuildDate>
    <atom:link href="https://sandeepgangarapu.com/tags/mab/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Who and when should one use Multi-Armed Bandits? A scenario.</title>
      <link>https://sandeepgangarapu.com/blog/technical/who-and-when-should-one-use-multi-armed-bandits-a-scenario/</link>
      <pubDate>Sun, 11 Oct 2020 23:13:00 -0600</pubDate>
      <guid>https://sandeepgangarapu.com/blog/technical/who-and-when-should-one-use-multi-armed-bandits-a-scenario/</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;💡 This is a humor piece. I am using the scenario to communicate the use of MAB&amp;rsquo;s. Please take it with a pinch of salt.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;A manager at a tech company that sells Alphonso mangoes online is stuck with one of the most difficult decisions. What should the color of the &amp;ldquo;Buy Now&amp;rdquo; button be? She recently came to know that data-driven decision making is the new fad and wants to follow suit. She has a meeting with her team of data scientists and marketing folks.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Thompson Sampling Algorithm for Normal outcome distribution</title>
      <link>https://sandeepgangarapu.com/blog/technical/thompson-sampling-algorithm-for-normal-outcome-distribution/</link>
      <pubDate>Thu, 10 Sep 2020 21:46:00 -0600</pubDate>
      <guid>https://sandeepgangarapu.com/blog/technical/thompson-sampling-algorithm-for-normal-outcome-distribution/</guid>
      <description>&lt;p&gt;Thompson Sampling is one of the most popular Multi-Armed bandit (MAB) algorithms - the main reason being its explainability (imagine explaining upper confidence bound to your manager) and decent performance in practice [1].&lt;/p&gt;
&lt;p&gt;Many blog posts on the Internet show how to implement Thompson sampling (&lt;a href=&#34;https://visualstudiomagazine.com/articles/2019/06/01/thompson-sampling.aspx&#34;&gt;here&lt;/a&gt;, &lt;a href=&#34;https://towardsdatascience.com/hompson-sampling-for-multi-armed-bandit-problems-part-1-b750cbbdad34&#34;&gt;here&lt;/a&gt;, &lt;a href=&#34;https://peterroelants.github.io/posts/multi-armed-bandit-implementation/&#34;&gt;here&lt;/a&gt; and &lt;a href=&#34;https://medium.com/analytics-vidhya/multi-armed-bandit-analysis-of-thompson-sampling-algorithm-6375271f40d1&#34;&gt;here&lt;/a&gt;). Almost all of them consider Bernoulli outcome distribution (e.g. click or no click, purchase, or no purchase) and use the Beta-Bernoulli Bayesian update procedure for simulations and usually compare the performance (Regret) to other MAB algorithms like UCB or sometimes to A/B testing. However, none of them consider Gaussian outcome distribution and especially the case when both mean and variance of the distributions are unknown (which is usually the case when you are conducting an experiment where the outcome is continuous, ex: dollars spent, time spent, etc.).&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
