Imagine you’re a budding chef. A datacurious one, of course, so you start by taking a set of foods (pizza, salad, spaghetti, etc.) and ask 10 friends how much of each they ate in the past day.
Your goal: to find natural groups of foodies, so that you can better cater to each cluster’s tastes. For example, your fratboy friends might love wings and beer, your anime friends might love soba and sushi, your hipster friends probably dig tofu, and so on.
So how can you use the data you’ve gathered to discover different kinds of groups?
One way is to use a standard clustering algorithm like kmeans or Gaussian mixture modeling (see this previous post for a brief introduction). The problem is that these both assume a fixed number of clusters, which they need to be told to find. There are a couple methods for selecting the number of clusters to learn (e.g., the gap and prediction strength statistics), but the problem is a more fundamental one: most realworld data simply doesn’t have a fixed number of clusters.
That is, suppose we’ve asked 10 of our friends what they ate in the past day, and we want to find groups of eating preferences. There’s really an infinite number of foodie types (carnivore, vegan, snacker, Italian, healthy, fast food, heavy eaters, light eaters, and so on), but with only 10 friends, we simply don’t have enough data to detect them all. (Indeed, we’re limited to 10 clusters!) So whereas kmeans starts with the incorrect assumption that there’s a fixed, finite number of clusters that our points come from, no matter if we feed it more data, what we’d really like is a method positing an infinite number of hidden clusters that naturally arise as we ask more friends about their food habits. (For example, with only 2 data points, we might not be able to tell the difference between vegans and vegetarians, but with 200 data points, we probably could.)
Luckily for us, this is precisely the purview of nonparametric Bayes.*
*Nonparametric Bayes refers to a class of techniques that allow some parameters to change with the data. In our case, for example, instead of fixing the number of clusters to be discovered, we allow it to grow as more data comes in.
A Generative Story
Let’s describe a generative model for finding clusters in any set of data. We assume an infinite set of latent groups, where each group is described by some set of parameters. For example, each group could be a Gaussian with a specified mean $\mu_i$ and standard deviation $\sigma_i$, and these group parameters themselves are assumed to come from some base distribution $G_0$. Data is then generated in the following manner:
 Select a cluster.
 Sample from that cluster to generate a new point.
(Note the resemblance to a finite mixture model.)
For example, suppose we ask 10 friends how many calories of pizza, salad, and rice they ate yesterday. Our groups could be:
 A Gaussian centered at (pizza = 5000, salad = 100, rice = 500) (i.e., a pizza lovers group).
 A Gaussian centered at (pizza = 100, salad = 3000, rice = 1000) (maybe a vegan group).
 A Gaussian centered at (pizza = 100, salad = 100, rice = 10000) (definitely Asian).
 …
When deciding what to eat when she woke up yesterday, Alice could have thought girl, I’m in the mood for pizza and her food consumption yesterday would have been a sample from the pizza Gaussian. Similarly, Bob could have spent the day in Chinatown, thereby sampling from the Asian Gaussian for his day’s meals. And so on.
The big question, then, is: how do we assign each friend to a group?
Assigning Groups
Chinese Restaurant Process
One way to assign friends to groups is to use a Chinese Restaurant Process. This works as follows: Imagine a restaurant where all your friends went to eat yesterday…
 Initially the restaurant is empty.
 The first person to enter (Alice) sits down at a table (selects a group). She then orders food for the table (i.e., she selects parameters for the group); everyone else who joins the table will then be limited to eating from the food she ordered.
 The second person to enter (Bob) sits down at a table. Which table does he sit at? With probability $\alpha / (1 + \alpha)$ he sits down at a new table (i.e., selects a new group) and orders food for the table; with probability $1 / (1 + \alpha)$ he sits with Alice and eats from the food she’s already ordered (i.e., he’s in the same group as Alice).
 …
 The (n+1)st person sits down at a new table with probability $\alpha / (n + \alpha)$, and at table k with probability $n_k / (n + \alpha)$, where $n_k$ is the number of people currently sitting at table k.
Note a couple things:
 The more people (data points) there are at a table (cluster), the more likely it is that people (new data points) will join it. In other words, our groups satisfy a rich get richer property.
 There’s always a small probability that someone joins an entirely new table (i.e., a new group is formed).
 The probability of a new group depends on $\alpha$. So we can think of $\alpha$ as a dispersion parameter that affects the dispersion of our datapoints. The lower alpha is, the more tightly clustered our data points; the higher it is, the more clusters we have in any finite set of points.
(Also notice the resemblance between table selection probabilities and a Dirichlet distribution…)
Just to summarize, given n data points, the Chinese Restaurant Process specifies a distribution over partitions (table assignments) of these points. We can also generate parameters for each partition/table from a base distribution $G_0$ (for example, each table could represent a Gaussian whose mean and standard deviation are sampled from $G_0$), though to be clear, this is not part of the CRP itself.
Code
Since code makes everything better, here’s some Ruby to simulate a CRP:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 

And here’s some sample output:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 

Notice that as we increase $\alpha$, so too does the number of distinct tables increase.
Polya Urn Model
Another method for assigning friends to groups is to follow the Polya Urn Model. This is basically the same model as the Chinese Restaurant Process, just with a different metaphor.
 We start with an urn containing $\alpha G_0(x)$ balls of “color” $x$, for each possible value of $x$. ($G_0$ is our base distribution, and $G_0(x)$ is the probability of sampling $x$ from $G_0$). Note that these are possibly fractional balls.
 At each time step, draw a ball from the urn, note its color, and then drop both the original ball plus a new ball of the same color back into the urn.
Note the connection between this process and the CRP: balls correspond to people (i.e., data points), colors correspond to table assignments (i.e., clusters), alpha is again a dispersion parameter (put differently, a prior), colors satisfy a richgetricher property (since colors with many balls are more likely to get drawn), and so on. (Again, there’s also a connection between this urn model and the urn model for the (finite) Dirichlet distribution…)
To be precise, the difference between the CRP and the Polya Urn Model is that the CRP specifies only a distribution over partitions (i.e., table assignments), but doesn’t assign parameters to each group, whereas the Polya Urn Model does both.
Code
Again, here’s some code for simulating a Polya Urn Model:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 

And here’s some sample output, using a uniform distribution over the unit interval as the color distribution to sample from:
1 2 3 4 5 6 

Code, Take 2
Here’s the same code for a Polya Urn Model, but in R:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 

Here are some sample density plots of the colors in the urn, when using a unit normal as the base color distribution:
Notice that as alpha increases (i.e., we sample more new ball colors from our base; i.e., as we place more weight on our prior), the colors in the urn tend to a unit normal (our base color distribution).
And here are some sample plots of points generated by the urn, for varying values of alpha:
 Each color in the urn is sampled from a uniform distribution over [0,10]x[0,10] (i.e., a [0, 10] square).
 Each group is a Gaussian with standard deviation 0.1 and mean equal to its associated color, and these Gaussian groups generate points.
Notice that the points clump together in fewer clusters for low values of alpha, but become more dispersed as alpha increases.
StickBreaking Process
Imagine running either the Chinese Restaurant Process or the Polya Urn Model without stop. For each group $i$, this gives a proportion $w_i$ of points that fall into group $i$.
So instead of running the CRP or Polya Urn model to figure out these proportions, can we simply generate them directly?
This is exactly what the StickBreaking Process does:
 Start with a stick of length one.
 Generate a random variable $\beta_1 \sim Beta(1, \alpha)$. By the definition of the Beta distribution, this will be a real number between 0 and 1, with expected value $1 / (1 + \alpha)$. Break off the stick at $\beta_1$; $w_1$ is then the length of the stick on the left.
 Now take the stick to the right, and generate $\beta_2 \sim Beta(1, \alpha)$. Break off the stick $\beta_2$ into the stick. Again, $w_2$ is the length of the stick to the left, i.e., $w_2 = (1  \beta_1) \beta_2$.
 And so on.
Thus, the StickBreaking process is simply the CRP or Polya Urn Model from a different point of view. For example, assigning customers to table 1 according to the Chinese Restaurant Process is equivalent to assigning customers to table 1 with probability $w_1$.
Code
Here’s some R code for simulating a StickBreaking process:
1 2 3 4 5 6 7 8 9 10 11 12 13 

And here’s some sample output:
Notice that for low values of alpha, the stick weights are concentrated on the first few weights (meaning our data points are concentrated on a few clusters), while the weights become more evenly dispersed as we increase alpha (meaning we posit more clusters in our data points).
Dirichlet Process
Suppose we run a Polya Urn Model several times, where we sample colors from a base distribution $G_0$. Each run produces a distribution of colors in the urn (say, 5% blue balls, 3% red balls, 2% pink balls, etc.), and the distribution will be different each time (for example, 5% blue balls in run 1, but 1% blue balls in run 2).
For example, let’s look again at the plots from above, where I generated samples from a Polya Urn Model with the standard unit normal as the base distribution:
Each run of the Polya Urn Model produces a slighly different distribution, though each is “centered” in some fashion around the standard Gaussian I used as base. In other words, the Polya Urn Model gives us a distribution over distributions (we get a distribution of ball colors, and this distribution of colors changes each time) – and so we finally get to the Dirichlet Process.
Formally, given a base distribution $G_0$ and a dispersion parameter $\alpha$, a sample from the Dirichlet Process $DP(G_0, \alpha)$ is a distribution $G \sim DP(G_0, \alpha)$. This sample $G$ can be thought of as a distribution of colors in a single simulation of the Polya Urn Model; sampling from $G$ gives us the balls in the urn.
So here’s the connection between the Chinese Restaurant Process, the Polya Urn Model, the StickBreaking Process, and the Dirichlet Process:
 Dirichlet Process: Suppose we want samples $x_i \sim G$, where $G$ is a distribution sampled from the Dirichlet Process $G \sim DP(G_0, \alpha)$.
 Polya Urn Model: One way to generate these values $x_i$ would be to take a Polya Urn Model with color distribution $G_0$ and dispersion $\alpha$. ($x_i$ would be the color of the ith ball in the urn.)
 Chinese Restaurant Process: Another way to generate $x_i$ would be to first assign tables to customers according to a Chinese Restaurant Process with dispersion $\alpha$. Every customer at the nth table would then be given the same value (color) sampled from $G_0$. ($x_i$ would be the value given to the ith customer; $x_i$ can also be thought of as the food at table $i$, or as the parameters of table $i$.)
 StickBreaking Process: Finally, we could generate weights $w_k$ according to a StickBreaking Process with dispersion $\alpha$. Next, we would give each weight $w_k$ a value (or color) $v_k$ sampled from $G_0$. Finally, we would assign $x_i$ to value (color) $v_k$ with probability $w_k$.
Recap
Let’s summarize what we’ve discussed so far.
We have a bunch of data points $p_i$ that we want to cluster, and we’ve described four essentially equivalent generative models that allow us to describe how each cluster and point could have arisen.
In the Chinese Restaurant Process:
 We generate table assignments $g_1, \ldots, g_n \sim CRP(\alpha)$ according to a Chinese Restaurant Process. ($g_i$ is the table assigned to datapoint $i$.)
 We generate table parameters $\phi_1, \ldots, \phi_m \sim G_0$ according to the base distribution $G_0$, where $\phi_k$ is the parameter for the kth distinct group.
 Given table assignments and table parameters, we generate each datapoint $p_i \sim F(\phi_{g_i})$ from a distribution $F$ with the specified table parameters. (For example, $F$ could be a Gaussian, and $\phi_i$ could be a parameter vector specifying the mean and standard deviation).
In the Polya Urn Model:
 We generate colors $\phi_1, \ldots, \phi_n \sim Polya(G_0, \alpha)$ according to a Polya Urn Model. ($\phi_i$ is the color of the ith ball.)
 Given ball colors, we generate each datapoint $p_i \sim F(\phi_i)$.
In the StickBreaking Process:
 We generate group probabilities (stick lengths) $w_1, \ldots, w_{\infty} \sim Stick(\alpha)$ according to a StickBreaking process.
 We generate group parameters $\phi_1, \ldots, \phi_{\infty} \sim G_0$ from $G_0$, where $\phi_k$ is the parameter for the kth distinct group.
 We generate group assignments $g_1, \ldots, g_n \sim Multinomial(w_1, \ldots, w_{\infty})$ for each datapoint.
 Given group assignments and group parameters, we generate each datapoint $p_i \sim F(\phi_{g_i})$.
In the Dirichlet Process:
 We generate a distribution $G \sim DP(G_0, \alpha)$ from a Dirichlet Process with base distribution $G_0$ and dispersion parameter $\alpha$.
 We generate grouplevel parameters $x_i \sim G$ from $G$, where $x_i$ is the group parameter for the ith datapoint. (Note: this is not the same as $\phi_i$. $x_i$ is the parameter associated to the group that the ith datapoint belongs to, whereas $\phi_k$ is the parameter of the kth distinct group.)
 Given grouplevel parameters $x_i$, we generate each datapoint $p_i \sim F(x_i)$.
Also, remember that each model naturally allows the number of clusters to grow as more points come in.
Inference in the Dirichlet Process Mixture
So we’ve described a generative model that allows us to calculate the probability of any particular set of group assignments to data points, but we haven’t described how to actually learn a good set of group assignments.
Let’s briefly do this now. Very roughly, the Gibbs sampling approach works as follows:
 Take the set of data points, and randomly initialize group assignments.
 Pick a point. Fix the group assignments of all the other points, and assign the chosen point a new group (which can be either an existing cluster or a new cluster) with a CRPish probability (as described in the models above) that depends on the group assignments and values of all the other points.
 We will eventually converge on a good set of group assignments, so repeat the previous step until happy.
For more details, this paper provides a good description. Philip Resnick and Eric Hardisty also have a friendlier, more general description of Gibbs sampling (plus an application to naive Bayes) here.
Fast Food Application: Clustering the McDonald’s Menu
Finally, let’s show an application of the Dirichlet Process Mixture. Unfortunately, I didn’t have a data set of people’s food habits offhand, so instead I took this list of McDonald’s foods and nutrition facts.
After normalizing each item to have an equal number of calories, and representing each item as a vector of (total fat, cholesterol, sodium, dietary fiber, sugars, protein, vitamin A, vitamin C, calcium, iron, calories from fat, satured fat, trans fat, carbohydrates), I ran scikitlearn’s Dirichlet Process Gaussian Mixture Model to cluster McDonald’s menu based on nutritional value.
First, how does the number of clusters inferred by the Dirichlet Process mixture vary as we feed in more (randomly ordered) points?
As expected, the Dirichlet Process model discovers more and more clusters as more and more food items arrive. (And indeed, the number of clusters appears to grow logarithmically, which can in fact be proved.)
How many clusters does the mixture model infer from the entire dataset? Running the Gibbs sampler several times, we find that the number of clusters tends around 11:
Let’s dive into one of these clusterings.
Cluster 1 (Desserts)
Looking at a sample of foods from the first cluster, we find a lot of desserts and desserty drinks:
 Caramel Mocha
 Frappe Caramel
 Iced Hazelnut Latte
 Iced Coffee
 Strawberry Triple Thick Shake
 Snack Size McFlurry
 Hot Caramel Sundae
 Baked Hot Apple Pie
 Cinnamon Melts
 Kiddie Cone
 Strawberry Sundae
We can also look at the nutritional profile of some foods from this cluster (after zscaling each nutrition dimension to have mean 0 and standard deviation 1):
We see that foods in this cluster tend to be high in trans fat and low in vitamins, protein, fiber, and sodium.
Cluster 2 (Sauces)
Here’s a sample from the second cluster, which contains a lot of sauces:
 Hot Mustard Sauce
 Spicy Buffalo Sauce
 Newman’s Own Low Fat Balsamic Vinaigrette
And looking at the nutritional profile of points in this cluster, we see that it’s heavy in sodium and fat:
Cluster 3 (Burgers, Crispy Foods, HighCholesterol)
The third cluster is very burgery:
 Hamburger
 Cheeseburger
 FiletOFish
 Quarter Pounder with Cheese
 Premium Grilled Chicken Club Sandwich
 Ranch Snack Wrap
 Premium Asian Salad with Crispy Chicken
 Butter Garlic Croutons
 Sausage McMuffin
 Sausage McGriddles
It’s also high in fat and sodium, and low in carbs and sugar
Cluster 4 (Creamy Sauces)
Interestingly, even though we already found a cluster of sauces above, we discover another one as well. These sauces appear to be much more creambased:
 Creamy Ranch Sauce
 Newman’s Own Creamy Caesar Dressing
 Coffee Cream
 Iced Coffee with Sugar Free Vanilla Syrup
Nutritionally, these sauces are higher in calories from fat, and much lower in sodium:
Cluster 5 (Salads)
Here’s a salad cluster. A lot of salads also appeared in the third cluster (along with hamburgers and McMuffins), but that’s because those salads also all contained crispy chicken. The salads in this cluster are either crispfree or have their chicken grilled instead:
 Premium Southwest Salad with Grilled Chicken
 Premium Caesar Salad with Grilled Chicken
 Side Salad
 Premium Asian Salad without Chicken
 Premium Bacon Ranch Salad without Chicken
This is reflected in the higher content of iron, vitamin A, and fiber:
Cluster 6 (More Sauces)
Again, we find another cluster of sauces:
 Ketchup Packet
 Barbeque Sauce
 Chipotle Barbeque Sauce
These are still high in sodium, but much lower in fat compared to the other sauce clusters:
Cluster 7 (Fruit and Maple Oatmeal)
Amusingly, fruit and maple oatmeal is in a cluster by itself:
 Fruit & Maple Oatmeal
Cluster 8 (Sugary Drinks)
We also get a cluster of sugary drinks:
 Strawberry Banana Smoothie
 Wild Berry Smoothie
 Iced Nonfat Vanilla Latte
 Nonfat Hazelnut
 Nonfat Vanilla Cappuccino
 Nonfat Caramel Cappuccino
 Sweet Tea
 Frozen Strawberry Lemonade
 CocaCola
 Minute Maid Orange Juice
In addition to high sugar content, this cluster is also high in carbohydrates and calcium, and low in fat.
Cluster 9 (Breakfast Foods)
Here’s a cluster of highcholesterol breakfast foods:
 Sausage McMuffin with Egg
 Sausage Burrito
 Egg McMuffin
 Bacon, Egg & Cheese Biscuit
 McSkillet Burrito with Sausage
 Big Breakfast with Hotcakes
Cluster 10 (Coffee Drinks)
We find a group of coffee drinks next:
 Nonfat Cappuccino
 Nonfat Latte
 Nonfat Latte with Sugar Free Vanilla Syrup
 Iced Nonfat Latte
These are much higher in calcium and protein, and lower in sugar, than the other drink cluster above:
Cluster 11 (Apples)
Here’s a cluster of apples:
 Apple Dippers with Low Fat Caramel Dip
 Apple Slices
Vitamin C, check.
And finally, here’s an overview of all the clusters at once (using a different clustering run):
No More!
I’ll end with a couple notes:
 Kevin Knight has a hilarious introduction to Bayesian inference that describes some applications of nonparametric Bayesian techniques to computational linguistics (though I don’t think he ever quite says “nonparametric Bayes” directly).
 In the Chinese Restaurant Process, each customer sits at a single table. The Indian Buffet Process is an extension that allows customers to sample food from multiple tables (i.e., belong to multiple clusters).
 The Chinese Restaurant Process, the Polya Urn Model, and the StickBreaking Process are all sequential models for generating groups: to figure out table parameters in the CRP, for example, you wait for customer 1 to come in, then customer 2, then customer 3, and so on. The equivalent Dirichlet Process, on the other hand, is a parallel model for generating groups: just sample $G \sim DP(G_0, alpha)$, and then all your group parameters can be independently generated by sampling from $G$ at once. This duality is an instance of a more general phenomenon known as de Finetti’s theorem.
And that’s it.