BayesMallows is an R package for analyzing preference data in the form of rankings with the Mallows rank model, and its finite mixture extension, in a Bayesian framework. The model is grounded on the idea that the probability density of an observed ranking decreases exponentially with the distance to the location parameter. It is the first Bayesian implementation that allows wide choices of distances, and it works well with a large amount of items to be ranked. BayesMallows handles non-standard data: partial rankings and pairwise comparisons, even in cases including non-transitive preference patterns. The Bayesian paradigm allows coherent quantification of posterior uncertainties of estimates of any quantity of interest. These posteriors are fully available to the user, and the package comes with convienient tools for summarizing and visualizing the posterior distributions.
Preference data are usually collected when individuals are asked to rank a set of items according to a certain preference criterion. The booming of internet-related activities and applications in recent years has led to a rapid increase of the amount of data that have ranks as their natural scale, however often in the form of partial or indirect rankings (pairwise preferences, ratings, clicks). The amount of readily available software handling preference data has consequently increased consistently in the last decade or so, but not many packages are flexible enough to handle all types of data mentioned above. The typical tasks when analyzing preference data are rank aggregation, classification or clustering, and prediction, where the latter task refers to the estimation of the individual rankings of the assessors when completely or partially missing. These tasks can be addressed either via model-based inference or via heuristic machine learning algorithms, with or without uncertainty quantification. However, very few methods allow handling diverse data types, combining several inferential tasks with proper propagation of the uncertainties, while also providing individualized predictions. Our proposal goes exactly in this direction, thus making the scopes of BayesMallows broad when it comes to data handling and individual-level inference.
The R package BayesMallows is the first software conceived to answer the needs mentioned above in a unified framework: it implements full Bayesian inference for ranking data, and performs all of the tasks above in the framework of the Bayesian Mallows model (BMM) (Mallows 1957; Vitelli et al. 2018). More specifically, BayesMallows allows for data in the forms of complete rankings, partial rankings, as well as pairwise comparisons, including the case where some comparisons are inconsistent. In these situations, it provides all Bayesian inferential tools for rank modeling with the BMM: it performs rank aggregation (estimation of a consensus ranking of the items), it can cluster together the assessors providing similar preferences (estimating both cluster specific model parameters, and individual cluster assignments, with uncertainty), it performs data augmentation for estimating the latent assessor-specific full ranking of the items in all missing data situations (partial rankings, pairwise preferences). The latter in particular, i.e., the possibility of predicting individual preferences for unranked items, enables the model to be used as a probabilistic recommender system. BayesMallows also enlarges the pool of distances that can be used in the Mallows model, and it supports the rank distances most used in the literature: Spearman’s footrule (henceforth footrule), Spearman’s rank correlation (henceforth Spearman), Cayley, Hamming, Kendall, and Ulam distances (we refer to Diaconis 1988; Marden 1995 for details on these). Finally, BayesMallows implements the Iterative Proportional Fitting Procedure (IPFP) algorithm for computing the partition function for the Mallows model (MM) (Mukherjee 2016) and the importance sampling algorithm described in Vitelli et al. (2018). In addition to being used in the MM, these functions may be of interest in their own right.
Comparing with other available inferential software, we notice that not many packages allow for such flexibility, very few in combination with full Bayesian inference, and none when using the MM as outlined in Section 4 below. Often machine learning approaches focus on either rank aggregation (i.e., consensus estimation), or individual rank prediction, while BayesMallows handles both. Since the BMM is fully Bayesian, all posterior quantities of interest are automatically available from BayesMallows for the first time for the MM. In addition, the package also has tools for visualizing posterior distributions, and hence, posterior quantities as well as their associated uncertainties. Uncertainty quantification is often critical in real applications: for recommender systems, the model should not spam the users with very uncertain recommendations; when performing subtype identification for cancer patients, a very uncertain cluster assignment might have serious consequences for the clinical practice, for example in treatment choice. The package also works well with a fairly large number of items, thanks to computational approximations and efficient programming. In conclusion, BayesMallows provides the first fully probabilistic inferential tool for the MM with many different distances. It is flexible in the types of data it handles, and computationally efficient. We therefore think that this package will gain popularity, and prove its usefulness in many practical situations, many of which we probably cannot foresee now.
The paper is organized as follows. The BMM for ranking data is briefly reviewed in Section 2, as well as its model extensions to different data types and to mixtures. Section 3 includes details on the implementation of the inferential procedure. For a thorough discussion of both the model and its implementation we refer interested readers to Vitelli et al. (2018). An overview of existing R packages implementing the Mallows model (MM) is given in Section 4. The use of the BayesMallows package is presented, in the form of three case studies, in Sections 5, 6, and 7. Section 8 concludes the paper, also discussing model extensions that will come with new releases of the package.
In this section we give the background for understanding the functions in the BayesMallows package. More details can be found in Vitelli et al. (2018) and Liu et al. (2019). The section is organized as follows: we first clarify the notations that we will use throughout the paper (Section 2.1). We then briefly describe the BMM for complete ranking data (Section 2.2), also focusing on the relevance of the choice of distance (Section 2.3). The last two sections focus on model extensions: partial and pairwise data (Section 2.4), non-transitive pairwise comparisons (Section 2.5), and mixtures (Section 2.6).
Let us denote with
The MM for ranking data (Mallows 1957) specifies the probability density
for a ranking
In the complete data case,
Inference on the model parameters is based on a Metropolis-Hastings (M-H) Markov Chain Monte Carlo (MCMC) algorithm, described in Vitelli et al. (2018). Some details relevant for a correct use of this package are also given in Section 3.1.
The partition function
When complete rankings of all items are not readily available, the BMM
can still be used by applying data augmentation techniques. Partial
rankings can occur because ranks are missing at random, because the
assessors have only ranked their top-
Let us start by considering the case of top-
The above procedure can also handle more general situations where
missing rankings are not necessarily the bottom ones, and where each
assessor is asked to provide the mutual ranking of some possibly random
subset
In the case of pairwise comparison data, let us call
It can happen in real applications that individual pairwise comparison
data are non-transitive, that is, they may contain a pattern of the form
The key ingredient of this generalization consists of adding one layer
of latent variables to the model hierarchy, accounting for the fact that
assessors can make mistakes. The main assumption is that the assessor
makes pairwise comparisons based on her latent full rankings
Two models for (6) are considered in Crispino et al. (2019): the
Bernoulli model, which accounts for random mistakes, and the Logistic
model, which lets the probability of making a mistake depend on the
similarity of the items being compared. The Bernoulli model states that:
The assumption, implicit in the discussion so far, that there exists a unique consensus ranking shared by all assessors is unrealistic in most real applications: the BMM thus handles the case where the rankings provided by all assessors can be modeled as a finite mixture of MMs. In the following brief discussion we assume that the data consist of complete rankings, but BayesMallows can fit a mixture based on any kinds of preference data described so far. See Section 4.3 of Vitelli et al. (2018) for details.
Let
In this section we briefly give some additional details regarding the implementation of the models described in Section 2. The BMM implementation is thoroughly described in Vitelli et al. (2018).
In order to obtain samples from the posterior density of equation
(4), BayesMallows implements an MCMC scheme
iterating between (i) updating
As mentioned in Section 2.4, the MCMC procedure for
sampling from the posterior densities corresponding to the partial data
cases (Algorithm 3 of Vitelli et al. 2018) has an additional M-H step to
account for the update of the augmented data rankings
When the distance in the BMM is footrule or Spearman, the partition
function
The package contains integer sequences for exact calculation of the
partition function with footrule distance for up to
The IS procedure can be used to compute an off-line approximation
The IPFP algorithm (Mukherjee 2016 Theorem 1.8) yields a numeric
evaluation of
A simulation experiment was conducted comparing the methods for
estimating
To obtain random samples from the MM with Cayley, Hamming, Kendall, or
Ulam distance and fixed
This section gives an overview of existing packages for fitting the MM.
lmm
is possible
when dbm
returns the MLE of dbm
generates a matrix of size pmr
was not able to handle a
ranking dataset with BayesMallows provides many new functionalities not implemented in these packages, as will be illustrated in the use cases of the following three sections.
We illustrate the case of complete rankings with the potato datasets
described in Liu et al. (2019 4). In short, a bag of 20 potatoes was
bought, and 12 assessors were asked to rank the potatoes by weight,
first by visual inspection, and next by holding the potatoes in hand.
These datasets are available in BayesMallows as matrices with names
potato_weighing
and potato_visual
, respectively. The true ranking of
the potatoes’ weights is available in the vector potato_true_ranking
.
In general, compute_mallows
expects ranking datasets to have one row
for each assessor and one column for each item. Each row has to be a
proper permutation, possibly with missing values. We are interested in
the posterior distribution of both the level of agreement between
assessors, as described by
First, we do a test run to check convergence of the MCMC algorithm, and
then get trace plots with assess_convergence
.
bmm_test <- compute_mallows(potato_visual)
assess_convergence(bmm_test)
By default, assess_convergence
returns a trace plot for items
argument.
assess_convergence(bmm_test, parameter = "rho", items = 1:5)
The corresponding plot is shown in Figure 2b, illustrating that the MCMC algorithm seems to have converged after around 1,000 iterations.
|
|
|
|
potato_visual
dataset. The plots indicating good mixing for
both parameters after about 500 iterations.
potato_visual
dataset. The posterior mass is symmetrically
centered between 9 and 13, with a mean around
11.From the trace plots, we decide to discard the first 1,000 MCMC samples
as burn-in. We rerun the algorithm to get 500,000 samples after burn-in.
The object bmm_visual
has S3
class "BayesMallows", so we plot the
posterior distribution of plot.BayesMallows
.
bmm_visual <- compute_mallows(potato_visual, nmc = 501000)
bmm_visual$burnin <- 1000 # Set burn-in to 1000
plot(bmm_visual) # Use S3 method for plotting
The plot is shown in Figure 3. We can
also get posterior credible intervals for
compute_posterior_intervals
, which returns both highest posterior
density intervals (HPDI) and central intervals in a tibble
(Müller and Wickham 2018). BayesMallows uses tibble
s rather than data.frame
s,
but both are accepted as function inputs. We refer to tibble
s as
dataframes henceforth.
compute_posterior_intervals(bmm_visual, decimals = 1L)
# A tibble: 1 x 6
parameter mean median conf_level hpdi central_interval
<ch <dbl> <dbl> <ch <ch <chr>
1 alpha 10.9 10.9 95 % [9.4,12.3] [9.5,12.3]
Next, we can go on to study the posterior distribution of
plot(bmm_visual, parameter = "rho", items = 1:20)
If the items
argument is not provided, and the number of items exceeds
five, five items are picked at random for plotting. To show all
potatoes, we explicitly set items = 1:20
. The corresponding plots are
shown in Figure 4.
potato_visual
dataset. Most potatoes have
highly peaked posterior distributions, indicating low uncertainty about
their
ranking.Updating alpha_jump
argument, we can tell the MCMC algorithm to update
alpha_jump
-th iteration. To update
bmm <- compute_mallows(potato_visual, nmc = 501000, alpha_jump = 10)
On a MacBook Pro 2.2 GHz Intel Core i7 running R version 3.5.1, the
above call ran in 2.0 seconds on average over 1,000 replications using
microbenchmark
(Mersmann 2018), while it took 4.2 seconds using the default value
alpha_jump = 1
, i.e., updating
By default, compute_mallows
uses the footrule distance, but the user
can also choose to use Cayley, Kendall, Hamming, Spearman, or Ulam
distance. Running the same analysis of the potato data with Spearman
distance is done with the command
bmm <- compute_mallows(potato_visual, metric = "spearman", nmc = 501000)
For the particular case of Spearman distance, BayesMallows only has integer sequences for computing the exact partition function with 14 or fewer items. In this case a precomputed importance sampling estimate is part of the package, and used instead.
Unless the argument error_model
to compute_mallows
is set, pairwise
preference data are assumed to be consistent within each assessor. These
data should be provided in a dataframe with the following three columns,
with one row per pairwise comparison.
assessor
is an identifier for the assessor; either a numeric
vector containing the assessor index, or a character vector
containing the unique name of the assessor.bottom_item
is a numeric vector containing the index of the item
that was disfavored in each pairwise comparison.top_item
is a numeric vector containing the index of the item that
was preferred in each pairwise comparison.A dataframe with this structure can be given in the preferences
argument to compute_mallows
. compute_mallows
will generate the full
set of implied rankings for each assessor using the function
generate_transitive_closure
, as well as an initial ranking matrix
consistent with the pairwise preferences, using the function
generate_initial_ranking
.
We illustrate with the beach preference data containing stated pairwise
preferences between random subsets of 15 images of beaches, by 60
assessors (Vitelli et al. 2018 6.2). This dataset is provided in the
dataframe beach_preferences
.
We start by generating the transitive closure implied by the pairwise preferences.
beach_tc <- generate_transitive_closure(beach_preferences)
The dataframe beach_tc
contains all the pairwise preferences in
beach_preferences
, with all the implied pairwise preferences in
addition. The latter are preferences that were not specifically stated
by the assessor, but instead are implied by the stated preferences. As a
consequence, the dataframe beach_tc
has 2921 rows, while
beach_preferences
has 1442 rows. Initial rankings, i.e., a set of full
rankings for each assessor that are consistent with the implied pairwise
preferences are then generated, and we set the column names of the
initial ranking matrix to "Beach 1", "Beach 2", ..., "Beach 15"
in order have these names appear as labels in plots and output.
beach_init_rank <- generate_initial_ranking(beach_tc)
colnames(beach_init_rank) <- paste("Beach", 1:ncol(beach_init_rank))
If we had not generated the transitive closure and the initial ranking,
compute_mallows
would do this for us, but when calling
compute_mallows
repeatedly, it may save time to have these precomputed
and saved for future re-use. In order to save time in the case of big
datasets, the functions for generating transitive closures and initial
rankings from transitive closures can all be run in parallel, as shown
in the examples to the compute_mallows
function. The key to the
parallelization is that each assessor’s preferences can be handled
independently of the others, and this can speed up the process
considerably with large dataset.
As an example, let us look at all preferences stated by assessor 1
involving beach 2. We use filter
from
dplyr (Wickham et al. 2018)
to obtain the right set of rows.
library("dplyr")
# All preferences stated by assessor 1 involving item 2
filter(beach_preferences, assessor == 1, bottom_item == 2 | top_item == 2)
# A tibble: 1 x 3
assessor bottom_item top_item
<dbl> <dbl> <dbl>
1 1 2 15
Assessor 1 has performed only one direct comparison involving beach 2, in which the assessor stated that beach 15 is preferred to beach 2. The implied orderings, on the other hand, contain two preferences involving beach 2:
# All implied orderings for assessor 1 involving item 2
filter(beach_tc, assessor == 1, bottom_item == 2 | top_item == 2)
assessor bottom_item top_item
1 1 2 6
2 1 2 15
In addition to the statement that beach 15 is preferred to beach 2, all the other orderings stated by assessor 1 imply that this assessor prefers beach 6 to beach 2.
As with the potato data, we can do a test run to assess the convergence
of the MCMC algorithm. However, this time we provide the initial
rankings beach_init_rank
to the rankings
argument and the transitive
closure beach_tc
to the preferences
argument of compute_mallows
.
We also set save_aug = TRUE
to save the augmented rankings in each
MCMC step, hence letting us assess the convergence of the augmented
rankings.
bmm_test <- compute_mallows(rankings = beach_init_rank,
preferences = beach_tc, save_aug = TRUE)
Running assess_convergence
for parameter = "Rtilde"
,
and also specify which items and assessors to plot. Let us start by
considering items 2, 6, and 15 for assessor 1, which we studied above.
assess_convergence(bmm_test, parameter = "Rtilde",
items = c(2, 6, 15), assessors = 1)
The resulting plot is shown in Figure 5a. It illustrates how the augmented rankings vary, while also obeying their implied ordering.
|
|
|
|
By further investigation of beach_tc
, we would find that no orderings
are implied between beach 1 and beach 15 for assessor 2. With the
following command, we create trace plots to confirm this:
assess_convergence(bmm_test, parameter = "Rtilde",
items = c(1, 15), assessors = 2)
The resulting plot is shown in Figure 5b. As
expected, the traces of the augmented rankings for beach 1 and 15 for
assessor 2 do cross each other, since no ordering is implied between
them. Ideally, we should look at trace plots for augmented ranks for
more assessors to be sure that the algorithm is close to convergence. We
can plot assessors 1-8 by setting assessors = 1:8
. We also quite
arbitrarily pick items 13-15, but the same procedure can be repeated for
other items.
assess_convergence(bmm_test, parameter = "Rtilde",
items = 13:15, assessors = 1:8)
The resulting plot is shown in Figure 6, indicating good mixing.
Based on the convergence diagnostics, and being fairly conservative, we discard the first 2,000 MCMC iterations as burn-in, and take 100,000 additional samples.
bmm_beaches <- compute_mallows(rankings = beach_init_rank, preferences = beach_tc,
nmc = 102000, save_aug = TRUE)
bmm_beaches$burnin <- 2000
The posterior distributions of compute_posterior_intervals
:
compute_posterior_intervals(bmm_beaches, parameter = "rho")
# A tibble: 15 x 7
item parameter mean median conf_level hpdi central_interval
<fct> <chr> <dbl> <dbl> <chr> <chr> <chr>
1 Beach 1 rho 7 7 95 % [7] [6,7]
2 Beach 2 rho 15 15 95 % [15] [15]
3 Beach 3 rho 3 3 95 % [3,4] [3,4]
4 Beach 4 rho 12 12 95 % [11,13] [11,14]
5 Beach 5 rho 9 9 95 % [8,10] [8,10]
6 Beach 6 rho 2 2 95 % [1,2] [1,2]
7 Beach 7 rho 9 8 95 % [8,10] [8,10]
8 Beach 8 rho 12 11 95 % [11,13] [11,14]
9 Beach 9 rho 1 1 95 % [1,2] [1,2]
10 Beach 10 rho 6 6 95 % [5,6] [5,7]
11 Beach 11 rho 4 4 95 % [3,4] [3,5]
12 Beach 12 rho 13 13 95 % [12,14] [12,14]
13 Beach 13 rho 10 10 95 % [8,10] [8,10]
14 Beach 14 rho 13 14 95 % [12,14] [11,14]
15 Beach 15 rho 5 5 95 % [5,6] [4,6]
We can also rank the beaches according to their cumulative probability
(CP) consensus (Vitelli et al. 2018 5.1) and their maximum posterior
(MAP) rankings. This is done with the function compute_consensus
, and
the following call returns the CP consensus:
compute_consensus(bmm_beaches, type = "CP")
# A tibble: 15 x 3
ranking item cumprob
<dbl> <chr> <dbl>
1 1 Beach 9 0.896
2 2 Beach 6 1
3 3 Beach 3 0.738
4 4 Beach 11 0.966
5 5 Beach 15 0.953
6 6 Beach 10 0.971
7 7 Beach 1 1
8 8 Beach 7 0.528
9 9 Beach 5 0.887
10 10 Beach 13 1.00
11 11 Beach 8 0.508
12 12 Beach 4 0.717
13 13 Beach 12 0.643
14 14 Beach 14 0.988
15 15 Beach 2 1
The column cumprob
shows the probability of having the given rank or
lower. Looking at the second row, for example, this means that beach 6
has probability 1 of having latent ranking 2 or lower. Next, beach 3 has
probability 0.738 of having latent rank 3 or lower. This is an example
of how the Bayesian framework can be used to not only rank items, but
also to give posterior assessments of the uncertainty of the rankings.
The MAP consensus is obtained similarly, by setting type = "MAP"
.
Keeping in mind that the ranking of beaches is based on sparse pairwise
preferences, we can also ask: for beach plot_top_k
plots
these probabilities. By default, it sets k = 3
, so a heatplot of the
probability of being ranked top-3 is obtained with the call:
plot_top_k(bmm_beaches)
The plot is shown in Figure 7. The left part of
the plot shows the beaches ranked according to their CP consensus, and
the probability predict_top_k
returns a dataframe with all the
underlying probabilities. For example, in order to find all the beaches
that are among the top-3 of assessors 1-5 with more than 90 %
probability, we would do:
predict_top_k(bmm_beaches) %>%
filter(prob > 0.9, assessor %in% 1:5)
# A tibble: 6 x 3
# Groups: assessor [4]
assessor item prob
<dbl> <chr> <dbl>
1 1 Beach 11 0.955
2 1 Beach 6 0.997
3 3 Beach 6 0.997
4 3 Beach 9 1
5 4 Beach 9 1.00
6 5 Beach 6 0.979
Note that assessor 2 does not appear in this table, i.e., there are no beaches for which we are at least 90 % certain that the beach is among assessor 2’s top-3.
BayesMallows comes with a set of sushi preference data, in which 5,000 assessors each have ranked a set of 10 types of sushi (Kamishima 2003). It is interesting to see if we can find subsets of assessors with similar preferences. The sushi dataset was analyzed with the BMM by Vitelli et al. (2018), but the results in that paper differ somewhat from those obtained here, due to a bug in the function that was used to sample cluster probabilities from the Dirichlet distribution.
The function compute_mallows_mixtures
computes multiple Mallows models
with different numbers of mixture components. It returns a list of
models of class BayesMallowsMixtures
, in which each list element
contains a model with a given number of mixture components. Its
arguments are n_clusters
, which specifies the number of mixture
components to compute, an optional parameter cl
which can be set to
the return value of the makeCluster
function in the
parallel package, and
an ellipsis (...
) for passing on arguments to compute_mallows
.
Hypothesizing that we may not need more than 10 clusters to find a
useful partitioning of the assessors, we start by doing test runs with
1, 4, 7, and 10 mixture components in order to assess convergence. We
set the number of Monte Carlo samples to 5,000, and since this is a test
run, we do not save cluster assignments nor within-cluster distances
from each MCMC iteration and hence set save_clus = FALSE
and
include_wcd = FALSE
. We also run the computations in parallel on four
cores, one for each mixture component.
library("parallel")
cl <- makeCluster(4)
bmm <- compute_mallows_mixtures(n_clusters = c(1, 4, 7, 10),
rankings = sushi_rankings, nmc = 5000,
save_clus = FALSE, include_wcd = FALSE, cl = cl)
stopCluster(cl)
The function assess_convergence
automatically creates a grid plot when
given an object of class BayesMallowsMixtures
, so we can check the
convergence of
assess_convergence(bmm)
The resulting plot is given in Figure 8a,
showing that all the chains seem to be close to convergence quite
quickly. We can also make sure that the posterior distributions of the
cluster probabilities parameter = "cluster_probs"
.
assess_convergence(bmm, parameter = "cluster_probs")
The trace plots for each number of mixture components are shown in Figure 8b. Note that with only one cluster, the cluster probability is fixed at the value 1, while for other number of mixture components, the chains seem to be mixing well.
|
|
Trace of α. | Trace of τc. |
Given the convergence assessment of the previous section, we are fairly
confident that a burn-in of 5,000 is sufficient. We run 95,000
additional iterations, and try from 1 to 10 mixture components. Our goal
is now to determine the number of mixture components to use, and in
order to create an elbow plot, we set include_wcd = TRUE
to compute
the within-cluster distances in each step of the MCMC algorithm. Since
the posterior distributions of rho_thinning = 10
.
cl <- makeCluster(4)
bmm <- compute_mallows_mixtures(n_clusters = 1:10, rankings = sushi_rankings,
nmc = 100000, rho_thinning = 10, save_clus = FALSE,
include_wcd = TRUE, cl = cl)
stopCluster(cl)
plot_elbow(bmm, burnin = 5000) # Create elbow plot
The resulting elbow plot is a notched boxplot (Mcgill et al. 1978; Wickham 2016) shown in Figure 9, for which the barely visible upper and lower whiskers represent approximate 95 % confidence intervals. Although not clear-cut, we see that the within-cluster sum of distances levels off at around 5 clusters, and hence we choose to use 5 clusters in our model.
Having chosen 5 mixture components, we go on to fit a final model, still
running 95,000 iterations after burnin. This time we call
compute_mallows
and set n_clusters = 5
. We also set
save_clus = TRUE
and clus_thin = 10
to save the cluster assignments
of each assessor in every 10th iteration, and rho_thinning = 10
to
save the estimated latent rank every 10th iteration.
bmm <- compute_mallows(rankings = sushi_rankings, n_clusters = 5, save_clus = TRUE,
clus_thin = 10, nmc = 100000, rho_thinning = 10)
bmm$burnin <- 5000
We can plot the posterior distributions of plot.BayesMallows
as shown preivously for the
potato data. We can also show the posterior distributions of the cluster
probabilities, using:
plot(bmm, parameter = "cluster_probs")
Using the argument parameter = "cluster_assignment"
, we can visualize
the posterior probability for each assessor of belonging to each
cluster:
plot(bmm, parameter = "cluster_assignment")
The resulting plot is shown in
Figure 10. The underlying numbers can
be obtained using the function assign_cluster
.
We can find clusterwise consensus rankings using compute_consensus
.
The following call finds the CP consensuses, and then uses select
from
dplyr and spread
from
tidyr (Wickham and Henry 2018)
to create one column for each cluster. The result is shown in
Table 1.
library("tidyr")
compute_consensus(bmm) %>%
select(-cumprob) %>%
spread(key = cluster, value = item)
Cluster 1 | Cluster 2 | Cluster 3 | Cluster 4 | Cluster 5 | |
---|---|---|---|---|---|
1 | shrimp | fatty tuna | fatty tuna | fatty tuna | fatty tuna |
2 | sea eel | sea urchin | sea eel | tuna | sea urchin |
3 | egg | salmon roe | tuna | shrimp | shrimp |
4 | squid | sea eel | shrimp | tuna roll | tuna |
5 | salmon roe | tuna | tuna roll | squid | salmon roe |
6 | fatty tuna | shrimp | squid | salmon roe | squid |
7 | tuna | tuna roll | egg | egg | tuna roll |
8 | tuna roll | squid | cucumber roll | cucumber roll | sea eel |
9 | cucumber roll | egg | salmon roe | sea eel | egg |
10 | sea urchin | cucumber roll | sea urchin | sea urchin | cucumber roll |
Note that for estimating cluster specific parameters, label switching is
a potential problem that needs to be handled. BayesMallows ignores
label switching issues inside the MCMC, because it has been shown that
this approach is better for ensuring full convergence of the chain
(Celeux et al. 2000; Jasra et al. 2005). MCMC iterations can be re-ordered after
convergence is achieved, for example by using the implementation of
Stephens’ algorithm (Stephens 2000) provided by the R package
label.switching
(Papastamoulis 2016). A full example of how to assess label switching
after running compute_mallows
is provided by running the following
command:
help("label_switching")
For the sushi data analyzed in this section, no label switching is detected by Stephen’s algorithm.
In this paper we discussed the methodological background and computational strategies for the BayesMallows package, implementing the inferential framework for the analysis of preference data based on the Bayesian Mallows model, as introduced in Vitelli et al. (2018). The package aims at providing a general probabilistic tool, capable of performing various inferential tasks (estimation, classification, prediction) with a proper uncertainty quantification. Moreover, the package widens the applicability of the Mallows model, by providing reliable algorithms for approximating the associated partition function, which has been the bottleneck for a successful use of this general and flexible model so far. Finally, it handles a variety of preference data types (partial rankings, pairwise preferences), and it could possibly handle many others which can lie in the above mentioned categories (noisy continuous measurements, clicking data, ratings).
One of the most important features of the BayesMallows package is that, despite implementing a Bayesian model, and thus relying on MCMC algorithms, its efficient implementation makes it possible to manage large datasets. The package can easily handle up to hundreds of items, and thousands of assessors; an example is the Movielens data analyzed in Section 6.4 of (Vitelli et al. 2018). By using the log-sum-exp trick, the implementation of the importance sampler is able to handle at least ten thousand items without numerical overflow. We believe that all these features make the package a unique resource for fitting the Mallows model to large data, with the benefits of a fully probabilistic interpretation.
Nonetheless, we also recognize that the BayesMallows package can open
the way for further generalizations. The Bayesian Mallows model for
time-varying rankings that has been introduced in Asfaw et al. (2017) will be
considered for a future release. Some further extensions which we might
consider to implement in the BayesMallows in the future include:
fitting an infinite mixture of Mallows models for automatically
performing model selection; allowing for a non-uniform prior for
The authors would like to thank Arnoldo Frigessi and Elja Arjas for fruitful discussions.
BayesMallows, PerMallows, pmr, rankdist, microbenchmark, dplyr, parallel, tidyr, label.switching
Bayesian, Databases, MissingData, ModelDeployment
This article is converted from a Legacy LaTeX article using the texor package. The pdf version is the official version. To report a problem with the html, refer to CONTRIBUTE on the R Journal homepage.
Text and figures are licensed under Creative Commons Attribution CC BY 4.0. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".
For attribution, please cite this work as
Sørensen, et al., "BayesMallows: An R Package for the Bayesian Mallows Model", The R Journal, 2020
BibTeX citation
@article{RJ-2020-026, author = {Sørensen, Øystein and Crispino, Marta and Liu, Qinghua and Vitelli, Valeria}, title = {BayesMallows: An R Package for the Bayesian Mallows Model}, journal = {The R Journal}, year = {2020}, note = {https://rjournal.github.io/}, volume = {12}, issue = {1}, issn = {2073-4859}, pages = {324-342} }