“Dimensionality reduction” (DR) is a widely used approach to find low dimensional and interpretable representations of data that are natively embedded in high-dimensional spaces. DR can be realized by a plethora of methods with different properties, objectives, and, hence, (dis)advantages. The resulting low-dimensional data embeddings are often difficult to compare with objective criteria. Here, we introduce the dimRed and coRanking packages for the R language. These open source software packages enable users to easily access multiple classical and advanced DR methods using a common interface. The packages also provide quality indicators for the embeddings and easy visualization of high dimensional data. The coRanking package provides the functionality for assessing DR methods in the co-ranking matrix framework. In tandem, these packages allow for uncovering complex structures high dimensional data. Currently 15 DR methods are available in the package, some of which were not previously available to R users. Here, we outline the dimRed and coRanking packages and make the implemented methods understandable to the interested reader.
Dimensionality Reduction (DR) essentially aims to find low dimensional representations of data while preserving their key properties. Many methods exist in literature, optimizing different criteria: maximizing the variance or the statistical independence of the projected data, minimizing the reconstruction error under different constraints, or optimizing for different error metrics, just to name a few. Choosing an inadequate method may imply that much of the underlying structure remains undiscovered. Often the structures of interest in a data set can be well represented by fewer dimensions than exist in the original data. Data compression of this kind has the additional benefit of making the encoded information better conceivable to our brains for further analysis tasks like classification or regression problems.
For example, the morphology of a plant’s leaves, stems, and seeds reflect the environmental conditions the species usually grow in (e.g., plants with large soft leaves will never grow in a desert but might have an advantage in a humid and shadowy environment). Because the morphology of the entire plant depends on the environment, many morphological combinations will never occur in nature and the morphological space of all plant species is tightly constrained. Díaz et al. (2016) found that out of six observed morphological characteristics only two embedding dimensions were enough to represent three quarters of the totally observed variability.
DR is a widely used approach for the detection of structure in multivariate data, and has applications in a variety of fields. In climatology, DR is used to find the modes of some phenomenon, e.g., the first Empirical Orthogonal Function of monthly mean sea surface temperature of a given region over the Pacific is often linked to the El Niño Southern Oscillation or ENSO (e.g., Hsieh 2004). In ecology the comparison of sites with different species abundances is a classical multivariate problem: each observed species adds an extra dimension, and because species are often bound to certain habitats, there is a lot of redundant information. Using DR is a popular technique to represent the sites in few dimensions, e.g., Aart (1972) matches wolfspider communities to habitat and Morrall (1974) match soil fungi data to soil types. (In ecology the general name for DR is ordination or indirect gradient analysis.) Today, hyperspectral satellite imagery collects so many bands that it is very difficult to analyze and interpret the data directly. Resuming the data into a set of few, yet independent, components is one way to reduce complexity (e.g., see Laparra et al. 2015). DR can also be used to visualize the interiors of deep neural networks (e.g., see Han et al. 2017), where the high dimensionality comes from the large number of weights used in a neural network and convergence can be visualized by means of DR. We could find many more example applications here but this is not the main focus of this publication.
The difficulty in applying DR is that each DR method is designed to maintain certain aspects of the original data and therefore may be appropriate for one task and inappropriate for another. Most methods also have parameters to tune and follow different assumptions. The quality of the outcome may strongly depend on their tuning, which adds additional complexity. DR methods can be modeled after physical models with attracting and repelling forces (Force Directed Methods), projections onto low dimensional planes (PCA, ICA), divergence of statistical distributions (SNE family), or the reconstruction of local spaces or points by their neighbors (LLE).
As an example for how changing internal parameters of a method can have
a great impact, the breakthrough for Stochastic Neighborhood Embedding
(SNE) methods came when a Student’s
There are a number of software packages for other languages providing collections of methods: In Python there is scikit-learn (Pedregosa et al. 2011), which contains a module for DR. In Julia we currently find ManifoldLearning.jl for nonlinear and MultivariateStats.jl for linear DR methods. There are several toolboxes for DR implemented in Matlab (Van Der Maaten et al. 2009; Arenas-Garcia et al. 2013). The Shogun toolbox (Sonnenburg et al. 2017) implements a variety of methods for dimensionality reduction in C++ and offers bindings for a many common high level languages (including R, but the installation is anything but simple, as there is no CRAN package). However, there is no comprehensive package for R and none of the former mentioned software packages provides means to consistently compare the quality of different methods for DR.
For many applications it can be difficult to objectively find the right method or parameterization for the DR task. This paper presents the dimRed and coRanking packages for the popular programming language R. Together, they provide a standardized interface to various dimensionality reduction methods and quality metrics for embeddings. They are implemented using the S4 class system of R, making the packages both easy to use and to extend.
The design goal for these packages is to enable researchers, who may not necessarily be experts in DR, to apply the methods in their own work and to objectively identify the most suitable methods for their data. This paper provides an overview of the methods collected in the packages and contains examples as to how to use the packages.
The notation in this paper will be as follows:
When referring to functions
in the dimRed package or base R
simply the function name is mentioned, functions from other packages are
referenced with their namespace, as with package::function
.
In the following section we do not aim for an exhaustive explanation to every method in dimRed but rather to provide a general idea on how the methods work. An overview and classification of the most commonly used DR methods can be found in Figure 1.
In all methods, parameters have to be optimized or decisions have to be made, even if it is just about the preprocessing steps of data. The dimRed package tries to make the optimization process for parameters as easy as possible, but, if possible, the parameter space should be narrowed down using prior knowledge. Often decisions can be made based on theoretical knowledge. For example, sometimes an analysis requires data to be kept in their original scales and sometimes this is exactly what has to be avoided as when comparing different physical units. Sometimes decisions based on the experience of others can be made, e.g., the Gaussian kernel is probably the most universal kernel and therefore should be tested first if there is a choice.
All methods presented here have the embedding dimensionality, ndim
as a parameter for embed
). For methods based on
eigenvector decomposition, the result generally does not depend on the
number of dimensions, i.e., the first dimension will be the same, no
matter if we decide to calculate only two dimensions or more. If more
dimensions are added, more information is maintained, the first
dimension is the most important and higher dimensions are successively
less important. This means, that a method based on eigenvalue
decomposition only has to be run once if one wishes to compare the
embedding in different dimensions. In optimization based methods this is
generally not the case, the number of dimensions has to be chosen a
priori, an embedding of 2 and 3 dimensions may vary significantly, and
there is no ordered importance of dimensions. This means that comparing
dimensions of optimization-based methods is computationally much more
expensive.
We try to give the computational complexity of the methods. Because of
the actual implementation, computation times may differ largely. R is an
interpreted language, so all parts of an algorithm that are implemented
in R often will tend to be slow compared to methods that call efficient
implementations in a compiled language. Methods where most of the
computing time is spent for eigenvalue decomposition do have very
efficient implementations as R uses optimized linear algebra libraries.
Although, eigenvalue decomposition itself does not scale very well in
naive implementations (
Principal Component Analysis (PCA) is the most basic technique for
reducing dimensions. It dates back to Pearson (1901). PCA finds a
linear projection (
PCA is a rotation around the origin and there exist a forward and inverse mapping. PCA may suffer from a scale problem, i.e., when one variable dominates the variance simply because it is in a higher scale, to remedy this, the data can be scaled to zero mean and unit variance, depending on the use case, if this is necessary or desired.
Base R implements PCA in the functions prcomp
and princomp
; but
several other implementations exist i.e.,
pcaMethods
from Bioconductor which implements versions of PCA that can deal with
missing data. The dimRed package wraps prcomp
.
Kernel Principal Component Analysis (kPCA) extends PCA to deal with
nonlinear dependencies among variables. The idea behind kPCA is to map
the data into a high dimensional space using a possibly non-linear
function
If the columns of X are centered around
Here is an example calculating the kernel matrix using a Gaussian
kernel:
The kPCA method is very flexible and there exist many kernels for
special purposes. The most common kernel function is the Gaussian kernel
(Equation (2)). The flexibility comes at the price that the
method has to be finely tuned for the data set because some parameter
combinations are simply unsuitable for certain data. The method is not
suitable for very large data sets, because memory scales with
Diffusion Maps, Isomap, Locally Linear Embedding, and some other techniques can be seen as special cases of kPCA. In which case, an out-of-sample extension using the formula can be applied (Bengio et al. 2004). This can also yield applications for bigger data, where an embedding is trained with a sub-sample of all data and then the data is embedded using the formula.
Kernel PCA in R is implemented in the
kernlab package
using the function kernlab::kpca
, and supports a number of kernels and
user defined functions. For details see the help page for
kernlab::kpca
.
The dimRed package wraps kernlab::kpca
but additionally provides
forward and inverse methods (Bakir et al. 2004) which can be used to
fit out-of sample data or to visualize the transformation of the data
space.
What today is called Classical Scaling was first introduced by
Torgerson (1952). It uses an eigenvalue decomposition of
a transformed distance matrix to find an embedding that maintains the
distances of the distance matrix. The method works because of the same
reason that kPCA works, i.e., classical scaling can be seen as a kPCA
with kernel
The disadvantage is that is computationally much more demanding,
i.e., an eigenvalue decomposition of a cmdscale
function.
The dimRed package wraps cmdscale
and allows the specification
of arbitrary distance functions for calculating the distance matrix.
Additionally a forward method is implemented.
As Classical Scaling can deal with arbitrarily defined distances,
Tenenbaum et al. (2000) suggested to approximate the structure of the
manifold by using geodesic distances. In practice, a graph is created by
either keeping only the connections between every point and its
Isomap’s computational cost is dominated by the eigenvalue decomposition
and therefore scales with
In R, Isomap is implemented in the
vegan package.
vegan::isomap
calculates an Isomap embedding and vegan::isomapdist
calculates a geodesic distance matrix. The dimRed package uses its
own implementation. This implementation is faster mainly due to using a
KD-tree for the nearest neighbor search (from the
RANN package) and to a
faster implementation for the shortest path search in the
Points that lie on a manifold in a high dimensional space can be
reconstructed through linear combinations of their neighborhoods if the
manifold is well sampled and the neighbohoods lie on a locally linear
patch. These reconstruction weights,
Conceptually the method is similar to Isomap but it is computationally
much nicer because the weight matrix is sparse and there exist efficient
solvers. In R, LLE is implemented by the package
lle, the embedding can
be calculated with lle::lle
. Unfortunately the implementation does not
make use of the sparsity of the weight matrix
Laplacian Eigenmaps were originally developed under the name spectral clustering to separate non-convex clusters. Later it was also used for graph embedding and DR (Belkin and Niyogi 2003).
A number of variants have been proposed. First, a graph is constructed,
usually from a distance matrix, the graph can be made sparse by keeping
only the
The dimRed package implements the algorithm from
Belkin and Niyogi (2003). Analogously to LLE, Laplacian eigenmaps avoid
computational complexity by creating a sparse matrix and not having to
estimate the distances between all pairs of points. Then the
eigenvectors corresponding to the lowest eigenvalues larger than
Diffusion Maps (Coifman and Lafon 2006) take a distance matrix as input
and calculates the transition probability matrix
The idea is to simulate a diffusion process between the nodes of the
graph, which is more robust to short-circuiting than the diffusionMap::diffuse()
function, which is
available in the
diffusionMap
package. Additional points can be approximated into an existing
embedding using the formula (Bengio et al. 2004). The implementation
in dimRed is based on the diffusionMap::diffuse
function.
While Classical Scaling and derived methods (see section
2.3) use eigenvector decomposition to embed the data
in such a way that the given distances are maintained, non-Metric
Dimensional Scaling (nMDS, Kruskal 1964a,b) uses
optimization methods to reach the same goal. Therefore a stress
function,
smacof
(Leeuw and Mair 2009).
Several packages provide implementations for nMDS in R, for example
MASS and vegan
with the functions MASS::isoMDS
and vegan::monoMDS
. Related methods
include Sammons Mapping which con be found as MASS::sammon
. The
dimRed package wraps vegan::monoMDS
.
The data
Graph embedding algorithms generally suffer from long running times (though compared to other methods presented here they do not scale as badly) and many local optima. This is why a number of methods have been developed that try to deal with some of the shortcomings, for example, the Kamada-Kawai (Kamada and Kawai 1989), the Fruchtermann-Reingold (Fruchterman and Reingold 1991), or the DrL (Martin et al. 2007) algorithms.
There are a number of graph embedding algorithms included in the
igraph package, they
can be accessed using the igraph::layout_with_*
function family. The
dimRed package only wraps the three algorithms mentioned above;
there are many others which are not interesting for dimensionality
reduction.
Stochastic Neighbor Embedding [SNE; Hinton and Roweis (2003)] is a
technique that minimizes the Kullback-Leibler divergence of scaled
similarities of the points
The
The general runtime of
There exist a number of derived techniques for dimensionality reduction, e.g., NeRV (Venna et al. 2010) and JNE (Lee et al. 2013), that improve results but for which there do not yet exist packages on CRAN implementing them.
Independent Component Analysis (ICA) interprets the data
There are a number of algorithms for ICA, the most widely used is
fastICA (Hyvarinen 1999) because it provides a fast and robust way
to estimate
There are a number of packages in R that implement algorithms for ICA,
the dimRed package wraps the fastICA::fastICA()
function from
fastICA.
Dimensionality Reduction via Regression is a very recent technique
extending PCA (Laparra et al. 2015). Starting from a rotated
(PCA) solution
The use of KRR also has the advantage of making the method convex, here we list it under non-convex methods, because other types of regression may make it non-convex.
Mathematicaly, functions are limited to map one input to a single output point. Therefore, DRR reduces to PCA if manifolds are too complex; but it seems very useful for slightly curved manifolds. The initial rotation is important, because the result strongly depends on the order of dimensions in high dimensional space.
DRR is implemented in the package DRR. The package provides forward and inverse functions which can be used to train on a subset.
The advantage of unsupervised learning is that one does not need to
specify classes or a target variable for the data under scrutiny.
Instead the chosen algorithm arranges the input data. For example,
arranged into clusters or into a lower dimensional representation. In
contrast to a supervised problem, there is no natural way to directly
measure the quality of any output or to compare two methods by an
objective measure like for instance modeling efficiency or
classification error. The reason is that every method optimizes a
different error function, and it would be unfair to compare
However, there are a series of independent estimators on the quality of a low-dimensional embedding. The dimRed package provides a number of quality measures which have been proposed in the literature to measure performance of dimensionality reduction techniques.
The co-ranking matrix (Lee and Verleysen 2009) is a way to capture the
changes in ordinal distance. As before, let
The co-ranking matrix
The co-ranking matrix can be computed using function
coRanking::coranking()
and can be visualized using
coRanking::imageplot()
. A good embedding should scatter the values
around the diagonal of the matrix. If the values are predominantly in
the lower triangle, then the embedding collapses the original structure
causing far away points to be much closer; if the values are
predominantly in the upper triangle the points from the original
structure are torn apart. Nevertheless this method requires visual
inspection of the matrix. For an automated assessment of quality, a
scalar value that assigns a quality to an embedding is needed.
A number of metrics can be computed from the co-ranking matrix. For example:
LCMC(
This measure is normalized to one and takes
In R, the co-ranking matrix can be calculated using the the
coRanking::coranking
function. The dimRed package contains the
functions Q_local
, Q_global
, Q_NX
, LCMC
, and R_NX
to calculate
the above quality measures in addition to AUC_lnK_R_NX
.
Calculating the co-ranking matrix is a relatively expensive operation
because it requires sorting every row of the distance matrix twice. It
therefore scales with plot_R_NX
, which plots the
There are a number of other measures that can be computed from a co-ranking matrix, e.g., see (Lee and Verleysen 2009; Lueks et al.; 2011), or Babaee et al. (2013).
An old measure originally developed to compare clustering methods in the
field of phylogenetics is cophenetic correlation
(Sokal and Rohlf 1962). This method consists simply of the correlation
between the upper or lower triangles of the distance matrices (in
dendrograms they are called cophenetic matrices, hence the name) in a
high and low dimensional space. Additionally the distance measure and
correlation method can be varied. In the dimRed package this is
implemented in the cophenetic_correlation
function.
Some studies use a measure called “residual variance”
(Tenenbaum et al. 2000; Mahecha et al. 2007), which is defined as
The fairest and most common way to assess the quality of a
dimensionality reduction when the method provides a inverse mapping is
the reconstruction error. The dimRed package includes a function
to calculate the root mean squared error which is defined as:
The dimRed package provides the reconstruction_rmse
and
reconstruction_error
functions.
There are a number of test data sets that are often used to showcase a dimensionality reduction technique. Common ones being the 3d S-curve and the Swiss roll, among others. These data sets have in common that they usually have three dimensions, and well defined manifolds. Real world examples usually have more dimensions and often are much noisier, the manifolds may not be well sampled and exhibit holes and large pieces may be missing. Additionally, we cannot be sure if we can observe all the relevant variables.
The dimRed package implements a number of test datasets that are
being used in literature to benchmark methods with the function
dimRed::loadDataSet()
. For artificial datasets the number of points
and the noise level can be adjusted, the function also returns the
internal coordinates.
The dimRed package collects DR methods readily implemented in R, implements missing methods and offers means to compare the quality of embeddings. The package is open source and available under the GPL3 license. Released versions of the package are available through CRAN (https://cran.r-project.org/package=dimRed) and development versions are hosted on GitHub (https://github.com/gdkrmr/dimRed). The dimRed package provides a common interface and convenience functions for a variety of different DR methods so that it is made easier to use and compare different methods. An overview of the packages main functions can be found in Table 1.
Function | Description |
---|---|
embed |
Embed data using a DR method. |
quality |
Calculate a quality score from the result of embed . |
plot |
Plot a "dimRedData" or "dimRedResult" object, colors the points automatically, for exploring the data. |
plot_R_NX |
Compares the quality of various embeddings. |
dimRedMethodList |
Returns a character vector that contains all implemented DR methods. |
dimRedQualityList |
Returns a character vector that contains all implemented quality measures. |
Internally, the package uses S4 classes but for normal usage the user
does not need to have any knowledge on the inner workings of the S4
class system in R (cf. table 2). The package contains
simple conversion functions from and to standard R-objects like a
data.frame or a matrix. The "dimRedData"
class provides an container
for the data to be processed. The slot data
contains a matrix with
dimensions in columns and observations in rows, the slot meta
may
contain a data frame with additional information, e.g., categories or
other information of the data points.
Class Name | Function |
---|---|
"dimRedData" |
Holds the data for a DR. Fed to embed() . An as.dimRedData() methods exists for "data.frame" , "matrix" , and "formula" exist. |
"dimRedMethod" |
Virtual class, ancestor of all DR methods. |
"dimRedResult" |
The result of embed() , the embedded data. |
Each embedding method is a class which inherits from "dimRedMethod"
which means that it contains a function to generate "dimRedResult"
objects and a list of standard parameters. The class "dimRedResult"
contains the data in reduced dimensions, the original meta information
along with the original data, and, if possible, functions for the
forward and inverse mapping.
From a user-perspective the central function of the package is embed
which is called in the form embed(data, method, …)
, data
can take
standard R objects such as instances of "data.frame"
, "matrix"
, or
"formula"
, as input. The method
is given as a character vector. All
available methods can be listed by calling dimRedMethodList()
.
Method-specific parameters can be passed through ...
; when no
method-specific parameters are given, defaults are chosen. The embed
function returns an object of class "dimRedResult"
.
For comparing different embeddings, dimRed contains the function
quality
which relies on the output of embed
and a method name. This
function returns a scalar quality score; a vector that contains the
names of all quality functions is returned by calling
dimRedQualityList()
.
For easy visual examination, the package contains plot
methods for
"dimRedData"
and "dimRedResult"
objects in order to plot high
dimensional data using parallel plots and pairwise scatter plots.
Automatic coloring of data points is done using the available metadata.
The comparison of different DR methods, choosing the right parameters for a method, and the inspection of the results is simplified by dimRed. This section contains a number of examples to highlight the use of the package.
To compare methods of dimensionality reduction, first a test data set is
loaded using loadDataSet
, then the embed
function is used for DR
(embed
can also handle standard R types like matrix
and
data.frame
). This makes it very simple to apply different methods of
DR to the same data e.g., by defining a character vector of method names
and then iterating over these, say with lapply
. For inspection,
dimRed provides methods for the plot
function to visualize the
resulting embedding (Figure 2 b and d), internal
coordinates of the manifold are represented by color gradients. To
visualize how well embeddings represent different neighborhood sizes,
the function plot_R_NX
is used on a list of embedding results
(Figure 2 c). The plots in
figure 2 are produced by the following code:
## define which methods to apply
embed_methods <- c("Isomap", "PCA")
## load test data set
data_set <- loadDataSet("3D S Curve", n = 1000)
## apply dimensionality reduction
data_emb <- lapply(embed_methods, function(x) embed(data_set, x))
names(data_emb) <- embed_methods
## figure \ref{fig:plotexample}a, the data set
plot(data_set, type = "3vars")
## figures \ref{fig:plotexample}b (Isomap) and \ref{fig:plotexample}d (PCA)
lapply(data_emb, plot, type = "2vars")
## figure \ref{fig:plotexample}c, quality analysis
plot_R_NX(data_emb)
The function plot_R_NX
produces a figure that plots the neighborhood
size (
Therefore we can see from Figure 2c that
Often the quality of an embedding strongly depends on the choice of parameters, the interface of dimRed can be used to facilitate searching the parameter space.
Isomap has one parameter
## Load data
ss <- loadDataSet("3D S Curve", n = 500)
## Parameter space
kk <- floor(seq(5, 100, length.out = 40))
## Embedding over parameter space
emb <- lapply(kk, function(x) embed(ss, "Isomap", knn = x))
## Quality over embeddings
qual <- sapply(emb, function(x) quality(x, "Q_local"))
## Find best value for K
ind_max <- which.max(qual)
k_max <- kk[ind_max]
Figure 3a shows how the
It is also very easy to compare across methods and quality scores. The
following code produces a matrix of quality scores and methods, where
dimRedMethodList
returns a character vector with all methods. A
visualization of the matrix can be found in
Figure 4.
quality_results
matrix. The methods are ordered by mean
quality score. The reconstruction error was omitted, because a higher
value means a worse embedding, while in the present methods a higher
score means a better embedding. Parameters were not tuned for the
example, therefore it should not be seen as a general quality assessment
of the methods.
embed_methods <- dimRedMethodList()
quality_methods <- c("Q_local", "Q_global", "AUC_lnK_R_NX",
"cophenetic_correlation")
scurve <- loadDataSet("3D S Curve", n = 2000)
quality_results <- matrix(
NA, length(embed_methods), length(quality_methods),
dimnames = list(embed_methods, quality_methods)
)
embedded_data <- list()
for (e in embed_methods) \{
embedded_data[[e]] <- embed(scurve, e)
for (q in quality_methods)
try(quality_results[e, q] <- quality(embedded_data[[e]], q))
\}
This example showcases the simplicity with which different methods and quality criteria can be combined. Because of the strong dependencies on parameters it is not advised to apply this kind of analysis without tuning the parameters for each method separately. There is no automatized way to tune parameters in dimRed.
This paper presents the dimRed and coRanking packages and it provides a brief overview of the methods implemented therein. The dimRed package is written in the R language, one of the most popular languages for data analysis. The package is freely available from CRAN. The package is object oriented and completely open source and therefore easily available and extensible. Although most of the DR methods already had implementations in R, dimRed adds some new methods for dimensionality reduction, and coRanking adds methods for an independent quality control of DR methods to the R ecosystem. DR is a widely used technique. However, due to the lack of easily usable tools, choosing the right method for DR is complex and depends upon a variety of factors. The dimRed package aims to facilitate experimentation with different techniques, parameters, and quality measures so that choosing the right method becomes easier. The dimRed package wants to enable the user to objectively compare methods that rely on very different algorithmic approaches. It makes the life of the programmer easier, because all methods are aggregated in one place and there is a single interface and standardized classes to access the functionality.
We thank Dr. G. Camps-Valls and an anonymous reviewer for many useful comments. This study was supported by the European Space Agency (ESA) via the Earth System Data Lab project (http://earthsystemdatacube.org) and the EU via the H2020 project BACI, grant agreement No 640176.
dimRed, coRanking, kernlab, vegan, RANN, igraph, lle, diffusionMap, MASS, Rtsne, fastICA, DRR
ChemPhys, Cluster, Distributions, Econometrics, Environmetrics, GraphicalModels, MachineLearning, MixedModels, NaturalLanguageProcessing, NumericalMathematics, Optimization, Phylogenetics, Psychometrics, Robust, Spatial, TeachingStatistics
This article is converted from a Legacy LaTeX article using the texor package. The pdf version is the official version. To report a problem with the html, refer to CONTRIBUTE on the R Journal homepage.
Text and figures are licensed under Creative Commons Attribution CC BY 4.0. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".
For attribution, please cite this work as
Kraemer, et al., "dimRed and coRanking---Unifying Dimensionality Reduction in R", The R Journal, 2018
BibTeX citation
@article{RJ-2018-039, author = {Kraemer, Guido and Reichstein, Markus and D. Mahecha, Miguel}, title = {dimRed and coRanking---Unifying Dimensionality Reduction in R}, journal = {The R Journal}, year = {2018}, note = {https://rjournal.github.io/}, volume = {10}, issue = {1}, issn = {2073-4859}, pages = {342-358} }