clusterboot
(as parameter
clustermethod
; "CBI" stands for "clusterboot interface").
In some situations it could make sense to use them to compute a
clustering even if you don't want to run clusterboot
, because
some of the functions contain some additional features (e.g., normal
mixture model based clustering of dissimilarity matrices projected
into the Euclidean space by MDS or partitioning around medoids with
estimated number of clusters, noise/outlier identification in
hierarchical clustering).kmeansCBI(data,krange,k,scaling=FALSE,runs=1,criterion="ch",...)hclustCBI(data,k,cut="number",method,scaling=TRUE,noisecut=0,...)
hclusttreeCBI(data,minlevel=2,method,scaling=TRUE,...)
disthclustCBI(dmatrix,k,cut="number",method,noisecut=0,...)
noisemclustCBI(data,G,k,emModelNames,nnk,hcmodel=NULL,Vinv=NULL,
summary.out=FALSE,...)
distnoisemclustCBI(dmatrix,G,k,emModelNames,nnk,
hcmodel=NULL,Vinv=NULL,mdsmethod="classical",
mdsdim=4, summary.out=FALSE, points.out=FALSE,...)
claraCBI(data,k,usepam=TRUE,diss=inherits(data,"dist"),...)
pamkCBI(data,krange=2:10,k=NULL,criterion="asw", usepam=TRUE,
scaling=TRUE,diss=inherits(data,"dist"),...)
trimkmeansCBI(data,k,scaling=TRUE,trim=0.1,...)
tclustCBI(data,k,trim=0.05,...)
disttrimkmeansCBI(dmatrix,k,scaling=TRUE,trim=0.1,
mdsmethod="classical",
mdsdim=4,...)
dbscanCBI(data,eps,MinPts,diss=inherits(data,"dist"),...)
mahalCBI(data,clustercut=0.5,...)
mergenormCBI(data, G=NULL, k=NULL, emModelNames=NULL, nnk=0,
hcmodel = NULL,
Vinv = NULL, mergemethod="bhat",
cutoff=0.1,...)
speccCBI(data,k,...)
claraCBI
,
pamkCBI
and dbscanCBI
work with an
n*n
-dissimilarity matrix as well, see parameter diss
.dist
-object.hclustCBI
and disthclustCBI
see parameter cut
below. Some
methods have a k
parametscaling
is a numeric
vector with length equal to the number of variables, then each
variable is divided by the corresponding value from "ch"
or "asw"
. Decides whether number
of clusters is estimated by the Calinski-Harabasz criterion or by the
average silhouette width.cutree
is used to obtain a partition from a hierarchy
tree. cut="level"
means that the tree is cut at a particular
dissimilarity level, cut="number"
means thclust
.<=noisecut< code=""> in the
disthclustCBI
/hclustCBI
-partition are joined and declared as
noise/outliers.=noisecut<>
minlevel=1
means that all clusters in
the tree are given out by hclusttreeCBI
or
disthclusttreeCBI
, including one-point
clusters (but excluding the cluster with all
points). minlevel=2<
mclustBIC
. If
G
has more than one entry, the number of clusters is
estimated by the BIC.mclustBIC
.NNclean
, which is used to estimate the
initial noise for noisemclustCBI
and
distnoisemclustCBI
. See parameter k
in themclustBIC
.TRUE
, the result of
summary.mclustBIC
is added as component
mclustsummary
to the output of noisemclustCBI
and
distnoisemc
TRUE
, the matrix of MDS points
is added as component
points
to the output of noisemclustCBI
.TRUE
, data
will be considered as
a dissimilarity matrix. In claraCBI
, this requires
usepam=TRUE
.tclust
,
trimkmeans
.dbscan
.dbscan
.fixmahal
is used for fuzzy clustering, a crisp partition is generated and
points with cluster membership values above clustercut
are
considered as membemethod
to mergenormals
.mergenormals
.summary.out
and points.out
above):nc
includes the
noise component, and there should be another component
nccl
, being the number of clusters not including the
noise component.n
) for each cluster,
indicating whether a point is a member of this cluster
(TRUE
) or not. If a noise component is included, it
should always be the last vector in this list.n
,
partitioning the data. If the method produces a partition, it
should be the clustering. This component is only used for plots,
so you could do something like rep(1,n)
for
non-partitioning methods.nc
above.noisemclustCBI
and distnoisemclustCBI
,
see above.NNclean
, called by noisemclustCBI
and distnoisemclustCBI
.TRUE
if points were classified as
noise/outliers by disthclustCBI
.clusterboot
. Here is a brief overview. For further
details see the help pages of the involved clustering methods.
[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]clusterboot
, dist
,
kmeans
, kmeansruns
, hclust
,
mclustBIC
,
pam
, pamk
,
clara
,
trimkmeans
, dbscan
,
fixmahal
set.seed(20000)
face <- rFace(50,dMoNo=2,dNoEy=0,p=2)
dbs <- dbscanCBI(face,eps=1.5,MinPts=4)
dhc <- disthclustCBI(dist(face),method="average",k=1.5,noisecut=2)
table(dbs$partition,dhc$partition)
mergenormCBI(face,G=10,emModelNames="EEE",nnk=2)
Run the code above in your browser using DataLab