X-Git-Url: https://git.auder.net/?p=valse.git;a=blobdiff_plain;f=pkg%2FR%2FEMGLLF.R;h=1633821a344996c75aec2c85a5371d8c80d08423;hp=e393ec8633ddfb711f382c8188d39f46df467ee2;hb=3921ba9b5ea85bcc190245ac7da9ee9da1658b9f;hpb=e9db79707709c10947e89756eb5655c0747a2a1d diff --git a/pkg/R/EMGLLF.R b/pkg/R/EMGLLF.R index e393ec8..1633821 100644 --- a/pkg/R/EMGLLF.R +++ b/pkg/R/EMGLLF.R @@ -1,6 +1,6 @@ #' EMGLLF #' -#' Run a generalized EM algorithm developped for mixture of Gaussian regression +#' Run a generalized EM algorithm developped for mixture of Gaussian regression #' models with variable selection by an extension of the Lasso estimator (regularization parameter lambda). #' Reparametrization is done to ensure invariance by homothetic transformation. #' It returns a collection of models, varying the number of clusters and the sparsity in the regression mean. @@ -16,6 +16,7 @@ #' @param X matrix of covariates (of size n*p) #' @param Y matrix of responses (of size n*m) #' @param eps real, threshold to say the EM algorithm converges, by default = 1e-4 +#' @param fast boolean to enable or not the C function call #' #' @return A list (corresponding to the model collection) defined by (phi,rho,pi,LLF,S,affec): #' phi : regression mean for each cluster @@ -37,14 +38,8 @@ EMGLLF <- function(phiInit, rhoInit, piInit, gamInit, mini, maxi, gamma, lambda, } # Function in C - n <- nrow(X) #nombre d'echantillons - p <- ncol(X) #nombre de covariables - m <- ncol(Y) #taille de Y (multivarié) - k <- length(piInit) #nombre de composantes dans le mélange .Call("EMGLLF", phiInit, rhoInit, piInit, gamInit, mini, maxi, gamma, lambda, - X, Y, eps, phi = double(p * m * k), rho = double(m * m * k), pi = double(k), - llh = double(1), S = double(p * m * k), affec = integer(n), n, p, m, k, - PACKAGE = "valse") + X, Y, eps, PACKAGE = "valse") } # R version - slow but easy to read