
Last chance! 50% off unlimited learning
Sale ends in
Last chance! 50% off unlimited learning
Sale ends in
sm.regression(x, y, h, design.mat = NA, model = "none", test = TRUE,
weights = rep(1, nobs), ...)
h
is its standard deviation.y
when these are assumed to be the
residuals from a linear model."none"
, "no effect"
and "linear"
are possible.sm.options
function, through
a mechanism which limits their effect only to this call of the function;
those relevant for this function are the following:"none"
will prevent any graphical output from being produced.
With one covariate, the default setting "lines"
will produce the density
estimate, while the setting "se"
will in addition produce a variabiliF
, which creates a new plot.
This argument applies only with one covariate.ngrid
refers to the number of points
along the axis in each dimension.
Default: 50 and 20 for 1-, 2-dimensional data, respectively.FALSE
, evaluation and plotting take place at all grid
points where eye
in the persp
function. Here the argument
is automatically scatest
set to T
, then the p-value of the test
is also returned. When there is only one covariate, the weights associated
with different obserations, an estimate of the error standard deviation and
the standard error of the estimate are also returned. If a reference model
has been specified, this standard error refers to the comparison between
the estimate and the reference model, and the values defining the reference
model are also returned.display="none"
.hcv
, sm
, sm.ancova
, sm.binomial
, sm.poisson
, sm.regression.autocor
,
sm.survival
, sm.options
# An example with one covariate
x <- runif(100,-2, 2)
y <- x^2 + rnorm(50)
sm.regression(x, y, h=0.5)
# An example with one covariate
x <- cbind(runif(100,-2, 2), runif(100,-2, 2))
y <- x[,1]^2 + x[,2]^2 + rnorm(50)
sm.regression(x, y, h=c(1,1))
Run the code above in your browser using DataLab