The load function "expects" to receive a function that has two arguments: the first is data.frame, and the second is a vector of "indices" (possibly with repeating elements and probably not using all indices) for use in selecting strings and, probably with multiple duplicates or triples). Then these are samples with a replacement determined by the pattern of duplicates and three repetitions from the original data frame (several times defined by "R" with different "sets of choices"), passes the argument to the boot.function functions to their indices, and then collects the results of the number of functional applications R.
Regarding what the print
method is told for bootable objects, look at this (after checking the returned object with str ()
> duncan.boot$t0 [1] 5.003310e+00 1.053184e-05 > apply(duncan.boot$t, 2, mean) [1] 5.342895220 0.002607943 > apply(duncan.boot$t, 2, mean) - duncan.boot$t0 [1] 0.339585441 0.002597411
It becomes more obvious that the value of T0
comes from the original data, and the offset is the difference between the average value of the boot () and the values of T0
. I don’t think it makes sense to ask a question about why p-values based on parametric considerations grow due to an increase in estimated t-statistics. When you do this, you really are in two separate areas of statistical thought. I would interpret the increase in p-values as an effect of the sampling process that did not take into account the assumption of a normal distribution. It just says something about the distribution of the p-value sample (which is really just another sample statistic).
(Comment: The source code used during the development of R was Davison and Hinckley “Bootstrap Methods and Their Applications.” I do not claim that I support my answer above, but I decided to put it as a reference after Hagen Brenner asked about the sample with with the two indexes in the comments below: There are many unexpected aspects of self-tuning that come up after you go beyond a simple parametric estimation, and I would go to this link first if I were to deal with more complex sampling situations.)
source share