211k views
0 votes
A tester sets 99% confidence limits on a certain statistic. The absolute difference between either limit and the mean is an example of:

2 Answers

3 votes
I found the missing choices:
the standard deviation
the margin of error
the variance
the population mean

The absolute difference between either limit and the mean is an example of THE VARIANCE.

Variance is defined as the
average of the squared differences from the Mean.
User Dustin Nielson
by
8.6k points
5 votes

Answer:

Margin of error.

Explanation:

Given that a tester sets 99% confiidence limits on a certain statistic

A confidence limit for 99% is obtained as

lower bound of confidence interval as Mean - critical 99% *std error and

upper bound of confidence interval as Mean + critical 99% * std error

Thus a confidence interval would be in interval form with lower limit and upper limit and the average of these two limits is the mean.

The absolute difference between either limit and the mean is the value

Critical value for 99% multiplied by Std error of the variable

= Margin of error

Hence answer is margin of error

User Andiana
by
8.0k points