Populations and Samples

The main difference between a population and a sample has to do with how
observations are assigned to the data set

Population Includes all of the elements from a set of data.

Sample Consists of one or more observations from the population.

Image result for population and sample

Probability Density function(PDF)

The PDF, or density of a continuous random variable, is a function that describes
the relative likelihood of a random variable X to take on a given value x.

In the mathematical fields of probability and statistics, a random variate x is a particular
outcome of a random variable X: the random variates which are other outcomes of
the same random variable might have different values.

The PDF also defines the expected value E[X] of a continuous distribution of X.

The expected value is a function of the probability distribution of the observed
value in our population. Moreover The sample mean of our sample is the observed mean value of our data.

Further If the experiment has been designed correctly, the sample mean
should converge to the expected value as more and more samples are included in the
analysis.

 

Central Limit Theorem

In probability theory, the central limit theorem (CLT) establishes that, in some situations, when independent random variables are added, their properly normalized sum tends toward a normal distribution (informally a "bell curve") even if the original variables themselves are not normally distributed.

The theorem is a key ("central") concept in probability theory because it implies that probabilistic and statistical methods that work for normal distributions can be applicable to many problems involving other types of distributions.

For example, suppose that a sample is obtained containing a large number of observations, each observation being randomly generated in a way that does not depend on the values of the other observations, and that the arithmetic average of the observed values is computed.

If this procedure is performed many times, the central limit theorem says that the distribution of the average will be closely approximated by a normal distribution.

A simple example of this is that if one flips a coin many times the probability of getting a given number of heads in a series of flips will approach a normal curve, with mean equal to half the total number of flips in each series. (In the limit of an infinite number of flips, it will equal a normal curve.)

Related image

Chebyshev's inequality

In probability theory, Chebyshev's inequality (also called the Bienaymé-Chebyshev inequality) guarantees that, for a wide class of probability distributions, no more than a certain fraction of values can be more than a certain distance from the mean.

Specifically, no more than 1/k2 of the distribution's values can be more than k standard deviations away from the mean (or equivalently, at least 1−1/(k^2) of the distribution's values are within k standard deviations of the mean). The rule is often called Chebyshev's theorem, about the range of standard deviations around the mean, in statistics.

The inequality has great utility because it can be applied to any probability distribution in which the mean and variance are defined.

Note the only requirement for applying Chebyshev's ineqaulity is that it has fixed mean.

 

Image result for Chebyshev's inequality gif


0 Comments

What do you think?