Theoretical Distribution
Theoretical Distribution
Theoretical Distribution
In the real world, experiments with single outcomes like heads or tails are rare; generally, experiments
are collection of set of events and carried by n number of times, the of outcome which, represents a
form of theoretical distribution. A random exponent assumed in theoretical distribution, given by a
function of the random variable, called probability function.
For example, if we toss a fair coin, the probability of getting a head is 12. If we toss it for 50 times,
the probability of getting a head is 25. We call this as the theoretical or expected frequency of the
heads. Actually, by tossing a coin, we may get 25, 30 or 35 heads, which we call as the observed
frequency.
Thus, the observed frequency and the expected frequency may equal or may differ from each other
due to fluctuation in the experiment.
1. Binomial Distribution
2. Poisson distribution
Binomial Distribution:
The prefix ‘Bi’ means two or twice. A binomial distribution is the probability of a trail with two and only
two outcomes. It is a type of distribution that has two different outcomes namely, ‘success’ and
‘failure’. In addition, it is applicable to discrete random variables only. In binomial probability
distribution, the number of ‘Success’ in a sequence of n experiments, where each time a question
is asked for yes-no, then the Boolean-valued outcome is represented either with
success/yes/true/one (probability p) or failure/no/false/zero (probability q = 1 − p). A single
success/failure test called a Bernoulli trial or Bernoulli experiment, and a series of outcomes called
a Bernoulli process. For n = 1, i.e. a single experiment, the binomial distribution is a Bernoulli
distribution. The binomial distribution is the base for the famous binomial test of statistical
importance.
Or
P(x:n,p) = nCx px (q)n-x
Where,
n = the number of experiments
x = 0, 1, 2, 3, 4, …
p = Probability of Success in a single experiment
q = Probability of Failure in a single experiment = 1 – p
The binomial distribution formula can also be written in the form of n-Bernoulli trials, where nCx =
n!/x!(n-x)!. Hence,
P(x:n,p) = n!/[x!(n-x)!].px.(q)n-x
There are two possible outcomes: true or false, success or failure, yes or no.
There is ‘n’ number of independent trials or a fixed number of n times repeated trials.
The probability of success or failure varies for each trial.
Only the number of success calculated, out of ‘n’ independent trials.
Every trial is an independent trial, which means the outcome of one trial does not affect
the outcome of another trial.
Finding the quantity of raw and used materials while making a product.
Taking a survey of positive and negative reviews from the public for any specific product
or place.
By using the YES/ NO survey, we can check whether the number of persons views the
particular channel.
To find the number of male and female employees in an organization.
The number of votes collected by a candidate in an election is counted based on 0 or 1
probability.
At first, we divide the time into n number of small intervals, such that n → ∞ and p denote the
probability of success, as we have already divided the time into infinitely small intervals so p → 0. So
the result must be that in that condition is n x p = λ (a finite constant).
A Poisson random variable “x” defines the number of successes in the experiment. This
distribution occurs when there are events that do not occur as the outcomes of a definite
number of outcomes. Poisson distribution is used under certain conditions. They are:
The probability of more than one success in unit time is very low.
Applications
Number of Network Failures per Week.
Number of Bankruptcies Filed per Month.
Number of Website Visitors per Hour.
Number of Arrivals at a Restaurant.
Number of Calls per Hour at a Call Center.
Number of Books Sold per Week.
Average Number of Storms in a City.
Normal Distribution:
The Normal Distribution defines a probability density function f(x) for the continuous random
variable X considered in the system. The random variables, which follow the normal distribution, are
ones whose values can assume any known value in a given range.
We can hence extend the range to – ∞ to + ∞ . Continuous Variables are such random variables
and thus, the Normal Distribution gives you the probability of your value being in a particular range
for a given trial. The normal distribution is very important in the statistical analysis due to the central
limit theorem.
The theorem states that any distribution become normally distributed when the number of variables
is sufficiently large. For instance, the binomial distribution tends to change into the normal distribution
with mean and variance.
In probability theory and statistics, the Normal Distribution, also called the Gaussian
Distribution, is the most significant continuous probability distribution. Sometimes it is also called
a bell curve. A large number of random variables are either nearly or exactly represented by the
normal distribution, in every physical science and economics. Furthermore, it can
be used to approximate other probability distributions, therefore supporting the usage of the word
‘normal ‘as in about the one, mostly used.
Where,
x is the variable
μ is the mean
σ is the standard deviation
Approximately 68% of the data falls within one standard deviation of the mean. (i.e.,
Between Mean- one Standard Deviation and Mean + one standard deviation)
Approximately 95% of the data falls within two standard deviations of the mean. (i.e.,
Between Mean- two Standard Deviation and Mean + two standard deviations)
Approximately 99.7% of the data fall within three standard deviations of the mean. (i.e.,
Between Mean- three Standard Deviation and Mean + three standard deviations)
In a normal distribution, the mean, mean and mode are equal.(i.e., Mean = Median=
Mode).
The total area under the curve should be equal to 1.
The normally distributed curve should be symmetric at the centre.
There should be exactly half of the values are to the right of the centre and exactly half
of the values are to the left of the centre.
The normal distribution is, defined by the mean and standard deviation.
The mean and median are the same and lie in the middle of the distribution
The normal distribution curve must have only one peak. (i.e., Unimodal)
The curve approaches the x-axis, but it never touches, and it extends farther away from
the mean.
About 68 percent of its values lie within one standard deviation of the mean.
About 95 percent of its values lie within two standard deviations of the mean.
Almost all of its values lie within three standard deviations of the mean.
Applications
The normal distributions are closely associated with many things such as: