Christian Battaglia
September 4, 2019
7 min read
Discrete data, as the name suggests, can take only specified values. For example, when you roll a die, the possible outcomes are 1, 2, 3, 4, 5 or 6 and not 1.5 or 2.45.
Continuous data can take any value within a given range. The range may be finite or infinite. For example, A girl’s weight or height, the length of the road. The weight of a girl can be any value from 54 kgs, or 54.5 kgs, or 54.5436kgs.
A Bernoulli distribution has only two possible outcomes, namely 1 (success) and 0 (failure), and a single trial.
When you roll a fair die, the outcomes are 1 to 6. The probabilities of getting these outcomes are equally likely and that is the basis of a uniform distribution. Unlike Bernoulli Distribution, all the n number of possible outcomes of a uniform distribution are equally likely.
Uniform distribution if:
for

A binomial distribution is the sum of independent and identically distributed Bernoulli random variables. Think of it as running a Bernoulli trial (n) times and counting the successes — like flipping a coin 10 times and asking "how many heads?"
The bell curve. The most important distribution in statistics thanks to the Central Limit Theorem: the sum of many independent random variables tends toward a normal distribution regardless of their individual distributions. Defined by its mean (\mu) and standard deviation (\sigma).
Models the number of events occurring in a fixed interval of time or space — like the number of emails you receive per hour or the number of typos per page. The key parameter is (\lambda), the average rate of occurrence.

Bayesian probability is an interpretation of the concept of probability, in which, instead of frequency or propensity of some phenomenon, probability is interpreted as reasonable expectation representing a state of knowledge or as quantification of a personal belief.
an extension of propositional logic that enables reasoning with hypotheses
Quantum Bayesianism (abbreviated QBism, pronounced "cubism") is an interpretation of quantum mechanics that takes an agent's actions and experiences.
This interpretation is distinguished by its use of a subjective Bayesian account of probabilities to understand the quantum mechanical Born rule as a normative addition to good decision-making.
The Born rule (also called the Born law, Born's rule, or Born's law), formulated by German physicist Max Born in 1926, is a physical law[citation needed] of quantum mechanics giving the probability that a measurement on a quantum system will yield a given result. In its simplest form it states that the probability density of finding the particle at a given point is proportional to the square of the magnitude of the particle's wavefunction at that point. The Born rule is one of the key principles of quantum mechanics.
The old quantum theory is a collection of results from the years 1900–1925 which predate modern quantum mechanics. The theory was never complete or self-consistent, but was rather a set of heuristic corrections to classical mechanics. The theory is now understood as the semi-classical approximation to modern quantum mechanics.
Where the are the momenta of the system and the are the corresponding coordinates. The quantum numbers are integers and the integral is taken over one period of the motion at constant energy (as described by the Hamiltonian). The integral is an area in phase space, which is a quantity called the action and is quantized in units of Planck's (unreduced) constant. For this reason, Planck's constant was often called the quantum of action.
events that led to quantum theory:
biggest concepts:
![]()
The mathematics of probability can be developed on an entirely axiomatic basis that is independent of any interpretation:
first:
probability of an event is a non-negative real number:
where is the event space. It follows that is always finite, in contrast with more general measure theory. Theories which assign negative probability relax the first axiom.
second:
assumption of unit measure: that the probability that at least one of the elementary events in the entire sample space will occur is 1
third:
Any countable sequence of disjoint sets (synonymous with mutually exclusive events) satisfies
The probability of the empty set.
In some cases, is not the only event with probability .
Monotonicity
.
If A is a subset of, or equal to B, then the probability of A is less than, or equal to the probability of B.
numeric bound
It immediately follows from the monotonicity property that
Kolmogorov's axioms imply that:
The probability of neither heads nor tails, is 0.
The probability of either heads or tails, is 1.