MOC - Introduction to Statistics and Statistical Programming

Scale of measure

A finite data sequence with real data has values (unique values, sometimes called Item Expression) where because the same value can be measured multiple times.

We interpret Stochastic Independence as causal independence, which means that we will look at the progression of data sequences to determine if they are independent of another.

Use Dichotomisation to make data usable for Runs test for long 01-sequence which determines whether a data sequence is random or not.

We want to be able to infer the underlying random mechanism of a data sequence. Ordered Data Sequences makes it a bit easier as the order usually doesn’t play a role.

We can use a Barplot or Histogram to plot frequencies of items. We can describe the probability for every interval in an ECDF.

CDF to PDF relation.

→ Infer it from the data (Calculate every p-quantile)

Chapter 3 - Random Variables

For real random Random Variables we mostly have events as intervals or sets (in this case they are generated by repeated application of unions and complementations of intervals). The set of all these intervals is called the Borel Sigma Algebra of on which we can define the Probability Measure which goes to from an event out of the Borel Sigma Algebra.

Identity Map The probability of the identity map for is .

Linear Transformation where ECDF: Let there be a data sequence with Item Expression (ordered values) and frequencies .

p-Value Statistical Simulation

Chapter 5 - Estimators

We would like to extend CDFs to vector valued items. That way we can for example repeat the same experiment multiple times.

So we describe such a repeated experiment with a Random Variable where every is a specific observation of that experiment.

We can define the CDF like this:

If all random variables are independent we can write: This also implies which implies uncorrelatedness.

We can use Standard Estimators to estimate Expectation, Variance and CDF from a data sample.

Chapter 6 - Gauß Test

General Confidence Intervals for Gauß Tests

We have with probability :

Chapter 11 - Distribution Parameters cont’d

Plug-In-Method Maximum Likelihood Estimation

Chapter 12 - Curve Fitting

  • We have a data sequence and want to find a function that best describes it.
  • “Best describing” can be quantified using an error metric.

We now want to find or choose a function that minimizes this error. This method is also known as the Method of Least Deviation. This method doesn’t necessarily result in a unique minimal error (if at all minimal error) as often only numerical solutions exist.

This method leads to Least Squares Method which is a popular way of solving the described problem.

is also called the regression function.

The Residuum can be used to study the underling random mechanisms.

Definition of the problem We can choose a function from a set of all possible polynomials.

Least Squares This method uses a Linear Model and the least squares error function.

See Mean Squared Error for more details.

Solutions depend on the complexity of the model which is mostly determined by the dimensions of :

Linear Gaussian Model

Geometric View

From the geometric view, the least squares method can be understood as an orthogonal projection

Simple Linear Regression

The method is often numerically stable, even for non-linear models.

Chapter 17 - Limit Theorems and Simulations

In this first subchapter we learn about the Strong Law of Large Numbers which we can use to define the Glivenko-Cantelli theorem.

The fundamental theorems become important when we want to talk about Kolmogorov in the next subchapter.

In this subchapter we define a general goodness of fit test that can test wether the unknown Distribution of a data sequence euqals some distribution .

We can use this general approach to definde the Kolmogorov Smirnov Test which uses the Kolmogorov Distribution for its Test Statistic.

Furthermore we can use the Lilliefors Test of Normality to determine whether a data sequence from an unknown distribution falls into the class of normal distributions,