- Applied probability
- >
- Expected utility
- >
- Optimal decisions
- >
- Loss functions

- Decision theory
- >
- Expected utility
- >
- Optimal decisions
- >
- Loss functions

- Inference
- >
- Statistical inference
- >
- Model selection
- >
- Loss functions

- Mathematical modeling
- >
- Statistical models
- >
- Model selection
- >
- Loss functions

- Statistical analysis
- >
- Statistical inference
- >
- Model selection
- >
- Loss functions

- Statistical theory
- >
- Statistical inference
- >
- Model selection
- >
- Loss functions

- Statistical theory
- >
- Statistical models
- >
- Model selection
- >
- Loss functions

Taguchi loss function

The Taguchi loss function is graphical depiction of loss developed by the Japanese business statistician Genichi Taguchi to describe a phenomenon affecting the value of products produced by a company.

Cross entropy

In information theory, the cross-entropy between two probability distributions and over the same underlying set of events measures the average number of bits needed to identify an event drawn from the

Loss function

In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a

Squared error loss

No description available.

Hinge loss

In machine learning, the hinge loss is a loss function used for training classifiers. The hinge loss is used for "maximum-margin" classification, most notably for support vector machines (SVMs). For a

Mean squared prediction error

In statistics the mean squared prediction error or mean squared error of the predictions of a smoothing or curve fitting procedure is the expected value of the squared difference between the fitted va

Mean squared error

In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—tha

Sum of absolute differences

In digital image processing, the sum of absolute differences (SAD) is a measure of the similarity between image blocks. It is calculated by taking the absolute difference between each pixel in the ori

Huber loss

In statistics, the Huber loss is a loss function used in robust regression, that is less sensitive to outliers in data than the squared error loss. A variant for classification is also sometimes used.

Sum of absolute transformed differences

The sum of absolute transformed differences (SATD) is a block matching criterion widely used in fractional motion estimation for video compression. It works by taking a frequency transform, usually a

© 2023 Useful Links.