One Piece of Math a Day

This is a mirror of the daily Samuel Vaiter's tweet series available here. One piece of math = one slide = one tweet every (working) day. All content is available under the CC BY-SA license.

October 2024

Johnson–Lindenstrauss lemma (1984)

2024-10-07

The Johnson–Lindenstrauss Lemma states that a set of high-dimensional points can be mapped into a much lower-dimensional space while approximately preserving pairwise distances. This is useful for dimensionality reduction, clustering, etc. [ref]

thumbnail

Stein's lemma

2024-10-04

Stein's Lemma states that for a normally distributed variable X, the expected value E[Xg(X)] = E[g’(X)] for any g absolutely continuous (derivative a.e.) such that E[|g’(X)|] < ∞. It is a central result for characterizing (close-to) Gaussian data [ref]

Twitter Post

thumbnail

Erdős–Rényi(–Gilbert) graph models

2024-10-03

Erdős–Rényi-Gilbert graph models are two random models introduced in 1959 sharing a lot of common properties. In particular, G(n, p) exhibits a sharp threshold for the connectedness for p ≈ n/ln n. [ref]

Twitter Post

GitHub File

thumbnail

Law(s) of Large Numbers

2024-10-02

The law of large numbers tells that the empirical mean of an integrable random sample converges towards its mean in probability (weak LLN) and almost surely (strong LLN). [ref]

Twitter Post

GitHub File

thumbnail

Tarski—Seidenberg: logical statement

2024-10-01

The Tarski-Seidenberg theorem in logical form states that the set of first-order formulas over the real numbers is closed under quantifier elimination. This means any formula with quantifiers can be converted into an equivalent quantifier-free formula. [ref]

Twitter Post

thumbnail

September 2024

Davis—Kahan sin(θ) inequality

2024-09-30

The Davis-Kahan sin(θ) theorem bounds the difference between the subspaces spanned by eigenvectors of two symmetric matrices, based on the difference between the matrices. It quantifies how small perturbations in a matrix affect its eigenvectors. [ref]

Twitter Post

thumbnail

Grid search

2024-09-27

Grid search is the most popular method for hyperparameter optimization in machine learning. Using a performance metric, it simply aims to “exhaustively” evaluate this metric on a discretisation of the sample space. It suffers from the curse of dimensionality.

Twitter Post

thumbnail

Determinant vs Permanent

2024-09-26

The determinant and permanent of a matrix may look similar, but they differ fundamentally: the determinant includes the signatures of permutations, while the permanent uses only addition. One is easy to compute, the others expected to be very hard. [ref]

Twitter Post

thumbnail

Chordal slope lemma

2024-09-25

If you pick three points along the curve in increasing order, the slope of the line between the first two points will always be less than or equal to the slope of the line between the last two points (chordal slope's lemma)

Twitter Post

thumbnail

Finite-difference in 1D (forward)

2024-09-24

Finite difference approximation is a numerical method to compute derivatives by discretization. The forward difference illustrated here is a first order approximation of the order of the stepsize. [ref]

Twitter Post

thumbnail

Barycenters in the Wasserstein space

2024-09-23

Barycenters in the Wasserstein space were studied by Agueh & Carlier in 2011. They define meaningful geometric mean for probability distribution, and are used in practice for instance for histogram transfer (Rabin et al. ’15) [ref]

Twitter Post

GitHub File

thumbnail

Tarski—Seidenberg: geometric statement

2024-09-20

Tarski—Seidenberg theorem claims that semialgebraic sets on 𝐑 are stable by projection. [ref]

Twitter Post

thumbnail

Semialgebraic set

2024-09-19

Semialgebraic sets can be expressed as finite union of polynomial equalities and inequalities. [ref]

Twitter Post

thumbnail

Banach fixed point & Picard’s iterations

2024-09-18

Banach fixed point theorem guarantees existence and uniqueness of fixed point of contraction mapping in metric spaces. It is at the root of the fixed point method.

Twitter Post

thumbnail

Dual and polar cones

2024-09-17

Dual cone and polar cone are sets in the dual of a vector space. They generalize the notion of orthogonal subspace.

Twitter Post

thumbnail

Voronoi diagram

2024-09-16

Voronoi diagram partitions the space in a proximity graph (cells) according to some sites. Each cell is the portion of the space where every points are closer to a given site than any other sites. [ref]

Twitter Post

GitHub File

thumbnail

Hamming distance

2024-09-13

Hamming distance is fundamental in coding theory: it corresponds to the number of different characters between two words of the same length. [ref]

Twitter Post

thumbnail

The rendering equation

2024-09-12

The rendering equation is the fundamental recursive integral equation of light transport proposed by Kajiya in 1986, at the heart of (ray) path-tracing and particle tracing. Numerous sampling methods have been proposed to solve it numerically. [ref]

Twitter Post

thumbnail

Random graph models

2024-09-11

Latent position models encompass popular random models of graphs such as SBM, Erdos-Renyi, etc. They are based on a latent space and a connectivity kernel, and strongly linked to graphons in the dense case. [ref]

Twitter Post

thumbnail

Clarke Generalized derivative

2024-09-10

Clarke derivative generalizes the notion of convex subdifferential to nonconvex locally Lipschitz functions — thanks to Rademacher’s theorem — as the convex hull of the limiting sequences of gradient converging. [ref]

Twitter Post

thumbnail

Complex step approximation

2024-09-09

Complex step approximation is a numerical method to approximate the derivative from a single function evaluation using complex arithmetic. It is some kind of “poor man” automatic differentiation. [ref]

Twitter Post

thumbnail

Stochastic Block Model

2024-09-06

Stochastic block model is a popular generative model for graph proposed by Holland-Laskey-Leinhardt in 1983 to model communities. [ref]

Twitter Post

GitHub File

thumbnail

Thomae's popcorn & Dirichlet functions

2024-09-05

Thomae’s popcorn function is a (1-periodic) discontinuous function on ℚ but continuous on ℝ\ℚ that is nowhere differentiable. Dirichlet is nowhere continuous on ℝ. [ref]

Twitter Post

GitHub File

thumbnail

Weisfeiler-Leman test

2024-09-04

(1-dimensional) Weisfeiler-Leman is a heuristic for the graph isomorphism problem. It is a color refinement algorithm akin to a basic message passing scheme. [ref]

Twitter Post

thumbnail

Talagrand's inequality

2024-09-03

Talagrand's inequality is a probabilistic isoperimetric inequality that allows to derive a concentration inequality for the median. This is an instance of "concentration of measure" that made him win the Abel Prize in 2024. [ref]

Twitter Post

thumbnail

Representation of closed set

2024-09-02

Every closed subset of 𝐑ⁿ is the zero set of a infinitely differentiable function from 𝐑ⁿ to 𝐑 (due to Whitney)

Twitter Post

thumbnail