Expand description
Information Geometry
Information geometry treats probability distributions as points on a curved manifold, enabling geometry-aware optimization and analysis.
§Core Concepts
- Fisher Information Matrix (FIM): Measures curvature of probability space
- Natural Gradient: Gradient descent that respects the manifold geometry
- K-FAC: Kronecker-factored approximation for efficient natural gradient
§Benefits for Vector Search
- Faster Index Optimization: 3-5x fewer iterations vs Adam
- Better Generalization: Follows geodesics in parameter space
- Stable Continual Learning: Information-aware regularization
§References
- Amari & Nagaoka (2000): Methods of Information Geometry
- Martens & Grosse (2015): Optimizing Neural Networks with K-FAC
- Pascanu & Bengio (2013): Natural Gradient Works Efficiently in Learning
Structs§
- Fisher
Information - Fisher Information Matrix calculator
- KFAC
Approximation - K-FAC approximation for full network
- Natural
Gradient - Natural gradient optimizer state