Lab
Quant Bootcamp
Quant Marathon
Certificate
Why ARPM?
Store
Login
English
中文
Español
العربیه
Русский
Deutsch
Italiano
Français
Português
Srpski
Login
Home
Lab
Summary
Topics
Channels
Sign up
Enter the Lab
Quant Bootcamp
Summary
Program
Benefits
Reviews
Sign Up
Quant Marathon
Summary
Program
Benefits
Reviews
Sign Up
Certificate
Summary
Testing
Sign up
Why ARPM?
Why?
Who is ARPM for?
Clients & Partners
About ARPM
FAQ
Store
All
Bootcamp
Marathon
Data Science courses
Lab
Certificate
Contact us
Summary
Topics
Channels
Sign up
Enter the Lab
Summary
Program
Benefits
Reviews
Sign Up
Summary
Program
Benefits
Reviews
Sign Up
Summary
Testing
Sign up
Why?
Who is ARPM for?
Clients & Partners
About ARPM
FAQ
All
Bootcamp
Marathon
Single courses
Lab
Certificate
Video lectures
Theory
Case studies
Data animations
Code
Documentation
Slides
Exercises
TOC
Search
Introduction
About the ARPM Lab
Organization of the ARPM Lab
Learning the ARPM Lab by topic
Learning the ARPM Lab by channel
Audience and prerequisites
Notation
Key notation tenets
Operators and special functions
Sets
Calculus
Probability and general distribution theory
Summary statistical features
Distributions
Stochastic processes
Time conventions and counting indexes
Risk drivers
Invariants
Performance
Asset classes
Portfolio
Factor models and learning
Views processing
Investor preferences/profile
Acronyms
Glossary
Symbols
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
Y
Z
Bibliography
Quantitative finance
About quantitative finance: P and Q
Differences between P and Q
Commonalities between P and Q
I. Valuation
[ 0a ]
Valuation foundations
[ 0a.1 ]
Background definitions
[ 0a.1.1 ]
Instruments
[ 0a.1.2 ]
Value
[ 0a.1.3 ]
Cash-flows
[ 0a.1.4 ]
Re-invested cash-flows
[ 0a.1.5 ]
Cash-flow adjusted value
[ 0a.1.6 ]
Profit-and-loss (P&L)
[ 0a.1.7 ]
Payoff
[ 0a.2 ]
Points of interest and pitfalls
[ 0a.2.1 ]
Value versus price
[ 0a.2.2 ]
Multi-currency conversions
[ 0a.2.3 ]
Actual versus simple P&L
[ 0b ]
Linear pricing theory: core
[ 0b.1 ]
Fundamental axioms
[ 0b.1.1 ]
Law of one price
[ 0b.1.2 ]
Linearity
[ 0b.1.3 ]
Absence of arbitrage
[ 0b.1.4 ]
Relationships among fundamental axioms
[ 0b.2 ]
Fundamental theorem of asset pricing
[ 0b.2.1 ]
Linear pricing equation
[ 0b.2.2 ]
Numeraire
[ 0b.2.3 ]
Identification issues
[ 0b.3 ]
Risk-neutral pricing
[ 0b.3.1 ]
Discrete-time rebalancing
[ 0b.3.2 ]
No rebalancing: forward measure
[ 0b.3.3 ]
Continuous rebalancing limit
[ 0b.4 ]
Capital asset pricing framework
[ 0b.4.1 ]
Maximum Sharpe ratio portfolio
[ 0b.4.2 ]
Efficient frontier
[ 0b.4.3 ]
Security market line
[ 0b.4.4 ]
Connections to CAPM and linear factor models
[ 0b.5 ]
Covariance principle
[ 0c ]
Linear pricing theory: further assumptions
[ 0c.1 ]
Completeness
[ 0c.1.1 ]
General statement
[ 0c.1.2 ]
Arrow-Debreu securities
[ 0c.1.3 ]
European options
[ 0c.2 ]
Equilibrium: capital asset pricing model
[ 0c.3 ]
Arbitrage pricing theory
[ 0c.3.1 ]
Standard derivation: linear factor model for instruments
[ 0c.3.2 ]
Alternative derivation: linear factor model for stochastic discount factor
[ 0c.4 ]
Intertemporal consistency
[ 0c.4.1 ]
The framework
[ 0c.4.2 ]
Intertemporal linear pricing equation
[ 0c.4.3 ]
Intertemporal fundamental theorem of asset pricing
[ 0d ]
Non-linear pricing theory
[ 0d.1 ]
Fundamental axioms
[ 0d.1.1 ]
Law of one price
[ 0d.1.2 ]
Non-linearity
[ 0d.1.3 ]
Arbitrage
[ 0d.2 ]
Valuation as evaluation
[ 0d.2.1 ]
Variance and other shift principles
[ 0d.2.2 ]
Certainty-equivalent principle
[ 0d.2.3 ]
Distortion principles
[ 0d.2.4 ]
Esscher principle
[ 0d.3 ]
Intertemporal consistency
[ 0d.3.1 ]
Continuous time variables
[ 0d.3.2 ]
Non-linear “martingales”?
[ 0d.4 ]
Point of interest and pitfalls
[ 0d.4.1 ]
Linear (mis)uses of non-linear pricing
[ 0e ]
Valuation implementation
[ 0e.1 ]
Portfolio value
[ 0e.1.1 ]
Long positions
[ 0e.1.2 ]
Short positions
[ 0e.1.3 ]
Generic positions
[ 0e.1.4 ]
Portfolios, funds
[ 0e.1.5 ]
Sum-of-parts
[ 0e.1.6 ]
Valuation recipe
[ 0e.2 ]
Equities
[ 0e.2.1 ]
Discounted cash-flows
[ 0e.2.2 ]
Multiples
[ 0e.3 ]
Options
[ 0e.3.1 ]
Bachelier
[ 0e.3.2 ]
Black-Scholes
[ 0e.3.3 ]
Heston
[ 0e.3.4 ]
Valuation recipe
[ 0e.4 ]
Fixed-income
[ 0e.4.1 ]
Vasicek
[ 0e.4.2 ]
Other models
[ 0e.4.3 ]
Valuation recipe
[ 0e.5 ]
Insurance
[ 0e.5.1 ]
Life insurance
[ 0e.5.2 ]
Non-life insurance
[ 0e.6 ]
Real assets
II. The “Checklist”
[ 1 ]
Risk drivers identification
[ 1.1 ]
Equities
[ 1.2 ]
Currencies
[ 1.2.1 ]
Exchange rates
[ 1.2.2 ]
Contracts
[ 1.3 ]
Fixed-income
[ 1.3.1 ]
Zero-coupon bond
[ 1.3.2 ]
Rolling value
[ 1.3.3 ]
Yield to maturity
[ 1.3.4 ]
Alternative representations
[ 1.3.5 ]
Parsimonious representations
[ 1.3.6 ]
Spreads
[ 1.4 ]
Derivatives
[ 1.4.1 ]
Call option
[ 1.4.2 ]
Rolling value
[ 1.4.3 ]
Implied volatility
[ 1.4.4 ]
Alternative representations
[ 1.4.5 ]
Parsimonious representations
[ 1.4.6 ]
Pure volatility products
[ 1.5 ]
Credit
[ 1.5.1 ]
Definitions
[ 1.5.2 ]
Conditioning risk drivers
[ 1.5.3 ]
Aggregate risk drivers
[ 1.6 ]
Insurance
[ 1.7 ]
Operations
[ 1.8 ]
High frequency
[ 1.8.1 ]
Market microstructure
[ 1.8.2 ]
Activity time
[ 1.8.3 ]
Time-changed variables
[ 1.9 ]
Strategies
[ 1.10 ]
Points of interest, pitfalls, practical tips
[ 1.10.1 ]
Spurious heteroscedasticity
[ 1.10.2 ]
Spot, par, yield, forward, zero, risk-free,...which rate and curve?
[ 2 ]
Quest for invariance
[ 2.1 ]
Efficiency: random walk
[ 2.1.1 ]
Continuous-state random walk
[ 2.1.2 ]
Discrete-state random walk
[ 2.1.3 ]
Flexible combinations
[ 2.2 ]
Mean-reversion (continuous state): ARMA
[ 2.2.1 ]
AR(1) process
[ 2.2.2 ]
AR(1) generalizations
[ 2.3 ]
Mean-reversion (discrete state)
[ 2.3.1 ]
Markov chains
[ 2.3.2 ]
Markov chains for credit
[ 2.3.3 ]
Structural models for credit
[ 2.3.4 ]
Connection Markov chains/structural models
[ 2.4 ]
Long memory: fractional integration
[ 2.5 ]
Seasonality
[ 2.6 ]
Volatility clustering
[ 2.6.1 ]
GARCH
[ 2.6.2 ]
Extensions of GARCH
[ 2.6.3 ]
Stochastic volatility
[ 2.7 ]
Multivariate quest
[ 2.7.1 ]
Copula-marginal models
[ 2.7.2 ]
Efficiency: multivariate random walk
[ 2.7.3 ]
Mean reversion (continuous state): VAR(I)MA
[ 2.7.4 ]
Mean reversion (discrete state)
[ 2.7.5 ]
Volatility clustering
[ 2.7.6 ]
Toward machine learning
[ 2.8 ]
Cointegration
[ 2.8.1 ]
Modeling
[ 2.8.2 ]
Detection
[ 2.9 ]
Points of interest, pitfalls, practical tips
[ 2.9.1 ]
Returns are not invariants
[ 2.9.2 ]
Sampling step size
[ 3 ]
Estimation
[ 3.1 ]
Setting the flexible probabilities
[ 3.1.1 ]
Exponential decay and time conditioning
[ 3.1.2 ]
Kernels and state conditioning
[ 3.1.3 ]
Joint state and time conditioning
[ 3.1.4 ]
Statistical power of flexible probabilities
[ 3.2 ]
Historical
[ 3.2.1 ]
From historical distribution to flexible probabilities
[ 3.2.2 ]
Location-dispersion: HFP ellipsoid
[ 3.2.3 ]
Kernel estimation with flexible probabilities
[ 3.2.4 ]
Exponential moving moments and statistics
[ 3.3 ]
Maximum likelihood
[ 3.3.1 ]
From maximum likelihood to flexible probabilities
[ 3.3.2 ]
Exponential family invariants
[ 3.3.3 ]
Location-dispersion: normal MLFP ellipsoid
[ 3.3.4 ]
Location-dispersion: t MLFP ellipsoid
[ 3.4 ]
Bayesian
[ 3.4.1 ]
Exponential family invariants
[ 3.4.2 ]
Normal invariants
[ 3.5 ]
Shrinkage
[ 3.5.1 ]
Mean shrinkage: James-Stein
[ 3.5.2 ]
Covariance shrinkage: Ledoit-Wolf
[ 3.5.3 ]
Correlation shrinkage: random matrix theory
[ 3.5.4 ]
Covariance shrinkage: sparse eigenvector rotations
[ 3.5.5 ]
Covariance shrinkage: glasso
[ 3.5.6 ]
Covariance shrinkage: factor analysis
[ 3.6 ]
Generalized method of moments
[ 3.6.1 ]
Method of moments
[ 3.6.2 ]
Generalized method of moments - exact specification
[ 3.6.3 ]
Generalized method of moments - over specification
[ 3.7 ]
Robustness
[ 3.7.1 ]
Local robustness
[ 3.7.2 ]
Global robustness
[ 3.8 ]
Missing data
[ 3.8.1 ]
Randomly missing data
[ 3.8.2 ]
Times series of different length
[ 3.8.3 ]
Missing series
[ 3.9 ]
(Dynamic) copula-marginal
[ 3.9.1 ]
Static copula
[ 3.9.2 ]
Application: credit risk
[ 3.9.3 ]
Dynamic copula
[ 3.10 ]
Points of interest, pitfalls, practical tips
[ 3.10.1 ]
Unconditional estimation
[ 3.10.2 ]
Standardization
[ 3.10.3 ]
Non-synchronous data
[ 3.10.4 ]
High-frequency volatility/correlation
[ 3.10.5 ]
Outlier detection
[ 3.10.6 ]
Backward/forward exponential decay
[ 3.10.7 ]
Combining estimation techniques
[ 4 ]
Projection
[ 4.1 ]
One-step historical projection
[ 4.2 ]
Univariate analytical projection
[ 4.3 ]
Multivariate analytical projection
[ 4.4 ]
Monte Carlo
[ 4.4.1 ]
Markov models
[ 4.4.2 ]
State-space models
[ 4.4.3 ]
Generalizations
[ 4.5 ]
Historical
[ 4.5.1 ]
Historical bootstrapping
[ 4.5.2 ]
Hybrid Monte Carlo-historical
[ 4.6 ]
Application to credit risk
[ 4.6.1 ]
Copula-marginal approach
[ 4.7 ]
Square-root rule and generalizations
[ 4.7.1 ]
Thin-tailed random walk
[ 4.7.2 ]
Thick-tailed random walk
[ 4.7.3 ]
Multivariate random walk
[ 4.7.4 ]
General processes
[ 4.8 ]
Points of interest, pitfalls, practical tips
[ 4.8.1 ]
Consecutive (non-)overlapping sequences
[ 4.8.2 ]
Semi-analytical random walk projection via fast Fourier transform
[ 4.8.3 ]
GARCH generalized next-step
[ 4.8.4 ]
Projection step at a later stage of the Checklist
[ 4.8.5 ]
Pitfalls for square-root rule
[ 4.8.6 ]
Thick tails from thin tails
[ 4.8.7 ]
Martingales
[ 4.8.8 ]
Linear versus compounded returns
[ 4.8.9 ]
FFT for moving averages
[ 4.8.10 ]
Scenario projection enhancements by probability twisting
[ 4.8.11 ]
Jumps versus diffusion
[ 5 ]
Pricing at the horizon
[ 5.1 ]
Exact repricing
[ 5.1.1 ]
Equities
[ 5.1.2 ]
Currencies
[ 5.1.3 ]
Fixed-income
[ 5.1.4 ]
Derivatives
[ 5.1.5 ]
Credit
[ 5.1.6 ]
High frequency
[ 5.1.7 ]
Strategies
[ 5.2 ]
Carry
[ 5.2.1 ]
Equities
[ 5.2.2 ]
Currencies
[ 5.2.3 ]
Fixed-income
[ 5.2.4 ]
Derivatives
[ 5.2.5 ]
Other asset classes
[ 5.3 ]
Taylor approximations
[ 5.3.1 ]
Equities
[ 5.3.2 ]
Fixed-income
[ 5.3.3 ]
Derivatives
[ 5.3.4 ]
Other asset classes
[ 5.4 ]
Hybrid Taylor/repricing approximation
[ 5.4.1 ]
Global quadratic approximation
[ 5.4.2 ]
Interpolated residual
[ 5.5 ]
Joint P&L distribution
[ 5.5.1 ]
Analytical approach
[ 5.5.2 ]
Scenario-based approach
[ 5.6 ]
Testing the pricing function
[ 5.7 ]
Pitfalls and practical tips
[ 5.7.1 ]
Pricing at the horizon and arbitrage
[ 5.7.2 ]
Path dependence
[ 5.7.3 ]
“Pricing at the horizon”versus “asset pricing/valuation theory”
[ 5.7.4 ]
Black-Merton-Scholes is exactly correct!
[ 5.7.5 ]
Greeks for intra-day updates
[ 5.7.6 ]
Greeks at the horizon
[ 5.7.7 ]
Bond carry versus accrued interest
[ 5.7.8 ]
Option carry versus theta
[ 6 ]
Aggregation
[ 6.1 ]
Stock variables
[ 6.1.1 ]
Portfolio
[ 6.1.2 ]
Value
[ 6.1.3 ]
Exposure
[ 6.1.4 ]
Leverage
[ 6.2 ]
Credit value adjustment
[ 6.3 ]
Liquidity value adjustment
[ 6.4 ]
Static market/credit risk
[ 6.4.1 ]
Standardized holdings and weights
[ 6.4.2 ]
Scenario-probability distribution
[ 6.4.3 ]
Elliptical distribution
[ 6.4.4 ]
Quadratic-normal distribution
[ 6.5 ]
Dynamic market/credit risk
[ 6.6 ]
Stress-testing
[ 6.6.1 ]
Theory
[ 6.6.2 ]
Why have stress-tests
[ 6.6.3 ]
Panic copula
[ 6.6.4 ]
Extreme copula
[ 6.7 ]
Enterprise risk management
[ 6.7.1 ]
Portfolio: balance sheet
[ 6.7.2 ]
Performance: income statement
[ 6.7.3 ]
Operational risk
[ 6.7.4 ]
Banking simplified regulatory framework
[ 6.7.5 ]
Insurance simplified framework
[ 6.8 ]
Points of interest and pitfalls
[ 6.8.1 ]
Solvency and collateral
[ 6.8.2 ]
CreditRisk+ approximation
[ 7 ]
Ex-ante evaluation
[ 7.1 ]
Stochastic dominance
[ 7.2 ]
Satisfaction/risk measures
[ 7.3 ]
Mean-variance trade-off
[ 7.3.1 ]
Mean
[ 7.3.2 ]
Variance
[ 7.3.3 ]
Standard deviation
[ 7.3.4 ]
Mean-variance trade-off
[ 7.3.5 ]
A strange success story
[ 7.4 ]
The fundamental risk quadrangle
[ 7.4.1 ]
Relevant cases
[ 7.4.2 ]
Generalizations
[ 7.5 ]
Expected utility and certainty-equivalent
[ 7.5.1 ]
Common examples
[ 7.5.2 ]
Computation
[ 7.6 ]
Value at Risk and quantile
[ 7.6.1 ]
Definition
[ 7.6.2 ]
Computation
[ 7.7 ]
Expected shortfall and sub-quantile
[ 7.7.1 ]
Definition
[ 7.7.2 ]
Computation
[ 7.8 ]
Spectral/distortion satisfaction measures
[ 7.8.1 ]
Definition
[ 7.8.2 ]
Common examples
[ 7.8.3 ]
Computation
[ 7.9 ]
Coherent satisfaction measures
[ 7.9.1 ]
Definition
[ 7.9.2 ]
Common examples
[ 7.9.3 ]
Computation
[ 7.10 ]
Induced expectations
[ 7.10.1 ]
Definition
[ 7.10.2 ]
Common examples
[ 7.10.3 ]
Computation
[ 7.11 ]
Non-dimensional ratios
[ 7.11.1 ]
Information ratio
[ 7.11.2 ]
Downside ratios
[ 7.11.3 ]
Correlation
[ 7.12 ]
Enterprise risk management
[ 7.13 ]
Pitfalls, points of interest and practical tips
[ 7.13.1 ]
The Arrow-Pratt approximation of the certainty-equivalent
[ 7.13.2 ]
Utility versus quantile
[ 7.13.3 ]
Utility versus spectrum functions
[ 7.13.4 ]
The Buhlmann and Esscher expectations are not distortion expectations
[ 7.13.5 ]
Satisfaction measures under normality
[ 8 ]
Ex-ante attribution
[ 8a ]
Ex-ante attribution: performance
[ 8a.1 ]
Bottom-up exposures
[ 8a.1.1 ]
Pricing factors
[ 8a.1.2 ]
Style factors/smart beta
[ 8a.2 ]
Top-down exposures: factors on demand
[ 8a.2.1 ]
Analytical computation
[ 8a.2.2 ]
Cardinality constraints
[ 8a.3 ]
Relationship between bottom-up and top-down exposures
[ 8a.3.1 ]
Subportfolios
[ 8a.3.2 ]
Style factors/smart beta
[ 8a.4 ]
Joint distribution
[ 8a.4.1 ]
Elliptical distribution
[ 8a.4.2 ]
Scenario-probability distribution
[ 8a.5 ]
Application: hedging
[ 8a.6 ]
Pitfalls and practical tips
[ 8a.6.1 ]
Estimation versus attribution
[ 8a.6.2 ]
The ex-ante attribution is not a regression on past data
[ 8b ]
Ex-ante attribution: risk
[ 8b.1 ]
General criteria
[ 8b.1.1 ]
Isolated/“first in”proportional attribution
[ 8b.1.2 ]
“Last in”proportional attribution
[ 8b.1.3 ]
Sequential attribution
[ 8b.1.4 ]
Shapley attribution
[ 8b.2 ]
Euler decomposition
[ 8b.2.1 ]
Standard deviation and variance
[ 8b.2.2 ]
Certainty-equivalent
[ 8b.2.3 ]
Quantile
[ 8b.2.4 ]
Sub-quantile
[ 8b.2.5 ]
Spectral satisfaction measures
[ 8b.2.6 ]
Coherent measures
[ 8b.2.7 ]
Economic capital
[ 8b.3 ]
Linear attribution for induced expectations
[ 8b.3.1 ]
Actuarial pricing
[ 8b.4 ]
Minimum-torsion bets attribution of variance
[ 8b.4.1 ]
Minimum-torsion bets
[ 8b.4.2 ]
Effective number of bets
[ 9 ]
Construction
[ 9a ]
Construction: portfolio optimization
[ 9a.1 ]
Mean-variance principles
[ 9a.2 ]
Sensitivity analysis
[ 9a.3 ]
Analytical solutions of the mean-variance problem
[ 9a.4 ]
Portfolio replication
[ 9a.5 ]
Benchmark allocation
[ 9a.6 ]
Pitfalls and practical tips
[ 9b ]
Construction: estimation and model risk
[ 9b.1 ]
Estimation risk measurement
[ 9b.2 ]
Sample-based allocation
[ 9b.3 ]
Prior allocation
[ 9b.4 ]
Estimation (=input) uncertainty: Bayesian allocation
[ 9b.5 ]
Allocation (=output) uncertainty: Robust allocation
[ 9b.6 ]
Diversification management
[ 9c ]
Construction: cross-sectional strategies
[ 9c.1 ]
Simplistic portfolio construction
[ 9c.1.1 ]
Backtesting
[ 9c.2 ]
Advanced portfolio construction
[ 9c.2.1 ]
Signal-induced factor
[ 9c.2.2 ]
Flexible factor
[ 9c.2.3 ]
Backtesting
[ 9c.3 ]
Relationship with FLAM and APT
[ 9c.3.1 ]
Signal-induced moments of the P&L
[ 9c.3.2 ]
Fundamental law of active management
[ 9c.3.3 ]
APT assumption
[ 9c.4 ]
Multiple portfolios
[ 9c.4.1 ]
Characteristic portfolios
[ 9c.4.2 ]
Signal-induced factors
[ 9c.4.3 ]
Flexible factors
[ 9c.4.4 ]
Relationship with FLAM and APT
[ 9c.5 ]
Points of interest, pitfalls, practical tips
[ 9c.5.1 ]
Machine learning
[ 9c.5.2 ]
Generalized fundamental law of active management
[ 9d ]
Construction: time series strategies
[ 9d.1 ]
The market
[ 9d.1.1 ]
Risky investment
[ 9d.1.2 ]
Low-risk investment
[ 9d.1.3 ]
Strategies
[ 9d.2 ]
Expected utility maximization
[ 9d.2.1 ]
The objective
[ 9d.2.2 ]
Optimization
[ 9d.3 ]
Option based portfolio insurance
[ 9d.3.1 ]
Payoff design
[ 9d.3.2 ]
Partial differential equation
[ 9d.3.3 ]
Budget
[ 9d.3.4 ]
Policy
[ 9d.3.5 ]
A unified approach
[ 9d.4 ]
Rolling horizon heuristics
[ 9d.4.1 ]
Constant proportion portfolio insurance
[ 9d.4.2 ]
Drawdown control
[ 9d.5 ]
Signal induced strategy
[ 9d.6 ]
Convexity analysis
[ 10 ]
Execution
[ 10.1 ]
Market impact modeling
[ 10.1.1 ]
Liquidity curve
[ 10.1.2 ]
Exogenous impact
[ 10.1.3 ]
Endogenous impact
[ 10.2 ]
Order scheduling
[ 10.2.1 ]
Trading P&L decomposition
[ 10.2.2 ]
Model P&L
[ 10.2.3 ]
Moments of model P&L
[ 10.2.4 ]
Model P&L optimization
[ 10.2.5 ]
Quasi-optimal P&L distribution
[ 10.3 ]
Order placement
[ 10.3.1 ]
Step 1: order scheduling
[ 10.3.2 ]
Step 2: order placement
[ 10.4 ]
Points of interest, pitfalls, practical tips
[ 10.4.1 ]
Mean-variance optimization in complex models
[ 10.4.2 ]
Price manipulation
[ 10.4.3 ]
Testing
III. Performance analysis
[ 11 ]
Performance attribution
IV. Financial toolbox
[ 12 ]
Performance definitions
[ 12.1 ]
Holding P&L of a portfolio
[ 12.1.1 ]
Long positions
[ 12.1.2 ]
Short positions
[ 12.1.3 ]
Generic positions
[ 12.1.4 ]
Portfolios, funds
[ 12.2 ]
Trading P&L
[ 12.2.1 ]
Single transaction
[ 12.2.2 ]
Multiple transactions in one position
[ 12.2.3 ]
Portfolio rebalancing
[ 12.3 ]
Implementation shortfall
[ 12.4 ]
Returns
[ 12.4.1 ]
Basic definitions
[ 12.4.2 ]
Standard linear returns and weights
[ 12.4.3 ]
Generalized linear returns
[ 12.4.4 ]
Generalized weights and aggregation
[ 12.4.5 ]
Investments with capital injection
[ 12.4.6 ]
Log-returns
[ 12.5 ]
Excess performance
[ 12.5.1 ]
Benchmark
[ 12.5.2 ]
Excess return
[ 12.6 ]
Path analysis
[ 12.7 ]
Pitfalls and practical tips
[ 12.7.1 ]
Linear versus compounded returns
[ 13 ]
Signals
[ 13.1 ]
Carry signals
[ 13.1.1 ]
Fixed-income
[ 13.1.2 ]
Foreign exchange
[ 13.2 ]
Value signals
[ 13.2.1 ]
Book
[ 13.2.2 ]
Pricing
[ 13.3 ]
Technical signals
[ 13.3.1 ]
Momentum
[ 13.3.2 ]
Filters
[ 13.3.3 ]
Cointegration
[ 13.4 ]
Microstructure signals
[ 13.4.1 ]
Trade autocorrelation
[ 13.4.2 ]
Order imbalance
[ 13.4.3 ]
Price prediction
[ 13.4.4 ]
Volume clustering
[ 13.5 ]
Fundamental and other signals
[ 13.6 ]
Signal processing
[ 13.6.1 ]
Smoothing
[ 13.6.2 ]
Scoring
[ 13.6.3 ]
Ranking
[ 14 ]
Black-Litterman
[ 14.1 ]
Equilibrium distribution
[ 14.1.1 ]
Performance model
[ 14.1.2 ]
Prior distribution of expected returns
[ 14.1.3 ]
Prior predictive performance distribution
[ 14.2 ]
Active views
[ 14.2.1 ]
Active views model
[ 14.2.2 ]
Active views statement
[ 14.2.3 ]
Posterior distribution of the expected returns
[ 14.3 ]
Black-Litterman posterior
[ 14.4 ]
Limit cases and generalizations
[ 14.4.1 ]
High confidence in prior
[ 14.4.2 ]
Low confidence in views
[ 14.4.3 ]
High confidence in views
[ 14.4.4 ]
Generalizations
[ 14.4.5 ]
From linear returns to risk drivers
[ 14.4.6 ]
From stock-like to generic asset classes
[ 14.4.7 ]
From normal to non-normal markets
[ 14.4.8 ]
From linear equality views to partial flexible views
Data science
V. Mathematics
[ 15 ]
Linear algebra primer
[ 15.1 ]
Vector spaces
[ 15.1.1 ]
Vector operations
[ 15.1.2 ]
Basis and coordinates
[ 15.1.3 ]
Vector subspaces
[ 15.2 ]
Linear transformations
[ 15.2.1 ]
Matrix representation
[ 15.2.2 ]
Composition
[ 15.2.3 ]
Invertibility
[ 15.3 ]
Inner product spaces
[ 15.3.1 ]
Symmetry
[ 15.3.2 ]
Positivity
[ 15.3.3 ]
Length, distance and angle
[ 15.3.4 ]
Orthogonal projection
[ 15.3.5 ]
Best prediction
[ 15.3.6 ]
Rotations
[ 15.4 ]
Metric and normed spaces
[ 15.4.1 ]
Norm
[ 15.4.2 ]
Distance
[ 15.4.3 ]
Divergence
[ 15.5 ]
Spectral decomposition
[ 15.5.1 ]
Eigenvalues and eigenvectors
[ 15.5.2 ]
Spectral theorem
[ 15.5.3 ]
Matrix notation
[ 15.5.4 ]
Generalizations
[ 15.6 ]
Matrix transpose-square-root
[ 15.6.1 ]
Gramian
[ 15.6.2 ]
Transpose-square-roots
[ 15.6.3 ]
Orthonormalization
[ 15.6.4 ]
Spectrum/principal components
[ 15.6.5 ]
Cholesky/Gram-Schmidt
[ 15.6.6 ]
Riccati/minimum torsion
[ 15.7 ]
Matrix operations
[ 15.7.1 ]
The vector space of matrices
[ 15.7.2 ]
Key operations
[ 15.7.3 ]
Pseudo-inverse
[ 15.7.4 ]
Useful identities
[ 15.8 ]
Matrix polynomials
[ 15.8.1 ]
Matrix polynomials factorization
[ 15.8.2 ]
Matrix polynomial inversion
[ 16 ]
Calculus primer
[ 16.1 ]
Differentiation
[ 16.1.1 ]
Univariate functions
[ 16.1.2 ]
Multivariate functions
[ 16.1.3 ]
Matrix-variate functions
[ 16.2 ]
Numerical derivatives
[ 16.2.1 ]
Univariate approximations
[ 16.2.2 ]
Multivariate approximations
[ 16.3 ]
Taylor expansion
[ 16.3.1 ]
Univariate functions
[ 16.3.2 ]
Multivariate functions
[ 16.4 ]
Integration
[ 16.4.1 ]
Univariate integration
[ 16.4.2 ]
Fundamental theorem of calculus
[ 16.4.3 ]
Multivariate integration
[ 16.5 ]
Monotone functions
[ 16.5.1 ]
Univariate monotonicity
[ 16.5.2 ]
Entrywise monotonicity
[ 16.5.3 ]
Monotone maps
[ 16.6 ]
Convexity
[ 16.6.1 ]
Univariate convexity
[ 16.6.2 ]
Multivariate convexity
[ 17 ]
Functional analysis primer
[ 17.1 ]
Measure theory
[ 17.1.1 ]
Domains
[ 17.1.2 ]
Measures
[ 17.1.3 ]
Lebesgue’s decomposition
[ 17.2 ]
Functional algebra
[ 17.2.1 ]
Function spaces
[ 17.2.2 ]
Linear operators
[ 17.2.3 ]
Kernel representation
[ 17.3 ]
L2 spaces
[ 17.3.1 ]
Inner product
[ 17.3.2 ]
Dirac delta
[ 17.3.3 ]
Riesz representation theorem
[ 17.3.4 ]
Convolution
[ 17.3.5 ]
Unitary operators
[ 17.4 ]
Fourier transform
[ 17.4.1 ]
The convolution theorem
[ 17.4.2 ]
Fourier integral transform
[ 17.4.3 ]
Discrete time Fourier transform
[ 17.4.4 ]
Fourier series
[ 17.4.5 ]
Discrete Fourier transform
[ 17.5 ]
Spectral theorem
[ 17.5.1 ]
Mercer kernels
[ 17.5.2 ]
Eigenvalues and eigenfunctions
[ 17.5.3 ]
Motivation
[ 17.5.4 ]
Matrix-valued kernels
[ 17.6 ]
Bochner’s theorem
[ 17.6.1 ]
Spectral representation
[ 17.6.2 ]
Matrix-valued kernels
[ 17.7 ]
Mercer’s theorem
[ 17.7.1 ]
Spectral representation
[ 17.7.2 ]
Reproducing kernel Hilbert spaces
[ 17.7.3 ]
Matrix-valued kernels
[ 17.8 ]
Functional calculus
[ 17.8.1 ]
Gateaux derivative
[ 17.8.2 ]
Frechet derivative
[ 17.8.3 ]
Second order derivative
[ 18 ]
Optimization primer
[ 18.1 ]
Fundamental concepts
[ 18.2 ]
Smooth programming
[ 18.2.1 ]
Smooth programming
[ 18.2.2 ]
Lagrange multipliers
[ 18.2.3 ]
Gradient descent
[ 18.2.4 ]
Newton’s method
[ 18.3 ]
Convex programming
[ 18.3.1 ]
Conic programming
[ 18.3.2 ]
Semidefinite programming
[ 18.3.3 ]
Second-order cone programming
[ 18.3.4 ]
Quadratic programming
[ 18.3.5 ]
Ridge, lasso, elastic nets
[ 18.3.6 ]
Linear programming
[ 18.4 ]
Integer ¯n-choose-k selection
[ 18.4.1 ]
Notation
[ 18.4.2 ]
Exact solution
[ 18.4.3 ]
General heuristics principles
[ 18.4.4 ]
Naive selection
[ 18.4.5 ]
Forward step-wise selection
[ 18.4.6 ]
Step-wise generalizations
[ 18.4.7 ]
Ridge, lasso, elastic net heuristics
[ 18.5 ]
Points of interest
[ 18.5.1 ]
Equivalent optimization problems
VI. Statistics
[ 19 ]
Distributions
[ 19.1 ]
Representations of a distribution
[ 19.1.1 ]
Random variables
[ 19.1.2 ]
Univariate distributions
[ 19.1.3 ]
Multivariate distributions
[ 19.2 ]
Marginalization
[ 19.3 ]
Conditioning
[ 19.3.1 ]
Conditional variables
[ 19.3.2 ]
Bayes theorem
[ 19.3.3 ]
Conditional features
[ 19.3.4 ]
Deterministic versus stochastic conditioning
[ 19.4 ]
Normal distribution
[ 19.4.1 ]
Pdf, cdf and characteristic function
[ 19.4.2 ]
Moments
[ 19.4.3 ]
Conditional distribution
[ 19.4.4 ]
Stochastic representations
[ 19.4.5 ]
Affine equivariance
[ 19.4.6 ]
Matrix-normal distribution
[ 19.5 ]
Notable univariate distributions
[ 19.5.1 ]
Chi-squared distribution
[ 19.5.2 ]
Gamma distribution
[ 19.5.3 ]
Quadratic-normal distribution
[ 19.6 ]
Notable multivariate distributions
[ 19.6.1 ]
Student t
[ 19.6.2 ]
Cauchy
[ 19.6.3 ]
Uniform distribution
[ 19.6.4 ]
Uniform inside the ellipsoid
[ 19.6.5 ]
Uniform on the ellipsoid
[ 19.6.6 ]
Lognormal distribution
[ 19.6.7 ]
Wishart distribution
[ 19.7 ]
Elliptical distributions
[ 19.7.1 ]
Fundamental concepts
[ 19.7.2 ]
Moments and dependence
[ 19.7.3 ]
Stochastic representations
[ 19.7.4 ]
Affine equivariance
[ 19.7.5 ]
Generation of elliptical scenarios
[ 19.7.6 ]
Scenario generation with dimension reduction
[ 19.8 ]
Scenario-probability distributions
[ 19.8.1 ]
Types of scenario-probability distributions
[ 19.8.2 ]
Probability mass and density function
[ 19.8.3 ]
Transformations and generalized expectations
[ 19.8.4 ]
Cumulative distribution function
[ 19.8.5 ]
Continuous cumulative distribution function
[ 19.8.6 ]
Quantile
[ 19.8.7 ]
Continuous quantile
[ 19.8.8 ]
Moments and other statistical features
[ 19.9 ]
Categorical distributions
[ 19.9.1 ]
Discriminant variables
[ 19.9.2 ]
Probabilities parametrization
[ 19.10 ]
Exponential family distributions
[ 19.10.1 ]
Normal distribution
[ 19.10.2 ]
Scenario-probability distribution
[ 19.11 ]
Mixture distributions
[ 19.11.1 ]
Binary case
[ 19.11.2 ]
Multi-class case
[ 19.12 ]
Other special classes of distributions
[ 19.12.1 ]
Stable distributions
[ 19.12.2 ]
Additive distributions
[ 19.12.3 ]
Infinitely divisible distributions
[ 19.13 ]
Distributions cheat sheet
[ 19.13.1 ]
Univariate distributions
[ 19.13.2 ]
Multivariate distributions
[ 19.13.3 ]
Matrix-valued distributions
[ 19.14 ]
Points of interest, pitfalls, practical tips
[ 19.14.1 ]
Probability spaces
[ 19.14.2 ]
Finite spaces
[ 19.14.3 ]
Discretized random variables
[ 20 ]
Copulas
[ 20.1 ]
Univariate results
[ 20.2 ]
Definition and properties
[ 20.2.1 ]
Grades
[ 20.2.2 ]
Copula
[ 20.2.3 ]
Sklar’s theorem
[ 20.2.4 ]
Copula invariance
[ 20.3 ]
Special classes of copulas
[ 20.3.1 ]
Elliptical copulas
[ 20.3.2 ]
Archimedean copulas
[ 20.4 ]
Implementation
[ 20.4.1 ]
Copula-marginal separation
[ 20.4.2 ]
Copula-marginal combination
[ 21 ]
Geometry of distributions
[ 21.1 ]
Distributions geometry
[ 21.1.1 ]
Fisher metric: length and volume
[ 21.1.2 ]
Flatness and geodesics
[ 21.1.3 ]
Duality: potentials and Legendre transformations
[ 21.1.4 ]
Distance and divergence
[ 21.2 ]
Exponential distributions geometry
[ 21.3 ]
Scenario-probability distribution geometry
[ 22 ]
Location and dispersion
[ 22.1 ]
Expectation and variance
[ 22.1.1 ]
Key definitions
[ 22.1.2 ]
Visualization: uncertainty band
[ 22.1.3 ]
Affine equivariance
[ 22.1.4 ]
Variational principles
[ 22.2 ]
Expectation and covariance
[ 22.2.1 ]
Key definitions
[ 22.2.2 ]
Affine equivariance
[ 22.2.3 ]
Linear algebra: spectral decomposition
[ 22.2.4 ]
Visualization: ellipsoid
[ 22.2.5 ]
Statistics: principal component analysis
[ 22.2.6 ]
Calculus: most important summary statistics
[ 22.2.7 ]
Probability
[ 22.2.8 ]
Variational principles
[ 22.3 ]
L2 geometry
[ 22.3.1 ]
Expectation inner product
[ 22.3.2 ]
Covariance (improper) inner product
[ 22.3.3 ]
Covariance versus expectation inner product
[ 22.3.4 ]
Least squares prediction
[ 22.3.5 ]
Orthonormal sets
[ 22.3.6 ]
Visualization: Euclidean vectors
[ 22.4 ]
Generalized affine equivariance
[ 22.4.1 ]
Univariate case
[ 22.4.2 ]
Multivariate case
[ 22.5 ]
Generalized variational principles
[ 22.5.1 ]
Key definitions
[ 22.5.2 ]
Relevant cases
[ 22.5.3 ]
Multivariate extensions
[ 22.6 ]
Points of interest, pitfalls, practical tips
[ 22.6.1 ]
Alternative visualizations in low dimension
[ 22.6.2 ]
Lp geometry
[ 23 ]
Correlation and generalizations
[ 23.1 ]
Measures of dependence
[ 23.1.1 ]
Schweizer-Wolff measure
[ 23.1.2 ]
Mutual information
[ 23.2 ]
Measures of concordance
[ 23.2.1 ]
Kendall’s tau
[ 23.2.2 ]
Spearman’s rho
[ 23.3 ]
Correlation
[ 23.3.1 ]
Related definitions
[ 23.4 ]
Points of interest, pitfalls, practical tips
[ 23.4.1 ]
Schweizer and Wolff measure via simulations
[ 24 ]
Statistical decision theory
[ 24.1 ]
Stochastic dominance fundamentals
[ 24.1.1 ]
Strong dominance
[ 24.1.2 ]
Weak dominance
[ 24.1.3 ]
Second and higher order dominance
[ 24.2 ]
No-data decision problem
[ 24.2.1 ]
Stochastic dominance failure
[ 24.2.2 ]
Bayesian evaluation
[ 24.2.3 ]
Robust evaluation
[ 24.2.4 ]
General functionals evaluation
[ 24.3 ]
Data-driven decision problem
[ 24.3.1 ]
Frequentist approach
[ 24.3.2 ]
Bayesian approach
[ 24.3.3 ]
Admissibility
[ 24.3.4 ]
Model and estimation risk
[ 24.3.5 ]
Unified approach
[ 24.4 ]
Sequential decision theory and dynamic programming
[ 25 ]
Useful algorithms
[ 25.1 ]
Moment-matching scenarios
[ 25.1.1 ]
Twisting scenarios
[ 25.1.2 ]
Twisting probabilities
[ 25.2 ]
Normalized empirical histogram
VII. Factor models and learning
[ 26 ]
Linear factor models
[ 26.1 ]
Overview
[ 26.1.1 ]
The r-squared
[ 26.1.2 ]
Dominant-residual models
[ 26.1.3 ]
Systematic-idiosyncratic models
[ 26.1.4 ]
Estimation
[ 26.2 ]
Regression LFM’s
[ 26.2.1 ]
Definition
[ 26.2.2 ]
Solution: factor loadings
[ 26.2.3 ]
Prediction and fit
[ 26.2.4 ]
Residuals features
[ 26.2.5 ]
Natural scatter specification
[ 26.2.6 ]
The ∥ projection operator
[ 26.2.7 ]
Estimation
[ 26.3 ]
Principal-component LFM’s
[ 26.3.1 ]
Definition
[ 26.3.2 ]
Solution: factor loadings and constructed factors
[ 26.3.3 ]
Identification issues
[ 26.3.4 ]
Prediction and fit
[ 26.3.5 ]
Residuals features
[ 26.3.6 ]
Natural scatter specification
[ 26.3.7 ]
Estimation
[ 26.4 ]
Factor-analysis LFM’s
[ 26.4.1 ]
Definition
[ 26.4.2 ]
Solution: factor loadings and idiosyncratic variances
[ 26.4.3 ]
Exact principal component with isotropic variances
[ 26.4.4 ]
Identification issues
[ 26.4.5 ]
Factor scores
[ 26.4.6 ]
Prediction and fit
[ 26.4.7 ]
Residuals features
[ 26.4.8 ]
Natural scatter specification
[ 26.4.9 ]
Estimation
[ 26.5 ]
Cross-sectional LFM’s
[ 26.5.1 ]
Definition
[ 26.5.2 ]
Solution: factor-construction matrix
[ 26.5.3 ]
Prediction and fit
[ 26.5.4 ]
Residuals features
[ 26.5.5 ]
Natural scatter specification
[ 26.5.6 ]
Systematic-idiosyncratic assumption
[ 26.5.7 ]
Estimation
[ 26.6 ]
Points of interest, pitfalls, practical tips
[ 26.6.1 ]
LFM’s are not a regression on past data(1) [⋆⋆]
[ 26.6.2 ]
LFM’s are not about returns(2-3) [⋆⋆]
[ 26.6.3 ]
LFM’s are not about stocks(4) [⋆⋆]
[ 26.6.4 ]
LFM’s “factors”are not “factors returns”(5) [⋆]
[ 26.6.5 ]
LFM’s are not systematic-idiosyncratic(6-7) [⋆⋆⋆]
[ 26.6.6 ]
LFM’s are not horizon-independent(8) [⋆⋆]
[ 26.6.7 ]
LFM’s are not a dimension reduction technique(9) [⋆]
[ 26.6.8 ]
LFM’s are not APT and CAPM(10-11-12) [⋆⋆⋆]
[ 26.6.9 ]
Factor analysis LFM’s are not idiosyncratic(14) [⋆⋆]
[ 26.6.10 ]
LFM’s do not extract premia-generating factors(16)[⋆⋆]
[ 26.6.11 ]
LFM’s are not always necessary(13-15-17-18) [⋆⋆⋆]
[ 26.6.12 ]
Affine versus linear formulation
[ 26.6.13 ]
Linear regression: a success story
[ 26.6.14 ]
Principal factors are not principal components
[ 26.6.15 ]
Performance of regression versus principal component
[ 26.6.16 ]
Conditional principal component
[ 26.6.17 ]
More general constraints
[ 27 ]
Machine learning foundations
[ 27.1 ]
Approaches to machine learning
[ 27.1.1 ]
Supervised learning
[ 27.1.2 ]
Unsupervised learning
[ 27.1.3 ]
Reinforcement learning
[ 27.1.4 ]
Linear factor models roots
[ 27.2 ]
Prediction
[ 27.2.1 ]
Point prediction
[ 27.2.2 ]
Probabilistic prediction
[ 27.2.3 ]
Point/probabilistic connections
[ 27.3 ]
Learning
[ 27.3.1 ]
Frequentist vs Bayesian
[ 27.3.2 ]
In-sample fit
[ 27.3.3 ]
Regularization
[ 27.3.4 ]
Decision theoretical interpretation and assessment
[ 27.4 ]
Inference
[ 27.4.1 ]
Marginalization and mode computation
[ 27.4.2 ]
Inference vs learning
[ 27.5 ]
Points of interest, pitfalls, practical tips
[ 27.5.1 ]
More about scoring rules and distributional divergences
[ 28 ]
Supervised learning: regression
[ 28.1 ]
Least squares regression
[ 28.1.1 ]
Error
[ 28.1.2 ]
Theoretical optimum
[ 28.1.3 ]
Optimization in practice
[ 28.1.4 ]
Linear basis
[ 28.1.5 ]
ANOVA
[ 28.1.6 ]
Trees
[ 28.1.7 ]
Neural networks
[ 28.1.8 ]
Gradient boosting
[ 28.1.9 ]
Kernel trick
[ 28.1.10 ]
Geometrical interpretation
[ 28.2 ]
Non-least squares regression
[ 28.2.1 ]
Error
[ 28.2.2 ]
Theoretical optimum
[ 28.2.3 ]
Optimization in practice
[ 28.2.4 ]
Linear basis
[ 28.2.5 ]
Advanced methods
[ 28.2.6 ]
Generalized point regression
[ 28.3 ]
Discriminant regression
[ 28.3.1 ]
Error
[ 28.3.2 ]
Theoretical optimum
[ 28.3.3 ]
Target parameters
[ 28.3.4 ]
Optimization in practice
[ 28.3.5 ]
Linear regression
[ 28.3.6 ]
Generalized linear models
[ 28.3.7 ]
Alternative generalizations
[ 28.4 ]
Points of interest
[ 28.4.1 ]
Alternative errors
[ 28.4.2 ]
Exponential tilting
[ 29 ]
Supervised learning: classification
[ 29.1 ]
Binary classification
[ 29.1.1 ]
Joint distribution
[ 29.1.2 ]
Error
[ 29.1.3 ]
Theoretical optimum
[ 29.1.4 ]
Receiver operating characteristic (ROC)
[ 29.1.5 ]
Optimization in practice
[ 29.1.6 ]
Perceptron
[ 29.1.7 ]
Support vector machines
[ 29.1.8 ]
Fisher discriminant analysis
[ 29.2 ]
Multinomial classification
[ 29.2.1 ]
Joint distribution
[ 29.2.2 ]
Error
[ 29.2.3 ]
Theoretical optimum
[ 29.2.4 ]
Classification via discriminants
[ 29.2.5 ]
Optimization in practice
[ 29.2.6 ]
Leveraging binary classifiers
[ 29.3 ]
Discriminant classification
[ 29.3.1 ]
Error
[ 29.3.2 ]
Theoretical optimum
[ 29.3.3 ]
Alternative approaches
[ 29.3.4 ]
Target parameters
[ 29.3.5 ]
Optimization in practice
[ 29.3.6 ]
Logistic regression
[ 29.3.7 ]
Naive Bayes classifiers
[ 29.3.8 ]
Probit regression
[ 29.3.9 ]
Neural networks
[ 29.3.10 ]
Trees
[ 29.3.11 ]
Gradient boosting
[ 29.4 ]
Points of interest, pitfalls, practical tips
[ 29.4.1 ]
Binary classification: alternative errors
[ 29.4.2 ]
Linear regression for classification
[ 30 ]
Unsupervised learning
[ 30.1 ]
Least squares autoencoders
[ 30.1.1 ]
Minimum torsion variables
[ 30.1.2 ]
k-means clustering
[ 30.1.3 ]
Kernel trick
[ 30.1.4 ]
Independent component analysis
[ 30.2 ]
Graphical models
[ 30.2.1 ]
Graphs
[ 30.2.2 ]
Definition
[ 30.2.3 ]
Probabilistic factor analysis
[ 30.2.4 ]
Mixture models
[ 30.2.5 ]
Naive Bayes models
[ 30.2.6 ]
Bayes networks
[ 30.2.7 ]
Markov networks
[ 31 ]
Generalized probabilistic inference
[ 31.1 ]
Minimum relative entropy
[ 31.1.1 ]
Base distribution and view variables
[ 31.1.2 ]
Point views
[ 31.1.3 ]
Distributional views
[ 31.1.4 ]
Partial views
[ 31.1.5 ]
Partial views on generalized expectations
[ 31.1.6 ]
Sanity check
[ 31.1.7 ]
Confidence
[ 31.1.8 ]
Relationship with Bayesian updating
[ 31.2 ]
Analytical implementation
[ 31.2.1 ]
Base distribution
[ 31.2.2 ]
Views
[ 31.2.3 ]
Updated distribution
[ 31.2.4 ]
Confidence
[ 31.2.5 ]
Relevant special cases
[ 31.3 ]
Flexible probabilities implementation
[ 31.3.1 ]
Base distribution
[ 31.3.2 ]
Views
[ 31.3.3 ]
Updated distribution
[ 31.3.4 ]
Confidence
[ 31.4 ]
Factor-based implementations
[ 31.5 ]
Copula opinion pooling
[ 31.5.1 ]
Base distribution
[ 31.5.2 ]
Views
[ 31.5.3 ]
Updated distribution
[ 31.5.4 ]
Confidence
[ 31.5.5 ]
The algorithm
[ 31.6 ]
Generalized shrinkage
[ 31.6.1 ]
Intuition
[ 31.6.2 ]
Classical shrinkage
[ 31.6.3 ]
Bayesian updating
[ 31.6.4 ]
Minimum relative entropy
[ 31.6.5 ]
Shrinkage
[ 31.6.6 ]
Regularization
[ 32 ]
Dynamic and spatial models
[ 32.1 ]
Overview
[ 32.2 ]
Least squares dynamic models
[ 32.3 ]
Levinson filtering
[ 32.4 ]
Wiener-Kolmogorov filtering
[ 32.4.1 ]
Solution
[ 32.4.2 ]
Causal filtering
[ 32.5 ]
Kriging
[ 32.6 ]
Dynamic principal component analysis
[ 32.6.1 ]
Solution
[ 32.6.2 ]
Computational issue
[ 32.7 ]
Functional principal component analysis
[ 32.8 ]
Probabilistic state space models
[ 32.8.1 ]
Probabilistic linear state-space models
[ 32.8.2 ]
Hidden Markov models
[ 32.8.3 ]
Hidden Markov VAR(1) models
VIII. Stochastic processes
[ 33 ]
Stochastic processes primer
[ 33.1 ]
Preliminary definitions
[ 33.1.1 ]
Probabilistic processes
[ 33.1.2 ]
Mean-covariance processes
[ 33.1.3 ]
Gaussian processes
[ 33.1.4 ]
Information sets and prediction
[ 33.2 ]
Main classes of processes
[ 33.2.1 ]
White noise
[ 33.2.2 ]
Random walk
[ 33.2.3 ]
Martingales
[ 33.2.4 ]
Markov processes
[ 33.2.5 ]
State space processes
[ 33.3 ]
Main properties
[ 33.3.1 ]
Stationarity
[ 33.3.2 ]
Ergodicity
[ 33.3.3 ]
Integrated processes
[ 33.3.4 ]
Cointegrated process
[ 33.4 ]
Stochastic processes decomposition
[ 33.4.1 ]
Mean-covariance linear
[ 33.4.2 ]
Probabilistic uncorrelated
[ 33.4.3 ]
Probabilistic independent
[ 33.5 ]
Random fields
[ 33.5.1 ]
Spatial fields
[ 33.5.2 ]
Probabilistic graphical models
[ 33.5.3 ]
Free fields
[ 33.5.4 ]
Functional data models
[ 33.5.5 ]
Gaussian random fields
[ 33.6 ]
Discrete time processes
[ 33.6.1 ]
Filtrations
[ 33.6.2 ]
Iterated expectations
[ 33.6.3 ]
Adapted processes
[ 33.6.4 ]
Martingales
[ 33.6.5 ]
Approximations of processes
[ 34 ]
Covariance (stationary) processes
[ 34.1 ]
Intuition
[ 34.2 ]
Main definitions
[ 34.2.1 ]
Non-stationary moments
[ 34.2.2 ]
Stationary moments
[ 34.2.3 ]
Detrending
[ 34.3 ]
Prediction - general case
[ 34.3.1 ]
Problem
[ 34.3.2 ]
Solution
[ 34.3.3 ]
Practical issues
[ 34.3.4 ]
Solution: prediction as orthogonal projection
[ 34.4 ]
Spectral representation
[ 34.4.1 ]
Spectral theorem - intuition
[ 34.4.2 ]
Spectral theorem - formal statement
[ 34.4.3 ]
Cramer decomposition - intuition
[ 34.4.4 ]
Cramer decomposition - formal statement
[ 34.5 ]
Filtering
[ 34.5.1 ]
Intuition
[ 34.5.2 ]
Formal definitions
[ 34.5.3 ]
Autocovariance function
[ 34.5.4 ]
Spectrum
[ 34.5.5 ]
Composition, inversion
[ 34.5.6 ]
Causality
[ 34.5.7 ]
Time domain filters
[ 34.5.8 ]
Frequency domain filters
[ 34.6 ]
Wold representation
[ 34.6.1 ]
Intuition
[ 34.6.2 ]
Formal statement
[ 34.6.3 ]
Weak white noise via Gram-Schmidt
[ 34.6.4 ]
Linearly regular component
[ 34.6.5 ]
Linearly deterministic component
[ 35 ]
Common covariance (stationary) processes
[ 35.1 ]
Vector autoregression of order one
[ 35.1.1 ]
Definitions
[ 35.1.2 ]
Weak random walk
[ 35.1.3 ]
Causal vector autoregression
[ 35.1.4 ]
Cointegrated vector autoregression
[ 35.1.5 ]
Prediction
[ 35.2 ]
VARMA processes
[ 35.2.1 ]
Definitions
[ 35.2.2 ]
Causal VARMA
[ 35.2.3 ]
Invertible VARMA
[ 35.3 ]
VARIMA(p,d,q)
[ 35.4 ]
Linear state space models
[ 35.4.1 ]
Definitions
[ 35.4.2 ]
Covariance stationary linear state-space models
[ 35.4.3 ]
Systematic-idiosyncratic
[ 35.4.4 ]
Prediction
[ 35.4.5 ]
The Kalman filter
[ 35.5 ]
Points of interest
[ 35.5.1 ]
VAR(1) as universal approximation
[ 36 ]
Invariance tests
[ 36.1 ]
Simple tests
[ 36.2 ]
Refinements and pitfalls
[ 36.2.1 ]
Circle-like covariance (not data)
[ 36.2.2 ]
Stronger tests based on copulas
[ 37 ]
Continuous time processes
[ 37.1 ]
Efficiency: Lévy processes
[ 37.1.1 ]
Numerical implementation
[ 37.1.2 ]
Infinite divisibility
[ 37.1.3 ]
Continuous state: Brownian diffusion
[ 37.1.4 ]
Discrete state: Poisson jumps
[ 37.1.5 ]
Lévy-Khintchine representation
[ 37.1.6 ]
Subordination
[ 37.2 ]
Mean-reversion (continuous state)
[ 37.2.1 ]
Ornstein-Uhlenbeck
[ 37.2.2 ]
Square-root process and other generalizations
[ 37.3 ]
Mean-reversion (discrete state)
[ 37.3.1 ]
Time-homogeneous Markov chain
[ 37.3.2 ]
Time-inhomogeneous Markov chains
[ 37.3.3 ]
Stationarity and unconditional distributions
[ 37.4 ]
Long memory: fractional Brownian motion
[ 37.4.1 ]
Fractional Brownian motion
[ 37.5 ]
Volatility clustering
[ 37.5.1 ]
Stochastic volatility
[ 37.5.2 ]
Time change
[ 37.5.3 ]
Connection between time-changed Brownian motion and stochastic volatility
[ 37.6 ]
Multivariate mean reversion
[ 37.6.1 ]
Definitions
[ 37.6.2 ]
Conditional distribution of MVOU
[ 37.6.3 ]
Stationarity and unconditional distribution of MVOU
[ 37.6.4 ]
Geometrical interpretation∗
[ 37.6.5 ]
Cointegrated Ornstein-Uhlenbeck
[ 37.6.6 ]
Relationship between (V)AR and (MV)OU
IX. Estimation theory
[ 38 ]
Probabilistic estimation and inference techniques
[ 38.1 ]
Maximum likelihood
[ 38.1.1 ]
Observable processes
[ 38.1.2 ]
Latent variables
[ 38.1.3 ]
EM algorithm
[ 38.1.4 ]
IID latent processes
[ 38.1.5 ]
Markov state-space processes
[ 38.1.6 ]
Networks
[ 38.2 ]
Bayesian statistics
[ 38.2.1 ]
Estimation
[ 38.2.2 ]
Prediction
[ 38.2.3 ]
Analytical solutions
[ 38.3 ]
Inference via Monte Carlo
[ 38.3.1 ]
Metropolis-Hastings
[ 38.4 ]
Inference and learning via variational techniques
[ 38.4.1 ]
IM projection
[ 38.4.2 ]
Inference
[ 38.4.3 ]
Learning
[ 38.4.4 ]
Analytical solution: exponential family
[ 38.4.5 ]
Variational solution
[ 38.4.6 ]
EM algorithm in population
[ 38.4.7 ]
Dimension reduction
[ 39 ]
Estimation and assessment
[ 39.1 ]
Probabilistic prediction assessment for invariants
[ 39.1.1 ]
Estimators as decisions
[ 39.1.2 ]
Frequentist approach
[ 39.1.3 ]
Bayesian approach
[ 39.1.4 ]
Analytical results
[ 39.1.5 ]
Monte Carlo simulations
[ 39.1.6 ]
Cross-validation
[ 39.2 ]
Bias versus variance
[ 39.3 ]
Point prediction assessment
[ 39.3.1 ]
Estimators as decisions
[ 39.3.2 ]
Frequentist approach
[ 39.3.3 ]
Bayesian approach
[ 39.3.4 ]
Historical with flexible probabilities estimators
[ 39.3.5 ]
Cross-validation
[ 39.4 ]
Probabilistic prediction assessment
[ 39.4.1 ]
Estimators as decisions
[ 39.4.2 ]
Frequentist approach
[ 39.4.3 ]
Bayesian approach
[ 39.4.4 ]
Maximum likelihood with flexible probabilities
[ 39.4.5 ]
Cross-validation
[ 40 ]
Bias reduction
[ 40.1 ]
Functional fit
[ 40.1.1 ]
Machine learning as functional optimization
[ 40.1.2 ]
Parametric functional optimization
[ 40.2 ]
Linear basis
[ 40.2.1 ]
Interactions/polynomials
[ 40.2.2 ]
Orthogonal series
[ 40.2.3 ]
Error minimization
[ 40.3 ]
Trees
[ 40.3.1 ]
CART
[ 40.3.2 ]
Splines
[ 40.3.3 ]
Voronoi diagrams
[ 40.3.4 ]
Error minimization
[ 40.4 ]
Neural networks
[ 40.4.1 ]
Neurons
[ 40.4.2 ]
Neural networks
[ 40.4.3 ]
Projection pursuit
[ 40.4.4 ]
Error minimization
[ 40.5 ]
Gradient boosting
[ 40.6 ]
Kernel trick
[ 41 ]
Estimation and regularization
[ 41.1 ]
Background
[ 41.1.1 ]
Panels
[ 41.1.2 ]
Assumptions
[ 41.1.3 ]
Estimation
[ 41.1.4 ]
Testing
[ 41.1.5 ]
Bias and variance
[ 41.2 ]
Regularization
[ 41.2.1 ]
Stepwise features selection
[ 41.2.2 ]
Ridge, lasso, elastic nets
[ 41.2.3 ]
Glasso
[ 41.2.4 ]
Categorical factors selection
[ 41.2.5 ]
Bayesian prior
[ 41.3 ]
Sparse principal component
[ 41.4 ]
Ensemble learning
[ 41.4.1 ]
Bagging
[ 41.4.2 ]
Flexible probabilities as random-variables
[ 41.4.3 ]
Flexible probabilities through conditioning
[ 41.4.4 ]
Ensemble weighting
[ 42 ]
Hypothesis testing
[ 42.1 ]
Hypothesis testing for invariants
[ 42.1.1 ]
Statistics
[ 42.1.2 ]
P-value
[ 42.1.3 ]
Univariate testing: the z-statistic
[ 42.1.4 ]
Multivariate testing: the Hotelling statistic
Featured case studies
X. Quantitative finance: the “Checklist”
[ 43 ]
Historical Checklist
[ 43.1 ]
Risk drivers identification
[ 43.2 ]
Quest for invariance
[ 43.3 ]
Estimation
[ 43.4 ]
Projection
[ 43.5 ]
Pricing at the horizon
[ 43.6 ]
Aggregation
[ 43.7 ]
Ex-ante evaluation
[ 43.8 ]
Ex-ante attribution
[ 43.8.1 ]
Ex-ante attribution: performance
[ 43.8.2 ]
Ex-ante attribution: risk
[ 43.9 ]
Construction
[ 43.10 ]
Execution
[ 44 ]
Monte Carlo Checklist
[ 44.1 ]
Risk drivers identification
[ 44.1.1 ]
Market
[ 44.1.2 ]
Credit
[ 44.2 ]
Quest for invariance
[ 44.2.1 ]
Market
[ 44.2.2 ]
Credit
[ 44.3 ]
Estimation
[ 44.3.1 ]
Market
[ 44.3.2 ]
Credit
[ 44.4 ]
Projection
[ 44.4.1 ]
Market
[ 44.4.2 ]
Credit
[ 44.5 ]
Pricing at the horizon
[ 44.6 ]
Aggregation
[ 44.7 ]
Ex-ante evaluation
[ 44.8 ]
Ex-ante attribution
[ 44.8.1 ]
Ex-ante attribution: performance
[ 44.8.2 ]
Ex-ante attribution: risk
[ 44.9 ]
Construction
[ 44.10 ]
Execution
XI. Data science: factor models and learning
[ 45 ]
Principal component analysis of the yield curve
[ 45.1 ]
Cross-sectional structure of the yield curve covariance
[ 45.2 ]
Finite set of times to maturity
[ 45.3 ]
The continuum limit
[ 46 ]
Machine learning for hedging
[ 46.1 ]
Least squares regression
[ 46.1.1 ]
Theoretical optimum
[ 46.1.2 ]
Linear least squares regression
[ 46.1.3 ]
Least squares regression tree
[ 46.2 ]
Least absolute distance regression
[ 46.2.1 ]
Theoretical optimum
[ 46.2.2 ]
Linear least absolute distance regression
[ 46.2.3 ]
Least absolute distance regression tree
[ 47 ]
Regression in the stock market
[ 47.1 ]
Time series models
[ 47.2 ]
Maximum likelihood
[ 47.2.1 ]
The model
[ 47.2.2 ]
Normal assumption
[ 47.2.3 ]
Student t assumption
[ 47.3 ]
Factor selection for regression
[ 47.3.1 ]
Stepwise regression selection
[ 47.3.2 ]
Lasso regression
[ 47.3.3 ]
Ridge regression
[ 47.4 ]
Bayesian
[ 47.4.1 ]
Normal conditional likelihood
[ 47.4.2 ]
Normal-inverse-Wishart prior distribution
[ 47.4.3 ]
Normal-inverse-Wishart posterior distribution
[ 47.4.4 ]
Student t predictive distribution
[ 47.4.5 ]
Classical equivalent
[ 47.4.6 ]
Uncertainty
[ 47.5 ]
Mixed approach
[ 48 ]
Credit default classification
[ 48.1 ]
Background
[ 48.2 ]
Fit and assessment
[ 48.3 ]
Logistic regression
[ 48.4 ]
Interactions
[ 48.5 ]
Encoding
[ 48.6 ]
Regularization
[ 48.7 ]
Trees
[ 48.8 ]
Gradient boosting
[ 48.9 ]
Cross-validation
[ 49 ]
Clustering for the stock market
[ 49.1 ]
k-means clustering
[ 49.2 ]
Shrinkage
Previous page
Next page
Private content
Content reserved to registered users.
Login
or
Sign Up
for free to access all chapters summaries.
Copy