4
A C B Small Model Middle Model Large Model Figure 1 Parameter Space parameters of a small model is an analytic set with singu Fisher information matrix depends on the parameter.

A C B Small Model Middle Model Large Model Figure 1 Parameter Space The set of parameters of a small model is an analytic set with singularities. Rank

Embed Size (px)

Citation preview

Page 1: A C B Small Model Middle Model Large Model Figure 1 Parameter Space The set of parameters of a small model is an analytic set with singularities. Rank

A

CB

Small Model

Middle Model

Large Model

Figure 1 Parameter Space

The set of parameters of a small model is an analytic set with singularities. Rank of the Fisher information matrix depends on the parameter.

Page 2: A C B Small Model Middle Model Large Model Figure 1 Parameter Space The set of parameters of a small model is an analytic set with singularities. Rank

H(w)

0

g(u)

RealManifold U

ResolutionMap

KullbackInformation

Figure 2 Resolution of Singularities

Hironaka’s theorem ensures that we can algorithmically find a resolution map which makes the Kullback informationbe a direct product of local coordinates.

H(g(u))  =  a(u) u

1

k1 u2

k2 … u3

k3

ParameterSpace W

Page 3: A C B Small Model Middle Model Large Model Figure 1 Parameter Space The set of parameters of a small model is an analytic set with singularities. Rank

A

C B

True distribution

Figure 3 Bias and Variance

The variance of a singular point is smaller than that of a regular point. If the number of training samples is not so large, then singular pointsA or B are selected in Bayesian estimation.

Page 4: A C B Small Model Middle Model Large Model Figure 1 Parameter Space The set of parameters of a small model is an analytic set with singularities. Rank

Figure 4 Learning Curve

The learning curve of a hierarchical learning machine is bounded by those of several smaller machines.

n: The number of training samples

G(n) : The generalization error