Post

A Ti–Yong–Shu Map of Probabilistic Modeling

A Ti–Yong–Shu Map of Probabilistic Modeling

This article is mainly generated by ChatGPT, I have tried many prompts and the results is still not perfrect. However I found it already quite useful to give me a mind map of this field.

The Bayesian worldview has become so influential across many areas of modern science that it’s almost impossible to ignore, whether one likes it or not. It often feels as if, unless you speak this language, people won’t take you seriously — to the point that it sometimes seems like a religion dressed up as science, or a science that resembles a religion. I believe we should treat it as one of many useful tools, rather than let it confine our way of thinking. A thoughtful reflection on this is Holes in Bayesian Statistics.

Anyway, let’s get back to the main topic.


The modern landscape of probabilistic modeling seems both rich and fragmented. From Bayes’ theorem to GANs, from variational inference to diffusion models, we find a vast collection of techniques—each powerful, yet sometimes disconnected. But is there a deeper unity?

To reveal this unity, we can borrow a classical conceptual framework from Chinese philosophy: 体 (Ti), 用 (Yong), and 术 (Shu)Essence, Function, and Method, which captures three complementary layers:

  • Ti (体, Essence): the worldview, the first principles.
  • Yong (用, Function): distinguishes why we model (infer, represent, generate, decide, explain).
  • Shu (术, Algorithm): distinguishes how we compute (maximize, sample, approximate, compete, propagate).

Each concrete model (ICA, VAE, GMM, etc.) is thus a point in this three-dimensional design space.

I. 体 (Ti): Foundational Principles

PrincipleEssence
1. Uncertainty as intrinsic to knowledgeModel distributions, not fixed values.
2. Latent generative processObserved data arise from hidden variables.
3. Learning = inferenceTo learn means to infer hidden causes or parameters.
4. Rational belief updatingBayes’ rule provides a consistent mechanism.
5. Structure and parsimonySimpler probabilistic structures generalize better.

II. 用 (Yong): Core Purposes and Functions

PurposeCore QuestionOutcome
InferenceWhat are the hidden parameters or causes given data?Posteriors, estimates.
Representation LearningHow to encode data into latent factors or components?Compressed latent space.
GenerationHow to synthesize new samples consistent with observed reality?Generative models.
Decision-makingHow to act optimally under uncertainty?Policies, expected utility.
Causal ReasoningHow to understand interventions and structure?Structural causal models.

III. 术 (Shu): Core Methods and Algorithms

Method FamilyDescriptionTypical Algorithms
Likelihood-based estimationOptimize data likelihood (possibly with priors).MLE, MAP, EM.
Sampling-based inferenceApproximate posterior by random samples.MCMC, Gibbs.
Optimization-based inferenceApproximate posterior by optimization.Variational inference, ELBO.
Energy minimization / message passingSolve inference in structured graphs.Belief propagation, mean-field.
Adversarial or contrastive learningLearn by distribution matching or contrastive objectives.GAN, Noise-Contrastive Estimation, Contrastive Divergence.

IV. 合 (He): Integrative Table — Mapping Ti → Yong → Shu

This is the fourth integrative layer — a cross-mapping table that shows for each major method:

  • what it is fundamentally for (Yong: purpose),
  • what principle or worldview it inherits (Ti connection), and
  • what technique or inference mechanism (Shu) it relies on.

Each method embodies one or more Yong (functions) and realizes it via certain Shu (techniques), under the common Ti (probabilistic worldview).

Model / MethodPrimary Yong (Function)Core Shu (Technique)Ti Connection (Foundational Principle)Remarks
MLEInference / Parameter estimationLikelihood maximizationLearning = inferencePure frequentist baseline.
MAPInference with prior knowledgeLikelihood + prior regularizationRational belief updatingBayesian variant of MLE.
EM AlgorithmInference (latent variables)Iterative E/M steps under MLELatent generative processGMM, HMM, Factor Analysis.
MCMCInference / SamplingMarkov chain samplingApproximate posterior inferenceFoundational for Bayesian computation.
Variational Inference (VI)Inference / ApproximationOptimization of ELBOLearning = inferenceCore to VAEs, topic models.
PCA / Probabilistic PCARepresentation learningClosed-form MLE under GaussianLatent generative processLinear Gaussian latent model.
ICARepresentation learning + inferenceMLE under independence constraintLatent sources assumptionLinks “Yong: representation” + “Shu: MLE”.
Sparse CodingRepresentation learningMAP (L1 prior on latent)Prior regularization = parsimonyBridge between inference & compression.
GMMGeneration + clusteringEM (MLE)Mixture latent processCanonical mixture model.
Bayesian NetworksCausal reasoning / inferenceExact or approximate inferenceStructured dependency modelingDirected graphical model.
MRF / CRFContextual inference / predictionEnergy minimization, message passingLocal dependency modelingUndirected structured model.
RBMRepresentation & generationContrastive Divergence (approx MLE)Energy-based latent structureBasis for deep belief nets.
Products of Experts (PoE)Generation / density modelingJoint energy minimizationCombining independent constraintsRBM is special case.
Field of Experts (FoE)Generation / image priorsMRF + learned filtersStructured local dependenciesNatural image modeling.
VAERepresentation + generationVariational inference (ELBO)Latent generative processDeep latent variable model.
GANGeneration (implicit)Adversarial training (min–max)Distributional realismNot explicit probability but shares generative Ti.
Normalizing FlowsGeneration / inferenceInvertible transformation, MLEProbabilistic bijectionExact likelihood models.
Diffusion ModelsGenerationReverse stochastic process trainingProbabilistic time-evolutionModern SOTA generative model.
HMM / Kalman FilterInference / sequence modelingEM, filtering, smoothingTemporal latent processSequential probabilistic structure.
Bayesian Decision TheoryDecision-makingExpected utility maximizationRational belief updatingFoundation for Bayesian RL.

V. Summary Diagram (Conceptual)

1
2
3
4
5
6
7
[体]  Foundations: Probability as logic of uncertainty
       ↓
[用]  Functions: Inference | Representation | Generation | Decision | Causality
       ↓
[术]  Techniques: MLE/MAP | EM | MCMC | VI | BP | Adversarial | Energy-based
       ↓
[合]  Concrete Methods: PCA, ICA, GMM, RBM, VAE, GAN, CRF, etc.

Each level unfolds naturally from the one above:

  • Understand “体”, and the functions (用) become inevitable.
  • Grasp “用”, and the diversity of methods (术) becomes intuitive.
  • The mapping (合) simply records how the abstract purposes manifest concretely.
This post is licensed under CC BY 4.0 by the author.