Home Page Icon
Home Page
Table of Contents for
Cover image
Close
Cover image
by John Kruschke
Doing Bayesian Data Analysis, 2nd Edition
Cover image
Title page
Table of Contents
Copyright
Dedication
Chapter 1: What's in This Book (Read This First!)
1.1 Real people can read this book
1.2 What's in this book
1.3 What's new in the second edition?
1.4 Gimme feedback (Be polite)
1.5 Thank you!
Part I: The Basics: Models, Probability, Bayes’ Rule, and R
Introduction
Chapter 2: Introduction: Credibility, Models, and Parameters
2.1 Bayesian inference is reallocation of credibility across possibilities
2.2 Possibilities are parameter values in descriptive models
2.3 The steps of bayesian data analysis
2.4 Exercises
Chapter 3: The R Programming Language
3.1 Get the software
3.2 A simple example of R in action
3.3 Basic commands and operators in R
3.4 Variable types
3.5 Loading and saving data
3.6 Some utility functions
3.7 Programming in R
3.8 Graphical plots: Opening and saving
3.9 Conclusion
3.10 Exercises
Chapter 4: What is This Stuff Called Probability?
4.1 The set of all possible events
4.2 Probability: Outside or inside the head
4.3 Probability distributions
4.4 Two-way distributions
4.5 Appendix: R code for figure 4.1
4.6 Exercises
Chapter 5: Bayes' Rule
5.1 Bayes' rule
5.2 Applied to parameters and data
5.3 Complete examples: Estimating bias in a coin
5.4 Why bayesian inference can be difficult
5.5 Appendix: R code for figures 5.1, 5.2, etc.
5.6 Exercises
Part II: All the Fundamentals Applied to Inferring a Binomial Probability
Introduction
Chapter 6: Inferring a Binomial Probability via Exact Mathematical Analysis
6.1 The likelihood function: Bernoulli distribution
6.2 A description of credibilities: The beta distribution
6.3 The posterior beta
6.4 Examples
6.5 Summary
6.6 Appendix: R code for figure 6.4
6.7 Exercises
Chapter 7: Markov Chain Monte Carlo
7.1 Approximating a distribution with a large sample
7.2 A simple case of the metropolis algorithm
7.3 The metropolis algorithm more generally
7.4 Toward gibbs sampling: Estimating two coin biases
7.5 Mcmc representativeness, accuracy, and efficiency
7.6 Summary
7.7 Exercises
Chapter 8: JAGS
8.1 Jags and its relation to R
8.2 A complete example
8.3 Simplified scripts for frequently used analyses
8.4 Example: difference of biases
8.5 Sampling from the prior distribution in jags
8.6 Probability distributions available in jags
8.7 Faster sampling with parallel processing in runjags
8.8 Tips for expanding jags models
8.9 Exercises
Chapter 9: Hierarchical Models
9.1 A single coin from a single mint
9.2 Multiple coins from a single mint
9.3 Shrinkage in hierarchical models
9.4 Speeding up jags
9.5 Extending the hierarchy: Subjects within categories
9.6 Exercises
Chapter 10: Model Comparison and Hierarchical Modeling
10.1 General formula and the bayes factor
10.2 Example: two factories of coins
10.3 Solution by MCMC
10.4 Prediction: Model averaging
10.5 Model complexity naturally accounted for
10.6 Extreme sensitivity to prior distribution
10.7 Exercises
Chapter 11: Null Hypothesis Significance Testing
11.1 Paved with good intentions
11.2 Prior knowledge
11.3 Confidence interval and highest density interval
11.4 Multiple comparisons
11.5 What a sampling distribution is good for
11.6 Exercises
Chapter 12: Bayesian Approaches to Testing a Point (“Null”) Hypothesis
12.1 The estimation approach
12.2 The model-comparison approach
12.3 Relations of parameter estimation and model comparison
12.4. Estimation or model comparison?
12.5. Exercises
Chapter 13: Goals, Power, and Sample Size
13.1 The will to power
13.2 Computing power and sample size
13.3 Sequential testing and the goal of precision
13.4 Discussion
13.5 Exercises
Chapter 14: Stan
14.1 HMC sampling
14.2 Installing stan
14.3 A complete example
14.4 Specify models top-down in stan
14.5 Limitations and extras
14.6 Exercises
Part III: The Generalized Linear Model
Introduction
Chapter 15: Overview of the Generalized Linear Model
15.1 Types of variables
15.2 Linear combination of predictors
15.3 Linking from combined predictors to noisy predicted data
15.4 Formal expression of the GLM
15.5 Exercises
Chapter 16: Metric-Predicted Variable on One or Two Groups
16.1 Estimating the mean and standard deviation of a normal distribution
16.2 Outliers and robust estimation: The t distribution
16.3 Two groups
16.4 Other noise distributions and transforming data
16.5 Exercises
Chapter 17: Metric Predicted Variable with One Metric Predictor
17.1 Simple linear regression
17.2 Robust linear regression
17.3 Hierarchical regression on individuals within groups
17.4 Quadratic trend and weighted data
17.5 Procedure and perils for expanding a model
17.6 Exercises
Chapter 18: Metric Predicted Variable with Multiple Metric Predictors
18.1 Multiple linear regression
18.2 Multiplicative interaction of metric predictors
18.3 Shrinkage of regression coefficients
18.4 Variable selection
18.5 Exercises
Chapter 19: Metric Predicted Variable with One Nominal Predictor
19.1 Describing multiple groups of metric data
19.2 Traditional analysis of variance
19.3 Hierarchical bayesian approach
19.4 Including a metric predictor
19.5 Heterogeneous variances and robustness against outliers
19.6 Exercises
Chapter 20: Metric Predicted Variable with Multiple Nominal Predictors
20.1 Describing groups of metric data with multiple nominal predictors
20.2 Hierarchical bayesian approach
20.3 Rescaling can change interactions, homogeneity, and normality
20.4 Heterogeneous variances and robustness against outliers
20.5 Within-subject designs
20.6 Model comparison approach
20.7 Exercises
Chapter 21: Dichotomous Predicted Variable
21.1 Multiple metric predictors
21.2 Interpreting the regression coefficients
21.3 Robust logistic regression
21.4 Nominal predictors
21.5 Exercises
Chapter 22: Nominal Predicted Variable
22.1 Softmax regression
22.2 Conditional logistic regression
22.3 Implementation in jags
22.4 Generalizations and variations of the models
22.5 Exercises
Chapter 23: Ordinal Predicted Variable
23.1 Modeling ordinal data with an underlying metric variable
23.2 The case of a single group
23.3 The case of two groups
23.4 The case of metric predictors
23.5 Posterior prediction
23.6 Generalizations and extensions
23.7 Exercises
Chapter 24: Count Predicted Variable
24.1 Poisson exponential model
24.2 Example: hair eye go again
24.3 Example: interaction contrasts, shrinkage, and omnibus test
24.4 Log-linear models for contingency tables
24.5 Exercises
Chapter 25: Tools in the Trunk
25.1 Reporting a bayesian analysis
25.2 Functions for computing highest density intervals
25.3 Reparameterization
25.4 Censored data in JAGS
25.5 What next?
Bibliography
Search in book...
Toggle Font Controls
Playlists
Add To
Create new playlist
Name your new playlist
Playlist description (optional)
Cancel
Create playlist
Sign In
Email address
Password
Forgot Password?
Create account
Login
or
Continue with Facebook
Continue with Google
Sign Up
Full Name
Email address
Confirm Email Address
Password
Login
Create account
or
Continue with Facebook
Continue with Google
Next
Next Chapter
Title page
Add Highlight
No Comment
..................Content has been hidden....................
You can't read the all page of ebook, please click
here
login for view all page.
Day Mode
Cloud Mode
Night Mode
Reset