• Tidak ada hasil yang ditemukan

Bayesian Networks With Examples in R pdf pdf

N/A
N/A
Protected

Academic year: 2019

Membagikan "Bayesian Networks With Examples in R pdf pdf"

Copied!
239
0
0

Teks penuh

  • Penulis:
    • M. Scutari
    • J.-B. Denis
  • Sekolah: UCL Genetics Institute
  • Mata Pelajaran: Genetics
  • Topik: Bayesian Networks: With Examples in R
  • Tipe: book
  • Tahun: 2015
  • Kota: London

I. The Discrete Case: Multinomial Bayesian Networks

This chapter introduces the fundamental concepts of Bayesian Networks (BNs) using a hypothetical transport survey. The focus is on modeling discrete data, with continuous data and more complex data types addressed in subsequent chapters. The chapter systematically guides the reader through building and interpreting a BN, emphasizing the practical application of the theoretical principles.

1.1 Introductory Example: Train Use Survey

This section introduces a hypothetical survey investigating transport usage patterns, focusing on cars and trains. Six discrete variables are defined: Age, Sex, Education, Occupation, Residence, and Travel. These variables are categorized into demographic and socioeconomic indicators, with Travel as the target variable. The example's context is relevant for understanding the practical applications of BNs in analyzing survey data and drawing inferences about relationships between variables.

1.2 Graphical Representation

This section details the representation of variable relationships using a directed acyclic graph (DAG). Each node represents a variable, and arcs represent direct dependencies. The concepts of parents, children, paths, and cycles are explained. The use of the bnlearn R package is introduced for creating and manipulating DAGs, illustrating the practical steps involved in constructing the graphical representation of a BN. The example uses the empty.graph, set.arc, and modelstring functions within bnlearn to build the DAG step-by-step, demonstrating how direct dependencies are coded within the graph.

1.3 Probabilistic Representation

Here, the joint probability distribution over the variables is defined. The concept of the global distribution is introduced, followed by the factorization of this distribution into smaller, local distributions based on the DAG's conditional independence relationships. This factorization greatly reduces the number of parameters, a key advantage of BNs. The section demonstrates how to define local distributions using R's array and matrix functions, creating conditional probability tables (CPTs) for each variable. The code illustrates the creation of CPTs for variables with different numbers of parents, showing how to specify probabilities within the tables, thus defining the probabilistic component of the BN model.

1.4 Estimating the Parameters: Conditional Probability Tables

This section focuses on estimating the parameters (conditional probabilities) of the local distributions from observed data. Maximum likelihood estimation (MLE) using empirical frequencies is described, demonstrated with R code using the bn.fit function from the bnlearn package. Bayesian estimation, offering a more robust alternative to MLE, is introduced and illustrated, highlighting the role of the prior distribution and the imaginary sample size (iss) parameter in influencing posterior estimates. The concept of smoothing conditional probabilities to avoid sparse tables and improve prediction is discussed, showing the effect of varying the iss value on the resulting estimates.

1.5 Learning the DAG Structure: Tests and Scores

This section addresses the scenario where the DAG structure is unknown and needs to be learned from data. This is a common task in fields like genetics and systems biology, where inferring network structures is crucial. The challenges of learning DAG structures are briefly introduced, setting the stage for a deeper discussion of structure learning algorithms in later chapters. The section touches on the use of conditional independence tests and network scores as methods for structure learning, creating a bridge to the more detailed theoretical exposition in Chapter 4.

1.6 Using Discrete BNs

This section delves into the practical application of discrete BNs, focusing on how to use the DAG structure and the CPTs for inference. Exact and approximate inference methods are briefly introduced, providing a foundation for the more in-depth treatment of inference algorithms in Chapter 4. This section connects the theoretical aspects of BNs to their practical use in making predictions and drawing inferences from the model.

1.7 Plotting BNs

This section describes how to visually represent the learned BNs, focusing on plotting both the DAG structure and the CPTs. This visualization helps in understanding the network's relationships and the probability distributions associated with each node. While specific R code for plotting may not be explicitly detailed, the importance of visualization in understanding and communicating BN models is highlighted.

1.8 Further Reading

This section provides pointers to further reading on the topics covered in the chapter, directing readers to more advanced resources for a deeper understanding of discrete Bayesian Networks.

II. The Continuous Case: Gaussian Bayesian Networks

This chapter extends the concepts introduced in Chapter 1 to handle continuous data using Gaussian Bayesian Networks. Similar steps are followed, but the probabilistic representation is now based on Gaussian distributions and the focus shifts to correlation coefficients instead of conditional probability tables.

III. More Complex Cases: Hybrid Bayesian Networks

This chapter addresses the modeling of more complex scenarios involving hybrid Bayesian networks that contain a mixture of discrete and continuous variables. Strategies for handling such mixed data types are presented, including techniques for discretizing continuous variables and using different probability distributions.

IV. Theory and Algorithms for Bayesian Networks

This chapter delves into the theoretical foundations of Bayesian Networks, covering topics such as conditional independence, graphical separation, Markov blankets, moral graphs, and structure and parameter learning algorithms. Inference methods and causal Bayesian networks are also discussed, providing a more rigorous mathematical treatment of the subject matter.

V. Software for Bayesian Networks

This chapter provides an overview of available software packages for working with Bayesian Networks, focusing on R packages such as bnlearn, deal, catnet, and pcalg, as well as other software like BUGS, BayesiaLab, Hugin, and GeNIe.

VI. Real-World Applications of Bayesian Networks

This chapter presents real-world applications of Bayesian Networks in diverse fields. Two case studies are analyzed in detail to illustrate the practical use of BNs in solving real-world problems.

VII. Graph Theory

This appendix provides background information on graph theory, covering basic definitions and concepts relevant to understanding the graphical representation of Bayesian Networks.

VIII. Probability Distributions

This appendix reviews relevant probability distributions, including discrete and continuous distributions, and their properties. This section provides the necessary background in probability theory for understanding the probabilistic aspects of Bayesian Networks.

IX. A Note about Bayesian Networks

This appendix clarifies the relationship between Bayesian Networks and Bayesian statistics, addressing potential misconceptions.

Referensi Dokumen

  • Categorical Data Analysis ( Agresti, A. )
  • Local Causal and Markov Blanket Induction for Causal Discovery and Feature Selection for Classification Part I: Algorithms and Empirical Evaluation ( Aliferis, C. F., Statnikov, A., Tsamardinos, I., Mani, S., and Xenofon, X. D. )
  • HUGIN – a Shell for Building Bayesian Belief Universes for Expert Systems ( Andersen, S., Olesen, K., Jensen, F., and Jensen, F. )
  • An Introduction to Multivariate Statistical Analysis ( Anderson, T. W. )
  • Probability and Measure Theory ( Ash, R. B. )

Gambar

Figure 1.1
Figure 1.2Conditional probability distributions for O given both possible values of E, that
Figure 1.6Barchart for the probability tables of Travel conditional on Residence and
Figure 2.1
+7

Referensi

Dokumen terkait