Some of the material in is restricted to members of the community. By logging in, you may be able to gain additional access to certain collections or items. If you have questions about access or logging in, please use the form on the Contact Page.
Tingir, S. (2019). Evaluating the Effectiveness of the Expectation-Maximization (EM) Algorithm for Bayesian Network Calibration. Retrieved from http://purl.flvc.org/fsu/fd/2019_Summer_Tingir_fsu_0071E_15106
Educators use various statistical techniques to explain relationships between latent and observable variables. One way to model these relationships is to use Bayesian networks as a scoring model. However, adjusting the conditional probability tables (CPT-parameters) to fit a set of observations is still a challenge when using Bayesian networks. A CPT provides the conditional probabilities of a single discrete variable with respect to other discrete variables. In general Bayesian networks, the CPTs that link the proficiency variable and observable outcomes are not necessarily monotonic, but they are often constrained to be monotonic in educational applications. The monotonicity constraint states that if an examinee shows an improvement on a proficiency variable (parent variable), the individual performance on an observable (child variable) should improve. For example, if a student has a higher writing skill, then this student is likely to score better on an essay task. For educational research, building parametric models (i.e., DiBello models) with the Expectation-Maximization algorithm provides monotonic conditional probability tables (CPT). This dissertation explored the effectiveness of the EM algorithm within the DiBello parameterization under different sample sizes, test forms, and item structures. The data generation model specifies two skill variables with a different number of items depending on the test forms. The outcome measures were the relative bias of the parameters to assess parameter recovery, Kullback-Leibler distance to evaluate the distance between CPTs, and Cohen's κ to assess classification agreement between data generation and estimation models. The simulation study results showed that a minimum sample size of 400 was sufficient to produce acceptable parameter bias and KL distance. A balanced distribution of simple and integrated type items produced less bias compared to an unbalanced item distribution. The parameterized EM algorithm stabilized the estimates for cells small sizes in CPTs, providing minimal KL distance values. However, the classification agreement between generated and estimated models was low.
A Dissertation submitted to the Department of Educational Psychology and Learning Systems in partial fulfillment of the requirements for the degree of Doctor of Philosophy.
Bibliography Note
Includes bibliographical references.
Advisory Committee
Russell G. Almond, Professor Directing Dissertation; Debajyoti Sinha, University Representative; Betsy J. Becker, Committee Member; Yanyun Yang, Committee Member.
Publisher
Florida State University
Identifier
2019_Summer_Tingir_fsu_0071E_15106
Tingir, S. (2019). Evaluating the Effectiveness of the Expectation-Maximization (EM) Algorithm for Bayesian Network Calibration. Retrieved from http://purl.flvc.org/fsu/fd/2019_Summer_Tingir_fsu_0071E_15106