MoTBFs (1.2)

0 users

Learning Hybrid Bayesian Networks using Mixtures of Truncated Basis Functions.

Learning, manipulation and evaluation of mixtures of truncated basis functions (MoTBFs), which include mixtures of polynomials (MOPs) and mixtures of truncated exponentials (MTEs). MoTBFs are a flexible framework for modelling hybrid Bayesian networks (I. Prez-Bernab, A. Salmern, H. Langseth (2015) ; H. Langseth, T.D. Nielsen, I. Prez-Bernab, A. Salmern (2014) ; I. Prez-Bernab, A. Fernndez, R. Rum, A. Salmern (2016) ). The package provides functionality for learning univariate, multivariate and conditional densities, with the possibility of incorporating prior knowledge. Structural learning of hybrid Bayesian networks is also provided. A set of useful tools is provided, including plotting, printing and likelihood evaluation. This package makes use of S3 objects, with two new classes called 'motbf' and 'jointmotbf'.

Maintainer: Ana D. Maldonado
Author(s): Inmaculada Prez-Bernab, Antonio Salmern, Thomas D. Nielsen, Ana D. Maldonado

License: LGPL-3

Uses: bnlearn, ggm, lpSolve, quadprog

Released 5 days ago.

2 previous versions



  (0 votes)


  (0 votes)

Log in to vote.


No one has written a review of MoTBFs yet. Want to be the first? Write one now.

Related packages:(20 best matches, based on common tags.)

Search for MoTBFs on google, google scholar, r-help, r-devel.

Visit MoTBFs on R Graphical Manual.