Source code for pgmpy.inference.ApproxInference

from pgmpy.models import BayesianNetwork, DynamicBayesianNetwork
from pgmpy.factors.discrete import DiscreteFactor

[docs]class ApproxInference(object): def __init__(self, model): """ Initializes the Approximate Inference class. Parameters ---------- model: Instance of pgmpy.models.BayesianNetwork or pgmpy.models.DynamicBayesianNetwork Examples -------- >>> from pgmpy.utils import get_example_model >>> model = get_example_model('alarm') >>> infer = ApproxInference(model) """ if not isinstance(model, (BayesianNetwork, DynamicBayesianNetwork)): raise ValueError( f"model should either be a Bayesian Network or Dynamic Bayesian Network. Got {type(model)}." ) model.check_model() self.model = model @staticmethod def _get_factor_from_df(df): """ Takes a groupby dataframe and converts it into a pgmpy.factors.discrete.DiscreteFactor object. """ variables = list(df.index.names) state_names = {var: list(df.index.unique(var)) for var in variables} cardinality = [len(state_names[var]) for var in variables] return DiscreteFactor( variables=variables, cardinality=cardinality, values=df.values, state_names=state_names, )
[docs] def get_distribution(self, samples, variables, joint=True): """ Computes distribution of `variables` from given data `samples`. Parameters ---------- samples: pandas.DataFrame A dataframe of samples generated from the model. variables: list (array-like) A list of variables whose distribution needs to be computed. joint: boolean If joint=True, computes the joint distribution over `variables`. Else, returns a dict with marginal distribution of each variable in `variables`. """ if joint == True: return self._get_factor_from_df( samples.groupby(variables).size() / samples.shape[0] ) else: return { var: self._get_factor_from_df( samples.groupby([var]).size() / samples.shape[0] ) for var in variables }
[docs] def query( self, variables, n_samples=int(1e4), evidence=None, virtual_evidence=None, joint=True, show_progress=True, ): """ Method for doing approximate inference based on sampling in Bayesian Networks and Dynamic Bayesian Networks. Parameters ---------- variables: list List of variables for which the probability distribution needs to be calculated. n_samples: int The number of samples to generate for computing the distributions. Higher `n_samples` results in more accurate results at the cost of more computation time. evidence: dict (default: None) The observed values. A dict key, value pair of the form {var: state_name}. virtual_evidence: list (default: None) A list of pgmpy.factors.discrete.TabularCPD representing the virtual/soft evidence. show_progress: boolean (default: True) If True, shows a progress bar when generating samples. Returns ------- Probability distribution: pgmpy.factors.discrete.TabularCPD The queried probability distribution. Examples -------- >>> from pgmpy.utils import get_example_model >>> from pgmpy.inference import ApproxInference >>> model = get_example_model("alarm") >>> infer = ApproxInference(model) >>> infer.query(variables=["HISTORY"]) <DiscreteFactor representing phi(HISTORY:2) at 0x7f92d9f5b910> >>> infer.query(variables=["HISTORY", "CVP"], joint=True) <DiscreteFactor representing phi(HISTORY:2, CVP:3) at 0x7f92d9f77610> >>> infer.query(variables=["HISTORY", "CVP"], joint=False) {'HISTORY': <DiscreteFactor representing phi(HISTORY:2) at 0x7f92dc61eb50>, 'CVP': <DiscreteFactor representing phi(CVP:3) at 0x7f92d915ec40>} """ # Step 1: Generate samples for the query samples = self.model.simulate( n_samples=n_samples, evidence=evidence, virtual_evidence=virtual_evidence, show_progress=show_progress, ) # Step 2: Compute the distributions and return it. return self.get_distribution(samples, variables=variables, joint=joint)