Assuming all variables you mention are categorical and the edge directions are from up to down:
Priors:
In the first Naive Bayes example, the conditional probability table (CPT) of 'class' consists solely of its prior distribution because it is a root node, i.e. does not have any parents. If 'class' can take on 2 states (e.g. black and white), its CPT will consist of 2 values.
In the second Bayesian Network (BN) example, the CPT of 'class' is dependent on 'cause1' and 'consequence'. Lets say 'consequence' has 3 states, 'cause1' has 4 states and as before 'class' has 2 states. In this case, the CPT of 'class' would contain 3*4*2 values. When you are learning this CPT, you can incorporate your prior beliefs as a dirichlet distribution (if all variables are categorical). For an example of how to incorporate your prior beliefs into a Maximum Likelihood Estimation process, have a look at these excellent lecture slides.
Inference: (or what you refer to as 'classification')
As per carrying out the classification, in Example 1, you can make use of the Bayes Rule to calculate P('class' = white) and P('class' = black). In the second (BN) example, you would have to use a belief propagation or variable elimination or junction tree algorithm to update the posterior probabilities of the 'class' node based on your observed nodes.
There is a straight-forward BNT example on how to accomplish this here. Plus, BNT toolbox comes with brief 'inference' examples that use the junction tree function, which you can find under the .../examples
folder.
Finally -some people may disagree but- as far as BNs go, I would advise against strictly interpreting A -> B as "A causes B" since the causality aspect of BNs, especially in the field of structure learning, is open to much debate.
I hope this helps.