Australasian Bayesian Network Modelling Society 2014: Call for Abstracts

We will be meeting in Rotorua, New Zealand, 26-27 November and welcome abstracts from anyone in the Bayesian network community, whether in applied sciences, government, industry or academic research. See

http://www.abnms.org/conferences/abnms2014/

Note that our conference is preceded by a two-day tutorial program covering Bayesian networks, including introduction, elicitation, GIS integration, OOBNs, sensitivity analysis and automated learning.

Reaction to the DTCA Begins

The DTCA is set to take effect at the beginning of 2015, criminalizing much of the research into high tech and applied science in Australia, whether conducted in universities or industry. I have summarized its main features and impacts on the university sector (see "Australia's act of Intellectual Terrorism" below). For impacts on industry see the website victimsofdsto.

I will be giving a presentation on DTCA at the Melbourne Future Day, 1 March.

Notes on Forecasting

— Yung En Chee, University of Melbourne

 

[Editor's note: These ideas were prepared by Yung in support of her participation in a forecasting project and describe techniques she found helpful for predicting specific world events within specific time frames. The context is one where specific possible outcomes of political, economic, etc. events were described, and probability ranges for those outcomes had to be supplied within a few days. The requirement was to provide predictions as accurate and precise as possible without succumbing to over- or under-confidence. I think these ideas are of general interest.]

 

My list of tools and strategies is a mash-up of ideas from other people, including Daniel Kahneman (Thinking Fast, Thinking Slow), a couple of bloggers who were on the Good Judgement team - ‘Dart-Throwing Chimp’ (Jay Ulfelder) and ‘Morendil’ (Laurent Bossavit) and personal reflection. See the following Morendil posts on lesswrong:

Bossavit's tools for prediction:

  1. Favour the status quo (this seems to me more like an empirically derived tip – if you do this you’ll come out ahead in the long run)
  2. Flip the question around
  3. Use reference classes
  4. Prepare lines of retreat (what would make me change my mind about this?) [this is equivalent to think of reasons/evidence that make it unlikely – a sort of correction for overconfidence]
  5. Abandon sunk costs
  6. Consider your loss function (this is more about strategic hedging in response to how the Brier score is computed)

These ideas are, in part, about the psychological aspects of making the prediction, but he doesn’t discuss strategies for finding information, reconciling or evaluating information, forming mental models etc.

Tools I use:

  1. Understand the conditions required for resolution of question.
  2. Iteration (rapid prototyping). Search and scan articles quickly to build up some mental model of what the situation is, who the actors are and understand any processes required for resolution of the question (point 3 below) – go quick, build a crude mental model then refine as more material is encountered/assimilated.
  3. Understand any processes required for resolution of question. E.g. standard operating protocol for declaration of disease status of OIE member (World Organization of Animal Health); process of granting of EU member candidacy; voting and veto powers re UN Security Council, process for obtaining full negotiating member status at TPP (Trans-Pacific Partnership), etc.
  4. If a question involves non-English speaking players make sure to search beyond the elite news sources (e.g. BBCNews, Guardian, Newsdaily, New York Times, Economist, WSJ) for primarily English speakers (e.g. Al-Jazeera, Al-Monitor, Kurdistan news, Ahram online etc.).
  5. Consider whether the uncertainty associated with the question is reducible or irreducible. Uncertainty may be reducible if it’s due to a lack of knowledge or understanding. If I suspect this is the case, I try to find a knowledgeable interpreter/analyst or try to construct Reference Classes and identify plausible Base Rates. Indicators of knowledge: knowledge of historical context, logical argumentation, ability to articulate reasons, provide reasons supported by evidence, ability to recognise and acknowledge where and when things aren’t known (e.g. Aurelan George Mulgan on Japan TPP; Reidar Visser on Iran). If the uncertainty is irreducible (and I consider things like oil, stock and currency prices to fall in this category) then don’t bother looking for more info.
  6. Use Reference Classes, try to estimate Base Rate from any available data and use that (see Kahneman: Thinking Fast, Thinking Slow).
  7. Use Bueno de Mesquita’s ‘factors’ (Position, Influence, Salience, Flexibility/Resolve, see below) to think through the interests of actors involved.
  8. Understand the status quo and consider favouring the status quo (empirically, we would expect the status quo to persist).
  9. Another empirically robust finding from interest group politics: well-organised private interests nearly always win out over diffuse, generally disorganized public interests.
  10. Try and articulate Reasons that would make an event LIKELY and and Reasons that would be it UNLIKELY (this requires adequate research or analysis or reasoning to furnish or construct reasons; will often rely on tally for informing prediction).
  11. Keep looking out for new info till just before the deadline and update estimates accordingly.
  12. Consciously consider our tendency towards What You See IS All There Is (WYSIATI) – this tendency is misleading! Remember that the capacity for unpredictable, unforeseeable events or chains of events is ever present, particularly with long time lines and adjust forecasts accordingly (e.g. UN Sec Council resolution on Mali).

Note: Bruce Bueno de Mesquita is a political scientist who claims to have built a system using Game Theory for making predictions (http://www.predictioneersgame.com/) I think he overreaches and he’s been criticised for not being transparent about his methods. I agree. Nevertheless, I find his key factors for assessing actor interests, stakes, power and influence useful.

  1. Relative potential influence - the relative potential ability of each player to persuade other stakeholders to adjust their approach to the issue to be more in line with the influencer’s perspective. Resources - the potential ability each player has to persuade other stakeholders to support a point of view on the issue in question. The ability to persuade may be derived from holding a position of authority, being an expert, commanding a large budget, or any other factor that makes others listen to someone.
  2. Policy position - position preferred by each stakeholder on the issue, taking constraints into account. This position is not likely to be the outcome the stakeholder expects or is prepared to accept, nor is it likely to be what the player wants in his or her heart of hearts. It is the position the stakeholder favors or advocates within the context of the situation. When a player’s position has not been articulated, it is best thought of as the answer to the following mind experiment: If the stakeholder were asked to write down his or her position, without knowing the values being written down by other stakeholders, what would he or she write down as the position he or she prefers on the issue continuum? To place a numeric value on the position, the investigator must first have defined the issue continuum. The continuum will either have a natural numeric interpretation, such as the percentage of uninsured on health care to be covered under a new policy or the analyst will need to develop numeric values that reflect the relative degree of difference across policy stances that are not inherently quantitative. It is important that the numerical values assigned to different positions (and they can range between any values) reflect the relative distance or proximity of the different solutions to one another.  An easy way to turn player preferences on an issue into numeric values is to place each player on the issue continuum you defined, locating then at the point on the continuum that reflects the policy they support. Then, use a ruler to measure how far each player is from one end of the line that reflects the range of choices. Let the left-hand end of the line equal 0. Then each other point on the line is simply its distance from 0 on the ruler.
  3. Salience - assesses how focused a stakeholder is on the issue. Its value is best thought of in terms of how prepared the stakeholder is to work on the issue when it comes up rather than some other issue on his or her plate. Would the stakeholder drop everything else to deal with the issue? Would the player work on it on a weekend day, come back from vacation, etc.? The more confidently it can be said that this issue takes priority over other matters in the stakeholder’s professional life (or personal life if the issue is about a personal or family matter), the higher the salience value.
  4. Flexibility/Resolve - evaluates the stakeholder’s preference for reaching an agreement as compared to sticking to his or her preferred position even if it means failing to reach an agreement.

BN Training Workshops

We are pleased to announce that Bayesian Intelligence will again be running a series of Bayesian network training workshops in Melbourne this year.

There will be six days of workshops, with the introductory BN training running in both April and June:

  • April 4th: Introduction to BNs
  • April 5th: More on BNs
  • June 27th: Introduction to BNs
  • June 28th: More on BNs
  • Sep 26th: Programming BN solutions with Netica (the basics)
  • Sep 27th: Programming BN solutions with Netica (advanced topics)

People are invited to register for any combination of the training days that best suits their background in BNs and their interests.

For more information, schedules and registration, please visit bayesian-intelligence.com/training/, or contact Steven via email or on 0425 801 277.

UAI 9th Bayesian Modeling Applications Workshop

— Ann Nicholson (Workshop co-chair)
When: Saturday, August 18, 2012
Where: Catalina Island, California
Website: abnms.org/uai2012-apps-workshop

This single day workshop is an excellent forum for presenting and hearing about real-world applications of Bayesian networks. It follows the 28th Int. Conference on Uncertainty in AI, the premier conference for presentation of research on Bayesian technology (Aug 15-17th). The call for papers is now out, with submission deadline May 5th (with a week’s extension very likely!).

The aim of the workshop is to foster discussion and interchange about novel contributions that can speak to both the academic and the larger research community. Accordingly, we seek submissions also from practitioners and tool developers as well as researchers. We welcome submissions describing real world applications, whether as stand-alone BNs or where the BNs are embedded in a larger software system. We encourage authors to address the practical issues involved in developing real-world applications, such as knowledge engineering methodologies, elicitation techniques, defining and meeting client needs, validation processes and integration methods, as well as software tools, including visualization and user interaction techniques to support these activities.

We particularly encourage the submission of papers that address the workshop theme of temporal modeling. Recently communities building dynamic Bayes networks (DBNs) and partially observable MDPs (POMDPs) are coming to realize that they are applying their methods to identical applications. Similarly POMDPs and other probabilistic methods are now established in the field of Automated Planning. Stochastic process models such as continuous time Bayes networks (CTBNs) should also be considered as part of this trend. Adaptive and on-line learning models also fit into this focus.

Ann Nicholson (Workshop co-chair)

Minimum Message Length: A Computational Bayesianism

—Lloyd Allison

Minimum message length (MML) inference is a computational implementation of Bayesian inference, an information-theoretic means of finding high posterior probability hypotheses, devised by Chris Wallace and David Boulton around 1968 (see Wallace's history of MML). MML seeks to minimise a two-part message length I(h,e) = I(e|h) + I(h), where h encodes a hypothesis and e some relevant evidence (data). So long as coding follows the principles developed by Claude Shannon, so that the codes enforce the efficiency principle that message lengths I(h) = - \log P(h), then minimising the MML message length is trivially equivalent to maximising posterior probability:

  • I(h,e) = I(e|h) + I(h)
  • - \log P(h,e) = - \log P(e|h) - \log P(h)
  • \log P(h,e) = \log [P(e|h)P(h)]
  • P(h,e) = P(e|h)P(h)

Since during this sequence we have multiplied by -1, we have also switched from minimising a message length to maximising a probability. And at the end, since P(h,e) and P(h|e) differ only by a positive multiple (see Bayes' theorem), maximising one is the same as maximising the other.

This foundation for minimum message length inference is quite elementary, so the fact that it was not in use before 1968 may be a little surprising. It is probably partly due to limits on computational capacity inhibiting Bayesian statistics and the related dominance of frequentist methods. That there remains any debate about computational Bayesianism, however, is even more surprising.

The application of Bayes' theorem is straightforward for discrete (multinomial) variables governed by a probability function. But consider a problem in which one or more variables are continuous, rather than discrete. Can Bayes' theorem apply?

  • Any continuous attribute (variable) can only be measured to some limited accuracy, \pm \epsilon /2, \epsilon > 0.
  • So, every datum that is possible under a model (theory, hypothesis) has a probability that is strictly greater than zero, and not just a probability density.
  • Any continuous parameter of a model can only be inferred (estimated) to some limited precision\pm \delta /2, \delta > 0.
  • So, every parameter estimate that is possible under a prior has a probability that is strictly greater than zero, and not just a probability density.
  • So, in continuous empirical domains both the data and the model spaces have a natural discretisation and Bayes' theorem can always be applied.

However, this is not to say that it is easy to go and make MML work in any given application; in fact it can be quite difficult. After the self evident observations above, a lot of hard work on efficient encodings, search algorithms, code books, invariance, Fisher information, fast approximations, robust heuristics, adaptations to specific problems, and all the rest, remained to be done. Fortunately, MML has been made to work in many general and useful applications, including, but not limited to:

For further information on MML you can peruse my MML web pages. For Chris Wallace's own account of MML see his book Inductive Inference by Minimum Message Length.