Zheng Zhang, 27.05.2010 Exact Matching slide 1 Outline Graph matching – Definition – Manifold application areas Exact matching – Definition – Different.

Slides:



Advertisements
Similar presentations
Introductory Mathematics & Statistics for Business
Advertisements

VBM Voxel-based morphometry
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Bayesian models for fMRI data
Midwestern State University Department of Computer Science Dr. Ranette Halverson CMPS 2433 – CHAPTER 4 GRAPHS 1.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
OverviewOverview Motion correction Smoothing kernel Spatial normalisation Standard template fMRI time-series Statistical Parametric Map General Linear.
Zheng Zhang, Exact Matchingslide 1 Exact (Graph) Matching Zheng Zhang e Presentation for VO Structural Pattern Recognition.
Outline Graph matching Definition Manifold application areas
CPSC 322, Lecture 9Slide 1 Search: Advanced Topics Computer Science cpsc322, Lecture 9 (Textbook Chpt 3.6) January, 23, 2009.
Chapter Seventeen HYPOTHESIS TESTING
Branch and Bound Similar to backtracking in generating a search tree and looking for one or more solutions Different in that the “objective” is constrained.
MAE 552 – Heuristic Optimization Lecture 27 April 3, 2002
MAE 552 – Heuristic Optimization Lecture 26 April 1, 2002 Topic:Branch and Bound.
Analysis of Algorithms CS 477/677
SubSea: An Efficient Heuristic Algorithm for Subgraph Isomorphism Vladimir Lipets Ben-Gurion University of the Negev Joint work with Prof. Ehud Gudes.
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Fundamentals of Hypothesis Testing: One-Sample Tests
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Physical Mapping of DNA Shanna Terry March 2, 2004.
Statistical Analysis A Quick Overview. The Scientific Method Establishing a hypothesis (idea) Collecting evidence (often in the form of numerical data)
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Statistical Decision Theory
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
METHODSDUMMIES BAYES FOR BEGINNERS. Any given Monday at pm “I’m sure this makes sense, but you lost me about here…”
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Significance Tests: THE BASICS Could it happen by chance alone?
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Coregistration and Spatial Normalisation
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Computational Intelligence II Lecturer: Professor Pekka Toivanen Exercises: Nina Rogelj
Bayes for Beginners Reverend Thomas Bayes ( ) Velia Cardin Marta Garrido.
Computational Intelligence: Methods and Applications Lecture 12 Bayesian decisions: foundation of learning Włodzisław Duch Dept. of Informatics, UMK Google:
Bayesian Inference and Posterior Probability Maps Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course,
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Confidence intervals and hypothesis testing Petter Mostad
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
1 Chapter 9 Hypothesis Testing. 2 Chapter Outline  Developing Null and Alternative Hypothesis  Type I and Type II Errors  Population Mean: Known 
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Lecture: Forensic Evidence and Probability Characteristics of evidence Class characteristics Individual characteristics  features that place the item.
Populations III: evidence, uncertainty, and decisions Bio 415/615.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
AP Statistics Section 11.1 B More on Significance Tests.
Problem Reduction So far we have considered search strategies for OR graph. In OR graph, several arcs indicate a variety of ways in which the original.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Chapter 13 Backtracking Introduction The 3-coloring problem
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Introduction to Graph Theory By: Arun Kumar (Asst. Professor) (Asst. Professor)
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:
The NP class. NP-completeness Lecture2. The NP-class The NP class is a class that contains all the problems that can be decided by a Non-Deterministic.
Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Zurich, February 2008 Bayesian Inference.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Hypothesis Tests for 1-Proportion Presentation 9.
BackTracking CS255.
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Structural Patten Recognition Instead of describing an object in terms of a feature vector, describe it by its structure Describe complicated objects.
Lecture: Forensic Evidence and Probability Characteristics of evidence
Elements of a statistical test Statistical null hypotheses
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Presentation transcript:

Zheng Zhang, Exact Matching slide 1 Outline Graph matching – Definition – Manifold application areas Exact matching – Definition – Different forms of exact matching – Necessary concepts Algorithms for exact matching – Techniques based on tree search – Other techniques Conclusion

Zheng Zhang, Exact Matchingslide 2 Graph matching Definition: –Graph matching is the process of finding a correspondence between the nodes and the edges of two graphs that satisfies some (more or less stringent) constraints ensuring that similar substructures in one graph are mapped to similar substructures in the other. [D. Conte, Pasquale Foggia, Carlo Sansone, and Mario Vento. Thirty Years of Graph Matching in Pattern Recognition.] in comparison with matching (graph theory): –A matching or independent edge set in a graph is a set of edges without common vertices. It may also be an entire graph consisting of edges without common vertices. [Wikipedia]

Zheng Zhang, Exact Matching slide 3 Graph matching Manifold application areas – 2D and 3D image analysis – Document processing – Biometric identification – Image databases – Video analysis – Biological and biomedical applications Categories of matching – Exact matching – Inexact matching – Other matching problems

Zheng Zhang, Exact Matching slide 4 Exact matching Definition – Exact graph matching is characterized by the fact that the mapping between the nodes of the two graphs must be edge-preserving in the sense that if two nodes in the first graph are linked by an edge, they are mapped to two nodes in the second graph that are linked by an edge as well. Different forms of exact matching – The most stringent form: graph isomorphism – A weaker form: subgraph isomorphism – A slightly weaker form: monomorphism – A still weaker form: homomorphism – Another interesting form: maximum common subgraph (MCS)

Zheng Zhang, Exact Matchingslide 5 Necessary concepts Morphism – A morphism is an abstraction derived from structure-preserving mappings between two mathematical structures. – In comparison with homomorphism: a homomorphism is a structure- preserving map between two mathematical structures. – If a morphism f has domain X and codomain Y, we write f :X→Y. Thus a morphism is represented by an arrow from its domain to its codomain. The collection of all morphisms from X to Y is denoted Hom(X,Y) or Mor(X,Y). Isomorphism – An isomorphism is a bijective homomorphism. – An isomorphism is a morphism f :X→Y in a category for which there exists an "inverse" f −1 :Y→X, with the property that both f −1 ·f = id X and f·f −1 =id Y.

Zheng Zhang, Exact Matchingslide 6 Necessary concepts Epimorphism – an surjective homomorphism Monomorphism – an injective homomorphism Endomoprphism – a homomorphism from an object to itself Automoprphism – an endomorphism which is also an isomorphism an isomorphism with itself

Zheng Zhang, Exact Matchingslide 7 Different forms of exact matching Graph isomorphism – A one-to-one correspondence must be found between each node of the first graph and each node of the second graph. – Graphs G(V G,E G ) and H(V H,E H ) are isomorphic if there is an invertible F from V G to V H such that for all nodes u and v in V G, (u,v) ∈ E G if and only if (F(u),F(v)) ∈ E H.

Zheng Zhang, Exact Matchingslide 8 Different forms of exact matching Subgraph isomorphism – It requires that an isomorphism holds between one of the two graphs and a node-induced subgraph of the other. Monomorphism – It requires that each node of the first graph is mapped to a distinct node of the second one, and each edge of the first graph has a corresponding edge in the second one; the second graph, however, may have both extra nodes and extra edges.

Zheng Zhang, Exact Matchingslide 9 Different forms of exact matching Homomorphism – It drops the condition that nodes in the first graph are to be mapped to distinct nodes of the other; hence, the correspondence can be many-to-one. – A graph homomorphism F from Graph G(V G,E G ) and H(V H,E H ), is a mapping F from V G to V H such that {x,y} ∈ E G implies {F(x),F(y)} ∈ E H.

Zheng Zhang, Exact Matchingslide 10 Different forms of exact matching Maximum common subgraph (MCS) – A subgraph of the first graph is mapped to an isomorphic subgraph of the second one. – There are two possible definitions of the problem, depending on whether node-induced subgraphs or plain subgraphs are used. – The problem of finding the MCS of two graphs can be reduced to the problem of finding the maximum clique (i.e. a fully connected subgraph) in a suitably defined association graph.

Zheng Zhang, Exact Matchingslide 11 Different forms of exact matching Properties – The matching problems are all NP-complete except for graph isomorphism, which has not yet been demonstrated whether in NP or not. – Exact graph matching has exponential time complexity in the worst case. However, in many PR applications the actual computation time can be still acceptable. – Exact isomorphism is very seldom used in PR. Subgraph isomorphism and monomorphism can be effectively used in many contexts. – The MCS problem is receiving much attention.

Zheng Zhang, Exact Matchingslide 12 Algorithms for exact matching Techniques based on tree search – mostly based on some form of tree search with backtracking Ullmann’s algorithm Ghahraman’s algorithm VF and VF2 algorithm Bron and Kerbosh’s algorithm Other algorithms for the MCS problem Other techniques – based on A* algorithm Demko’s algorithm – based on group theory Nauty algorithm

Zheng Zhang, Exact Matching slide 18 Techniques based on tree search The basic idea – A partial match (initially empty) is iteratively expanded by adding to it new pairs of matched nodes. – The pair is chosen using some necessary conditions, usually also some heuristic condition to prune unfruitful search paths. – Eventually, either the algorithm finds a complete matching, or no further vertex pairs may be added (backtracking) – For PR the algorithm may consider the attributes of nodes and edges in constraining the desired matching.

Zheng Zhang, Exact Matchingslide 19 Techniques based on tree search The backtracking algorithm – depth-first search(DFS): it progresses by expanding the first child node of the search tree that appears and thus going deeper and deeper until a goal node is found, or until it hits a node that has no children. – Branch and bound(B&B): it is a BFS(breadth-first search)-like search for the optimal solution. Branch is that a set of solution candidates is splitted into two or more smaller sets; bound is that a procedure upper and lower bounds.

Zheng Zhang, Exact Matchingslide 20 Techniques based on tree search Ullmann’s algorithm – Probably the most popular graph matching algorithm – Application for graph isomorphism, subgraph isomorphism and monomorphism, also for MCS problem – A refinement procedure based on matrix of possible future matched node pairs to prune unfruitful matches – The simple enumeration algorithm for the isomorphisms between a graph G and a subgraph of another graph H with the adjacency matrices A G and A H An M’ matrix with |V G | rows and |V H | columns can be used to permute the rows and columns of A H to produce a further matrix P. If, then M’ specifies an isomorphism between G and the subgraph of H. [J.R.Ullmann.An algorithm for subgraph isomorphism.]

Zheng Zhang, Exact Matchingslide 21 Techniques based on tree search Ullmann’s algorithm – Example for permutation matrix – The elements of M’ are 1’s and 0’s, such that each row contains 1 and each column contains 0 or 1

Zheng Zhang, Exact Matchingslide 22 Techniques based on tree search Ullmann’s algorithm – Construction of another matrix M (0) with the same size of M’ – Generation of all M’ by setting all but one of each row of M (0) – A subgraph isomorphism has been found if

Zheng Zhang, Exact Matchingslide 23 Techniques based on tree search Ullmann’s algorithm – Example

Zheng Zhang, Exact Matchingslide 24 Techniques based on tree search Ghahraman’s algorithm – Application for monomorphism – Use of the netgraph obtained from the Cartesian product of the nodes of two graph, monomorphisms correspond to particular subgraphs of the netgraph – A strong necessary condition and a weak one, then two versions of the algorithm to detect unfruitful partial solutions VF and VF2 algorithm – Application for isomorphism and subgraph isomorphism – VF algorithm defines a heuristic based on the analysis of the sets of nodes adjacent to the ones already considered in the partial mapping. – VF2 algorithm reduces the memory requirement from O(N 2 ) to O(N).

Zheng Zhang, Exact Matchingslide 25 Techniques based on tree search Bron and Kerbosh’s algorithm – Application for the clique detection and the MCS problem – Based on the use of a heuristic for pruning the search tree – Simplicity and an acceptable performance in most cases Other algorithms for the MCS problem – Balas and Yu’s alogrithm also defines a heristic, but based on graph colouring techniques. – McGregor’s alogrithm is not applicated for a maximum clique problem. – Koch’s alogrithm is applicated for a slightly simplified version of the MCS problem and suggests the use of the Bron and Kerbosh’s algorithm.

Zheng Zhang, Exact Matchingslide 26 Other techniques Based on the A* algorithm – It uses a distance-plus-cost heuristic function to determine the order in which the search visits nodes in the tree. The heuristic is a sum of two functions: the path-cost function, i.e. the cost from the starting node to the current node, and an admissible “heuristic estimate” of the distance to the goal. – Demko’s algorithm investigates a generalization of MCS to hypergraphs. Based on group theory – McKay’s Nauty(No automorphisms yes?) algorithm deals only with the isomorphism problem. It constructs the automorphism group of each of the input graphs and derives a canonical labeling. The isomorphism can be checked by verifying the equality of the adjacency matrices.

Zheng Zhang, Exact Matching slide 27 Conclusion The exact graph matching problem is of interest in a variety of different pattern recognition contexts. Of the exact graph matching problems, exact isomorphism is very seldom used in PR while subgraph isomorphism, monomorphism and the MCS problem are popularly proposed. Ullmann’ algorithm, VF2 algorithm and Nauty algorithm are mostly used algorithms, which based on the search methods, and may outperform others. Most modified algorithms adopt some conditions to prune the unfruitful partial matching.

Bayes for Beginners Reverend Thomas Bayes ( )

Overview of SPM Analysis Motion Correction Smoothing Spatial Normalisation General Linear Model Statistical Parametric Map fMRI time-series Parameter Estimates Design matrix Anatomical Reference

Template image Affine registration. (  2 = 472.1) Non-linear registration without Bayes constraints. (  2 = 287.3) Non-linear registration using Bayes. (  2 = 302.7) Without Bayesian constraints, the non- linear spatial normalisation can introduce unnecessary warps. Spatial Normalisation - Overfitting

Bayes and image segmentation Want to automatically classify regions into grey matter, white matter, CSF and non-brain tissue. How do we use prior information? (probabilities of each voxel being of a particular type derived from a database of 152 scans) Bayes! Want to automatically classify regions into grey matter, white matter, CSF and non-brain tissue. How do we use prior information? (probabilities of each voxel being of a particular type derived from a database of 152 scans) Bayes!

A big advantage of a Bayesian approach Allows a principled approach to the exploitation of all available data … with an emphasis on continually updating one’s models as data accumulate as seen in the consideration of what is learned from a positive mammogram

Bayesian Reasoning - Casscells, Schoenberger & Grayboys, Eddy, Gigerenzer & Hoffrage, 1995, Butterworth, Hoffrage, Lindsey, Hertwig & Gigerenzer, 2001 When PREVALENCE, SENSITIVITY, and FALSE POSITIVE RATES are given, most physicians misinterpret the information in a way that could be potentially disastrous for the patient.

Bayesian Reasoning ASSUMPTIONS 1% of women aged forty who participate in a routine screening have breast cancer 80% of women with breast cancer will get positive tests 9.6% of women without breast cancer will also get positive tests EVIDENCE A woman in this age group had a positive test in a routine screening PROBLEM What’s the probability that she has breast cancer?

Bayesian Reasoning ASSUMPTIONS 10 out of 1000 women aged forty who participate in a routine screening have breast cancer 800 out of 1000 of women with breast cancer will get positive tests 95 out of 1000 women without breast cancer will also get positive tests PROBLEM If 1000 women in this age group undergo a routine screening, about what fraction of women with positive tests will actually have breast cancer?

Bayesian Reasoning ASSUMPTIONS 100 out of 10,000 women aged forty who participate in a routine screening have breast cancer 80 of every 100 women with breast cancer will get positive tests 950 out of 9,900 women without breast cancer will also get positive tests PROBLEM If 10,000 women in this age group undergo a routine screening, about what fraction of women with positive tests will actually have breast cancer?

Bayesian Reasoning Before the screening: 100 women with breast cancer 9,900 women without breast cancer After the screening: A = 80 women with breast cancer and positive test B = 20 women with breast cancer and negative test C = 950 women without breast cancer and positive test D = 8,950 women without breast cancer and negative test Proportion of cancer patients with positive results, within the group of ALL patients with positive results: A/(A+C) = 80/(80+950) = 80/1030 = = 7.8%

Compact Formulation C = cancer present, T = positive test p(A|B) = probability of A, given B, ~ = not PRIOR PROBABILITY p(C) = 1% CONDITIONAL PROBABILITIES p(T|C) = 80% p(T|~C) = 9.6% POSTERIOR PROBABILITY (or REVISED PROBABILITY) p(C|T) = ? PRIORS

Bayesian Reasoning Before the screening: 100 women with breast cancer 9,900 women without breast cancer After the screening: A = 80 women with breast cancer and positive test B = 20 women with breast cancer and negative test C = 950 women without breast cancer and positive test D = 8,950 women without breast cancer and negative test Proportion of cancer patients with positive results, within the group of ALL patients with positive results: A/(A+C) = 80/(80+950) = 80/1030 = = 7.8%

Bayesian Reasoning Prior Probabilities: 100/10,000 = 1/100 = 1% = p(C) 9,900/10,000 = 99/100 = 99% = p(~C) Conditional Probabilities: A = 80/10,000 = (80/100)*(1/100) = p(T|C)*p(C) = B = 20/10,000 = (20/100)*(1/100) = p(~T|C)*p(C) = C = 950/10,000 = (9.6/100)*(99/100) = p(T|~C)*p(~C) = D = 8,950/10,000 = (90.4/100)*(99/100) = p(~T|~C) *p(~C) = Rate of cancer patients with positive results, within the group of ALL patients with positive results: A/(A+C) = 0.008/( ) = 0.008/0.103 = = 7.8%

-----> Bayes’ theorem p(T|C)*p(C) p(C|T) = ______________________ P(T|C)*p(C) + p(T|~C)*p(~C) A A + C

Comments Common mistake: to ignore the prior probability The conditional probability slides the revised probability in its direction but doesn’t replace the prior probability A NATURAL FREQUENCIES presentation is one in which the information about the prior probability is embedded in the conditional probabilities (the proportion of people using Bayesian reasoning rises to around half). Test sensitivity issue (or: “if two conditional probabilities are equal, the revised probability equals the prior probability”) Where do the priors come from?

-----> Bayes’ theorem p(X|A)*p(A) p(A|X) = ______________________ P(X|A)*p(A) + p(X|~A)*p(~A) Given some phenomenon A that we want to investigate, and an observation X that is evidence about A, we can update the original probability of A, given the new evidence X.

It relates the conditional density of a parameter (posterior probability) with its unconditional density (prior, since depends on information present before the experiment). The likelihood is the probability of the data given the parameter and represents the data now available. Bayes’ Theorem for a given parameter  p (  data) = p (data  ) p (  ) / p (data) 1/P (data) is basically a normalizing constant Posterior  likelihood x prior The prior is the probability of the parameter and represents what was thought before seeing the data. The posterior represents what is thought given both prior information and the data just seen.

Likelihood:p(y|  ) =  (M d, d -1 ) Prior:p(  ) =  (M p, p -1 ) Posterior: p(  y) ∝ p(y|  )*  p(  ) =  (M post, post -1 ) MpMp p -1 M post post -1 d -1 MdMd post = d + p M post = d M d + p M p  post Posterior Probability Distribution precision = 1/  2

Activations in fMRI…. Classical – ‘What is the likelihood of getting these data given no activation occurred?’ Bayesian option (SPM2) –‘What is the chance of getting these parameters, given these data?

What use is Bayes in deciding what brain regions are active in a particular study? Problems with classical frequentist approach – All inferences relate to disproving the null hypothesis – Never fully reject H 0, only say that the effect you see is unlikely to occur by chance – Corrections for multiple comparisons significance depends on the number of contrasts you look at – Very small effects can be declared significant with enough data Bayesian Inference offers a solution through Posterior Probability Maps (PPMs)

SPMs and PPMs PPMs: Show activations of a given size SPMs: show voxels with non-zero activations

PPMs Advantages Disadvantages One can infer a cause DID NOT elicit a response SPMs conflate effect-size and effect-variability Computationally demanding (priors are determined empirically) Utility of Bayesian approach is yet to be established

Frequentist vs. Bayesian by Berry 1. Probabilities of data vs. probabilities of parameters (& also data). 2. Evidence used: – Frequentist measures specific to experiment. – Posterior distribution depends on all available information. Makes Bayesian approach appealing, but assembling, assessing, & quantifying information is work.

3. Depend on probabilities of results that could occur vs. did occur: – Frequentist measures (e.g., p values, confidence intervals) incorporate probabilities of data that were possible but did not occur. – Posterior depends on data only through the likelihood, which is calculated from observed data. 4. Flexibility: – Frequentist measures depend on design; require that design be followed. – Bayesian view: update continually as data accumulate (only requirement is honesty). Sample size: need not choose in advance. Weigh costs/benefits; decide whether to start experiment. After experiment starts, decide whether to continue—stop at any time, for any reason.

5. Decision making – Frequentist: historically avoided. – Bayesian: tailored to decision analysis; losses and gains considered explicitly.

Statistics: A Bayesian Perspective D. Berry, 1996, Duxbury Press. – excellent introductory textbook, if you really want to understand what it’s all about. – “Using a Bayesian Approach in Medical Device Development”, also by Berry – a powerpoint presentation by Berry – Extremely clear presentation of the mammography example; highly polemical and fun too! – Bayes’ original essay Jaynes, E. T., 1956, `Probability Theory in Science and Engineering,' (No. 4 in `Colloquium Lectures in Pure and Applied Science,' Socony-Mobil Oil Co. USA. – A physicist’s take on Bayesian approaches. Proposes an interesting metric of probability using decibels (yes, the unit used for sound levels!). – a skeptical account of Bayesian approaches. The rest of the site is very informative and sensible about basic statistical issues. References

Bayes’ ending Bunhill Fields Burial Ground off City Road, EC1