NASSP Masters 5003F - Computational Astronomy - 2009 Lecture 23 Two areas of my current research interest: –Parallel CLEAN. –Looking for bursts in XMM.

Slides:



Advertisements
Similar presentations
The Maximum Likelihood Method
Advertisements

Lecture 15 Orthogonal Functions Fourier Series. LGA mean daily temperature time series is there a global warming signal?
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Exponential Functions Logarithmic Functions
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Lecture 11 (Was going to be –Time series –Fourier –Bayes but I haven’t finished these. So instead:) Radio astronomy fundamentals NASSP Masters 5003F -
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
NASSP Masters 5003F - Computational Astronomy Lecture 5: source detection. Test the null hypothesis (NH). –The NH says: let’s suppose there is no.
6-1 Stats Unit 6 Sampling Distributions and Statistical Inference - 1 FPP Chapters 16-18, 20-21, 23 The Law of Averages (Ch 16) Box Models (Ch 16) Sampling.
NASSP Masters 5003F - Computational Astronomy Lecture 13 Further with interferometry – Resolution and the field of view; Binning in frequency and.
Beyond Null Hypothesis Testing Supplementary Statistical Techniques.
Theoretical Probability Distributions We have talked about the idea of frequency distributions as a way to see what is happening with our data. We have.
21-May-15 Genetic Algorithms. 2 Evolution Here’s a very oversimplified description of how evolution works in biology Organisms (animals or plants) produce.
GLAST Science Support CenterAugust 9, 2004 Likelihood Analysis of LAT Data James Chiang (GLAST SSC – SLAC)
Mr Barton’s Maths Notes
RHESSI/GOES Observations of the Non-flaring Sun from 2002 to J. McTiernan SSL/UCB.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Lecture 2: Numerical Differentiation. Derivative as a gradient
Environmental Data Analysis with MatLab Lecture 24: Confidence Limits of Spectra; Bootstraps.
Human Growth: From data to functions. Challenges to measuring growth We need repeated and regular access to subjects for up to 20 years. We need repeated.
Source detection at Saclay Look for a fast method to find sources over the whole sky Provide list of positions, allowing to run maximum likelihood locally.
Review.
Linear and generalised linear models
Linear and generalised linear models
Inferential Statistics
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Lesson 4: Percentage of Amounts.
Function approximation: Fourier, Chebyshev, Lagrange
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Orthogonal Functions Numerical Methods in Geophysics Function Approximation The Problem we are trying to approximate a function f(x) by another function.
Statistical Techniques I EXST7005 Conceptual Intro to ANOVA.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
F OUNDATIONS OF S TATISTICAL I NFERENCE. D EFINITIONS Statistical inference is the process of reaching conclusions about characteristics of an entire.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
NASSP Masters 5003F - Computational Astronomy Lecture 19 EPIC background Event lists and selection The RGA Calibration quantities Exposure calculations.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Lecture 14: Stability and Control II Reprise of stability from last time The idea of feedback control Remember that our analysis is limited to linear systems.
Modern Navigation Thomas Herring
NASSP Masters 5003F - Computational Astronomy Lecture 8 Some further unix & python tips Good programming habits Catalog cross-matching Bayesian.
Copyright © 2009 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
1 Chapter 18 Sampling Distribution Models. 2 Suppose we had a barrel of jelly beans … this barrel has 75% red jelly beans and 25% blue jelly beans.
Fourier TheoryModern Seismology – Data processing and inversion 1 Fourier Series and Transforms Orthogonal functions Fourier Series Discrete Fourier Series.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 18 Sampling Distribution Models.
A statistical test for point source searches - Aart Heijboer - AWG - Cern june 2002 A statistical test for point source searches Aart Heijboer contents:
NASSP Masters 5003F - Computational Astronomy Lecture 12: The beautiful theory of interferometry. First, some movies to illustrate the problem.
NASSP Masters 5003F - Computational Astronomy Lecture 14 Reprise: dirty beam, dirty image. Sensitivity Wide-band imaging Weighting –Uniform vs Natural.
Synthesizing Natural Textures Michael Ashikhmin University of Utah.
NASSP Masters 5003F - Computational Astronomy Lecture 12 Complex numbers – an alternate view The Fourier transform Convolution, correlation, filtering.
Sampling and estimation Petter Mostad
Week 6. Statistics etc. GRS LX 865 Topics in Linguistics.
NASSP Masters 5003F - Computational Astronomy Lecture 16 Further with interferometry – Digital correlation Earth-rotation synthesis and non-planar.
Lecture 8 Source detection NASSP Masters 5003S - Computational Astronomy
Lecture 10 The catalog of sources –Resolved sources –Selection biases –Luminosity (and mass) functions –Volume- vs flux-limited surveys. Cross-matching.
From the population to the sample The sampling distribution FETP India.
NASSP Masters 5003F - Computational Astronomy Lecture 11 Single-dish spectra: baselines. Complex numbers refresher Fourier refresher Correlation.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Hypothesis Tests. An Hypothesis is a guess about a situation that can be tested, and the test outcome can be either true or false. –The Null Hypothesis.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
NASSP Masters 5003F - Computational Astronomy Lecture 4: mostly about model fitting. The model is our estimate of the parent function. Let’s express.
All-sky source search Aim: Look for a fast method to find sources over the whole sky Selection criteria Algorithms Iteration Simulations Energy information.
Lecture 15 Deconvolution CLEAN MEM Why does CLEAN work? Parallel CLEAN
Central Limit Theorem, z-tests, & t-tests
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Presentation transcript:

NASSP Masters 5003F - Computational Astronomy Lecture 23 Two areas of my current research interest: –Parallel CLEAN. –Looking for bursts in XMM data.

NASSP Masters 5003F - Computational Astronomy Applications of parallel CLEAN I’ve told you something about this already. –In lecture 15. There you had the nice pix, here we learn more about the nuts and bolts. The original CLEAN is a technique for deconvolving images made by aperture synthesis. CLEAN relies upon the assumption that the dirty image D can be described as a convolution between the dirty beam B and a ‘true sky image’ I which is supposed to be a set of point sources. Ie or

NASSP Masters 5003F - Computational Astronomy Applications of parallel CLEAN It turns out that there are quite a few cases where this assumption breaks down. Eg –If the observation bandwidth is large enough that the different spectral shape of different parts of the sky can no longer be neglected. This was the original application. –If the observed sky contains sources which vary over the duration of the observation. –Even if a point source is not located exactly in the centre of an image pixel, the assumption breaks down.

NASSP Masters 5003F - Computational Astronomy Applications of parallel CLEAN The Sault-Wieringa algorithm, as applied to the wide-band situation, approximates the spectrum S at each sky position by a sum over basis functions f j : Here the A j ( r ) are amplitudes which will vary over the sky. The basis function in the original paper were Taylor-series power functions, but any set which spans the space of possible spectra would do as well.

NASSP Masters 5003F - Computational Astronomy Applications of parallel CLEAN The algorithm then constructs, not one, but M beams. The jth beam is constructed from visibilities appropriate to a point source at the phase centre and with a spectrum equal to f j. The dirty image can then be expressed as a sum of convolutions: The algorithm finds the brightest pixel at each iteration, but then calculates the amount of each beam present, and subtracts λ times each from the residual image.

NASSP Masters 5003F - Computational Astronomy Parallel CLEAN of time-varying sources. There’s no reason why this can’t be extended to sources which vary with time, or even time- and spectrum-variable sources. It is just necessary to expand the flux in the appropriate set of basis functions: There are, however, a couple of issues, both arising out of the ‘more wiggly’ nature of many radio time series, as compared to spectra.

NASSP Masters 5003F - Computational Astronomy Problems caused by rapid time variation A radio spectrum nearly always is a slowly-varying function of frequency. Over a fractional bandwidth of up to a decade (ie, a ten-fold change), many can be quite well approximated by a simple power law. –A power law doesn’t need many terms of a Taylor series – 3 at maximum – for a good approximation. Time series may need many more orders for good approximation – and for this, a Taylor series begins to seem not the best choice of basis function.

NASSP Masters 5003F - Computational Astronomy Problems caused by rapid time variation Other possibilities: –A Fourier series. –Chebyshev (or Tchebyshev) polynomials. We’ll consider both in a little while. First let’s discuss the other problem: Suppose we consider the full frequency-time variation. If we use for example 3 orders of frequency basis function f and 9 orders of time basis function g (not unreasonable), that means we need 3 x 9 = 27 beams. It gets worse than this because, within the S-W algorithm, we must store 27x27 = 729 internal arrays!

NASSP Masters 5003F - Computational Astronomy Gram-Schmidt orthogonalization It turns out that we only need NxN arrays if the beams are non-orthogonal. If we make them orthogonal, just N (27 in our example) are needed – a substantial saving not just in memory use, but also in speed. The beams can be orthogonalized by a Gram-Schmidt process. –But... what does it mean to say that two images are ‘orthogonal’? Images aren’t vectors! Anything for which a scalar product can be defined can be treated like a generalized vector. Eg for two images I and B :

NASSP Masters 5003F - Computational Astronomy Gram-Schmidt v1v1 v2v2 v1v1 v2v2 v1v1 v2v2 v1v1 v2v2 v1v1 v2v2 u1u1 (v2·u1)u1(v2·u1)u1 v 2 -(v 2 ·u 1 )u 1 u2u2 u1u1 u1u1 u1u1 u 1 = v 1 /|v 1 | u 2 = (v 2 -[v 2 ·u 1 ]u 1 )/|same| u 3 = (v 3 -[v 3 ·u 1 ]u 1 -[v 3 ·u 2 ]u 2 )/|same| etc

NASSP Masters 5003F - Computational Astronomy Gram-Schmidt However, for parallel clean, we need to be able to recover the old basis. So we express the relation between vs and us via a matrix of coefficients: G is lower-triangular, with elements: –g i,I = |u i | –g i,j = for j<i. Somewhat surprisingly, construction of G is ill- conditioned – ie ok in theory, but in practice the errors accumulate and soon grow to dominate. –  Somewhat greater cunning is needed.

NASSP Masters 5003F - Computational Astronomy The Wikipedia algorithm: I’m not kidding, there’s a good article on Gram-Schmidt on Wikipedia. (There’s nothing in Press et al for once.) This is pretty stable. # In pythonish language: u = v[i] for i in range(len(v)): g[i,i] = u[i] = u[i] – g[i,j]*u[j] g[i,i] = sqrt( ) u[i] = u[i] / g[i,i]

NASSP Masters 5003F - Computational Astronomy Basis sets: (i) Fourier series. This is standard stuff, and easy to calculate. But! –Any function can of course be approximated by a Fourier series. –But unless the function (i) has periodic boundary conditions and (ii) is everywhere analytic (=“really smooth” in layman’s language), the approximation is limited by the Gibbs phenomenon.

NASSP Masters 5003F - Computational Astronomy Gibbs ‘ringing’. Max order = 10

NASSP Masters 5003F - Computational Astronomy Gibbs ‘ringing’. Max order = 100

NASSP Masters 5003F - Computational Astronomy Basis sets: (ii) Chebyshev polynomials. From Press et al, “Numerical Recipes” Note: Standard notation is T n. They are defined on an interval [-1,1].

NASSP Masters 5003F - Computational Astronomy Basis sets: (ii) Chebyshev polynomials. Recurrence relations: –Many ‘special’ functions have them, and they can be very useful aids for computation. Care is needed to choose one which converges though! Once again, there is a difference between a neat formula on paper and a stable computer algorithm. –For Chebyshevs we can use the following algorithm to calculate T n for as high n as we like: # First, transform x so its range becomes [-1,1]. T[0] = 1 T[1] = x for i in range(2,N): T[i] = 2.0*x*T[i-1] – T[i-2]

NASSP Masters 5003F - Computational Astronomy Chebyshevs don’t suffer from Gibbs! Max order = 10Max order = 5!

NASSP Masters 5003F - Computational Astronomy Off-pixel-centre sources RED = real, BLUE = imaginary. U-V ‘plane’‘Sky’ Fourier

NASSP Masters 5003F - Computational Astronomy Off-pixel-centre sources = x

NASSP Masters 5003F - Computational Astronomy Off-pixel-centre sources ~ x

NASSP Masters 5003F - Computational Astronomy The next research interest: x-ray flashes in XMM data? The standard XMM source detection deals with total counts.  List of sources. –A simple null-hypothesis variability test is performed on all such sources, and any interesting ones will have already been picked over. Q1: is a source which has its counts bunched in a short time easier to detect than one with counts spread out over the exposure duration? –A: seems reasonable, but we’ll do some figuring in a minute to check.

NASSP Masters 5003F - Computational Astronomy The next research interest: x-ray flashes in XMM data? Q2: Would x-ray bursts have been picked up by other satellites? –A: The only others worth considering are RXTE and the BAT instrument on Swift. Q3: Can we estimate how many candidates there might be? –Let’s do a logN-logS of the BAT detections, then extrapolate. –The other figures we need: XMM-Newton has a FOV with ½ degree diameter; It has been operating for 10 years.

NASSP Masters 5003F - Computational Astronomy Search for a typical GRB From P T O’Brien et al, astro-ph/ Average duration ~ 30 sAverage spectral index ~ 0.5

NASSP Masters 5003F - Computational Astronomy Q1 – which is easier to detect, a short or long burst? Let’s do some simple modelling. Suppose we have N time bins, each of which contains a number of x-ray counts c i. These data will of course be Poisson randoms. The null hypothesis is that the parent light curve is just that of the background, which we’ll assume we know to be a constant value B counts per bin. We want to compare two probabilities: –After summing the bin counts to get C, the probability P 1 that C or higher would arise by random fluctuations about the null hypothesis value of N*B; –The probability P 2 that B would produce (by random fluctuations) C in at least one of the bins.

NASSP Masters 5003F - Computational Astronomy The battle of the probabilities - an example Which is more likely to result by chance just from a constant background level? or Both give a total of five counts. Time

NASSP Masters 5003F - Computational Astronomy The battle of the probabilities P 1 is easy to calculate – it is just Similarly, the probability of getting C or greater in a single bin is just But it turns out that the probability of this happening in any of N bins is

NASSP Masters 5003F - Computational Astronomy The battle of the probabilities Let’s take as typical figures: –Background rate = 5.0e-5 counts/s (roughly the lower limit for XMM, summed over all instruments, all bands). –Detection probability P det chosen to be exp(-8) (same as for 1XMM and 2XMM source catalogs). –Exposure duration 30ks (a typical XMM value). –Bin duration 30 s (a typical GRB burst duration). Calculation shows that 6 counts are needed for P 1 <P det, but only 2 counts in any 10 second time bin for P 2 <P det. So the answer to Q1 is YES it is a reasonable project to look for bursts not picked up by the standard pipeline source detection – at least, statistically speaking.

NASSP Masters 5003F - Computational Astronomy Q2 – have Rossi and Swift got the scoop? Both Rossi (alias RXTE) and Swift have wide-area detectors. Their sensitivities are not all that high however: –Swift BAT: about 4e-9 erg cm -2 s -1. (Converted to XMM keV band, assuming a typical GRB spectral index of +0.5.) BAT detects about 0.3 GRB sr -1 day -1. –ASM on RXTE: 30 mCrab – works out in erg cm -2 s -1 to much worse than BAT. Forget it - RXTE is not in the running. Compared to –XMM: about 1e-12 erg cm -2 s -1. So it seems there is some territory to explore.

NASSP Masters 5003F - Computational Astronomy Q3 – how many to expect? Swift BAT XMM: ~8e4 GRB sr -1 day -1. FOV = 6e-5 sr;  expect 5 a day. This figure should be taken with a pinch of salt, but it is intriguing to say the least. Log 10 (flux) in erg cm -2 s -1 Log 10 (GRB rate) in GRB sr -1 day -1 Slope of 1.5 if evenly distributed in space

NASSP Masters 5003F - Computational Astronomy How to go about it? A variation on a technique described in J Scargle ApJ 504, 405 (1998) suggests itself. Scargle used a Bayesian analysis, but I am more interested in his model: –He compared model A which consisted of constant flux over the whole interval, with model B which cut the time series into two blocks, within each of which the flux was constant. Ie:

NASSP Masters 5003F - Computational Astronomy Comparison of two models: Model A: 0T T0T1T1 Model B: λ1λ1 λ2λ2 λ3λ3 (Ignoring background B(t))

NASSP Masters 5003F - Computational Astronomy Likelihood ratio The likelihood ratio between these models evaluates to (T 1 =NΔt and T=MΔt for bin duration Δt.) The procedure would mimic Scargle’s: –Calculate U for all possible T 1 ; –See how probable the maximum value is. Have to figure in the PSF too... and worry about how well the background is known... and instrumental noise... and probably some other things too.

NASSP Masters 5003F - Computational Astronomy Haven’t made it work yet......but maybe some student will.