SED data model v0.92
Ed Shaya
edward.j.shaya.1 at gsfc.nasa.gov
Tue Nov 23 06:29:55 PST 2004
Gilles DUVERT wrote:
>
>
> Ed Shaya wrote:
>
>>
>> You want the lower error bar to reach from the upper limit to the
>> x-axis.
>>
> I think, in view of what you object, that I would prefer upper limits
> to have a different status than measure. Besides, I'm ill at ease
> thinking measure in geometrical terms (or, rather, in "plot+axis"
> terms), this looks too linked with our day to day external
> representation of data, (a mathematician would perhaps read "measure"
> as a "norm" for topology )
>
Although it makes the model more complicated, it is probably worth
having an UpperLimitPoint and a LowerLimitPoint which contains only the
limit to be used ONLY when the limit value is not known. It could have
an optional NSigma which tells how many sigma limit was used and the
default should be 3. This way one does not confuse the upper limit
magnitude with the true error value or the true value.
>>> The risk would be, if any *measured* value is set, that is is taken
>>> at face value, when only the noise level has sense. Could you use
>>> some kind of blanking value for the measured value in this? (or is
>>> there a general concept of upper limit that would go in the
>>> Quantity::Accuracy data model?)
>>>
>> All I need say here is that if 4 independent experiments come up
>> with 0.9 sigma detections of some measurement that it would be
>> awful if they each published only the upper limits of the value.
>
>
> Here you suppose that the value "detected at 0.9 sigma" exists, or,
> rather, that I can pinpoint it in the graph, and put nice big
> errorbars on it for each of the 4 measurements. What I say is that
> this value does not exist until it is measured, and it is measured
> only when it is not an upper limit.
Are we having some subtle argument about when a measurement is a
measurement and when a detection is a detection?
If I measure 9 photons in a pixel but the sky noise is 10 photons we
keep track of those 9 photons by placing it in the point/value just like
any other measured value.
With just this measurement it is not a detection of anything except for
those 9 photons, but if I go back through the archives and find that
this location has been studied 3 times before, perhaps with different
filters, and each time this location has roughly 1 sigma positive
measurements, then I would say that an object is detected at roughly 4
sigma and my observation was just a 0.9 sigma detection. Perhaps you
would phrase this differently.
But, I hope that you are not supporting the idea that the USUAL
treatment for a measurement that is less than 3 sigma should be to
drop the value and just quote the upper limit.
You would not support blanking out all of the pixels in an image or
spectrum where the signal is below 3 times the sky noise and replace
them with upper limit values? Or, would you? We would need to
reprocess all existing data archives to treat upper limits in this way?
If 4 major physics laboratories come up with 1 sigma detections of the
mass of the neutrino at about 10eV should each publish just upper limits?
>
> Of the 4 groups coming up with this "0.9 sigma detection", the 3 last
> in time are morons: they were not able to devise an experiment with
> less uncertainty as the 1st pioneering group. Shall we continue to
> support them financially? ;^).
That is a political statement.
> Besides, since the experiments are not the same, and you want to get a
> measurement at the end by averaging values+errors, you have to prove
> that errors and "measure" in those different experimental setups can
> be averaged. Unless the 4 experiments are just 4 realizations of the
> 1st measurement, and, bingo, upper limits are still upper limits in
> this case....
>
We certainly need to have a system so that if the same measurement shows
up in several context that we and our applications can easily recognize
it. That is the point of putting IDs on observations.
Beyond that, it is the job of a distributed data system to be able to
pull together similar data entities for processing. Isn't the idea of
the VO to allow us to do analyses across multiple data sets and data
archives that is beyond our present capabilities.
> Fortunately, for people bold enough to claim (and they are numerous!)
> that their 0.9 sigma measurement (was really an upper limit) _is_ a
> measurement, they can use the normal value+error scheme.
This should be the normal mode and our applications should be upgraded
to recognize measurements below 3 sigma and plot them as yellow and
properly draw the upper limit symbol.
>
>> The possibility that some moron may not look at the quoted noise
>> levels before coming to some silly conclusion on a measurment does
>> not compensate for missing out on potentially important real
>> discoveries that properly archived data makes possible.
>
>
> a) The possibility of "some moron..." is huge.
Oh yes! This is a tradeoff between a bunch of people wasting some time
against losing precious perhaps irretrievable information.
> b) I would not place too much faith in discoveries (real, important)
> based on a sum of invalid measurements...
>
This is the crux of it. To me, a less than 3 sigma measurement is a
valid measurement. It tells me the probability of something being
there so that I can properly assess my risks in attempting to go deeper
in that direction. If it is an invalid measurement then I don't want it
published at all anywhere.
>
> Best,
> Gilles
>
Cheers,
Ed
More information about the dm
mailing list