[Quality] another can of worms
Martin Hill
mchill at dial.pipex.com
Thu Nov 20 15:22:38 PST 2003
Hi all
I've been wondering what to do about quality, and how to describe it.
We have a general one - if we allow anyone to publish to the VO, how do
we describe a datacenter's quality? There are datacenters setup for a
sky survey, with all the processing involved that has been set up by
teams of astronomers (presumably?) double checking each other. And there
is data published by a small group or an individual so that people can
access it, but which has not gone through such a rigorous set of tests.
We don't really want a general query to the registry returning both as
'equals value' (or do we?).
We also have a smaller scale one - each item of data may need to be
marked. For example, a sky survey may have items that have been marked
as 'possibly satellite track' or 'instrument feature' (is that the word?
such as diffraction spikes?)
It's surfaced occasionally in some groups as a 'placeholder' but doesn't
seem to have been dealt with in itself. Has anyone come up with a
framework (I can't find a general one)? If not, where do we start? Is
a 'placeholder' sufficient for all three groups (dal/dm/registry) to
start with?
I've posted this across all three groups for discussions about quality
and what it means/implies on a large scale. Obviously it should split
down when it becomes more specific! Such as whether [Quantity] should
include [Quality] as well as [Error]; and whether this is the same
[Quality] as that attached to a dataset... :-)
Cheers,
Martin
--
Software Engineer
AstroGrid @ ROE
Tel: +44 7901 55 24 66
www.astrogrid.org
More information about the dm
mailing list