Print

Print


I recommend this article as an entry point into a research program on 
information quality:

Stvilia, B., Gasser, L., Twidale, M. B. and Smith, L. C. (2007), A 
framework for information quality assessment. J. Am. Soc. Inf. Sci., 58: 
1720–1733. doi:10.1002/asi.20652 Available at: 
http://stvilia.cci.fsu.edu/wp-content/uploads/2011/03/IQAssessmentFramework.pdf

One cannot manage information quality (IQ) without first being able to 
measure it meaningfully and establishing a causal connection between the 
source of IQ change, the IQ problem types, the types of activities 
affected, and their implications. In this article we propose a general 
IQ assessment framework. In contrast to context-specific IQ assessment 
models, which usually focus on a few variables determined by local 
needs, our framework consists of comprehensive typologies of IQ 
problems, related activities, and a taxonomy of IQ dimensions organized 
in a systematic way based on sound theories and practices. The framework 
can be used as a knowledge resource and as a guide for developing IQ 
measurement models for many different settings. The framework was 
validated and refined by developing specific IQ measurement models for 
two large-scale collections of two large classes of information objects: 
Simple Dublin Core records and online encyclopedia articles.

Bob

On 5/6/2015 4:32 PM, Diane Hillmann wrote:
> You might try this blog post, by Thomas Bruce, who was my co-author on an
> earlier article (referred to in the post):
> https://blog.law.cornell.edu/voxpop/2013/01/24/metadata-quality-in-a-linked-data-context/
>
> Diane
>
> On Wed, May 6, 2015 at 5:24 PM, Kyle Banerjee <[log in to unmask]>
> wrote:
>
>>> On May 6, 2015, at 7:08 AM, James Morley <[log in to unmask]>
>> wrote:
>>>
>>> I think a key thing is to determine to what extent any definition of
>> 'completeness' is actually a representation of 'quality'.  As Peter says,
>> making sure not just that metadata is present but then checking it conforms
>> with rules is a big step towards this.
>>
>> This.
>>
>> Basing quality measures too much on the presence of certain data points or
>> the volume of data is fraught with peril. In experiments in the distant
>> past, my experience was that looking for structure and syntax patterns that
>> indicate good/bad quality as well as considering record sources was useful.
>> Also keep in mind that any scoring system is to some extent arbitrary, so
>> you don't want to read more into what it generates than appropriate.
>>
>> Kyle
>>
>