Print

Print


Greg,

At NCSU we are working with FGDC metadata in our NDIIPP project  
focusing on preservation of state and local government digital 
geospatial data.  Much of the state data and some of the local data 
arrives with FGDC metadata.  In our ingest workflow we are validating, 
normalizing, synchronizing and remediating existing FGDC metadata, and 
batch generating FGDC metadata where none exists.  In the latter case it 
is a question of focusing on the subset of software-generated technical 
metadata elements while also adding additional information from agency 
templates and selectively adding information from data inventory 
details.  We are normalizing to the ESRI Profile of FGDC in order to 
take advantage of additional technical metadata elements that are 
available.  Not all of this has been fully automated yet as we are 
working out the kinks and refining the work flow. 

We've participated in some FGDC-ISO crosswalk tests in order to prepare 
for the upcoming shift from FGDC ver. 2 to the North American profile of 
the ISO standard.  In exchange cases such as this we are stripping the 
ESRI Profile elements and synchronization tags for standards 
compliance.  Tools we are using include ArcCatalog, the NPS Metadata 
Toolkit, mp, and cns.  In the past we also used an in-house modified 
version of the NOAA Metadata Collector extension for batch processes 
(rewriting the dialog-driven tool as a non-dialog batch processor for 
extraction of technical elements)--the NPS Toolkit has superceded that 
for batch processes but there are questions about ongoing support of the 
NPS software.  ArcObjects-based processing may be a next direction for 
some of the things that we'd like to automate further--e.g., better 
control of synchronization.

Our partner on the NDIIPP project is the state GIS agency, which 
conducts metadata outreach and training for local agencies.  As part of 
the project we are compiling metadata quality information to be passed 
back to the metadata outreach effort to inform the training process and 
hopefully improve the quality of metadata seen in the longer term.  
Quality issues in this case being things like a change in datum or 
change in format not being recorded, and other innaccuracies.   Lack of 
concurrency between the data and metadata is a common problem.

We've also mapped FGDC elements, along with other technical and 
administrative metadata cultivated in the ingest preparation process, to 
DSpace QDC for repository ingest.  The plan is to include the FGDC 
record itself as well as a METS wrapper as bitstreams in the Dspace 
item--we're still fleshing this part out.  The Dspace QDC mapping is 
seen as a single spoke in the repository preparation process, as we'd 
also like to investigate mapping to FEDORA SIP's, etc.

In earlier projects going back to the mid-late 90's we've been involved 
in a variety of other projects with FGDC metadata components:  serving 
as the NPS  metasearch source (using Isite to index and serve SGML 
records via Z39.50 GEO profile); cross-walking FGDC to MARC as a 
beta-tester of a USGS tool, to incorporate geospatial data into the 
library catalog; and developing a public access discovery system that 
involves disambiguating key access elements from FGDC metadata 
(including key access elements or facets that are not discretely defined 
in FGDC) to inform discovery and selection.  Also, over the years we've 
provided metadata authoring training and support to grad students and 
research staff working on various campus projects.

Best regards,

Steve Morris

 

gmarch wrote:

>Hello,
>
>I am curious to know if any DLF partners are actively involved with FGDC 
>metadata related projects?
>
>Thank you,
>
>Greg
>
>Gregory March
>Graduate Research Assistant
>Map Library
>15 Hoskins Library
>University of Tennessee
>Knoxville, TN 37996
>(865)974-4315
>[log in to unmask]
>  
>

-- 
Steve Morris
Head of Digital Library Initiatives
North Carolina State University Libraries
Phone: (919) 515-1361  Fax: (919) 515-3031 
[log in to unmask]