You might want to have a look at Zebra for the indexing engine. http://indexdata.dk/zebra/ Great tool for making a file of MARC records Z searchable. John On 27-Oct-05, at 2:58 PM, Eric Lease Morgan wrote: > On Oct 27, 2005, at 2:06 PM, Andrew Nagy wrote: > > >>> http://mylibrary.ockham.org/ >>> >> >> I have been thinking of ways, similiar to what you have done that you >> mentioned below with the Ockham project, to allow more modern day >> access >> with our library catalog. I have been beginning to think about >> devising >> a way to index/harvest our entire catalog (and allow this indexing >> process to run every so often) to allow our own custom access >> methods. >> We could then generate our own custom RSS feeds of new books, allow >> more >> efficient/enticing search interfaces, etc. >> >> Do you know of any existing software for indexing or harvesting a >> catalog into another datastore (SQL Database, XML Database, etc). >> I am >> sure I could fetch all of the records somehow through Z39.50 and >> dump it >> into a MySQL database, but maybe there is some better method? >> > > I too have thought about harvesting content from my local catalog and > providing new interfaces to the content, and I might go about this in > a number of different ways. > > 1. I might use OAI to harvest the content, cache is locally, and > provide services against the cache. This cache might be saved on a > file system, but more likely into a relational database. > > 2. I might simply dump all the MARC records from my catalog, > transform them into something more readable, say sets of HTML/XML > records, and provide services against these files. > > The weakest link in my chain would be my indexer. Relational > databases are notoriously ill-equipped to handle free text searching. > Yes, you can implement it and you can use various database-specific > features to implement free text searching, but they still won't work > as well as an indexer. My only experience with indexers lies in > things like swish-e and Plucene. I sincerely wonder whether or not > these indexers would be up to the task. > > Supposing I could find/use an indexer that was satisfactory, I would > then provide simple and advanced (SRU/OpenSearch) search features > against the index of holdings. Search results would then be enhanced > with the features such as borrow, re-new, review, put on reserve, > save as citation, email, "get it for me", put on hold, "what's new?", > view as RSS, etc. These services would require a list of authorized > users of the system -- a patron database. > > In short, since I would have direct access to the data, and since I > would have direct to the index, I would use my skills to provide > services them. For the most part, I don't mind back-end, > administrative, data-entry interfaces to our various systems, but I > do have problems with the end-user interfaces. Let me use those back- > ends to create and store my data, then give me unfettered access to > the data and I will provide my own end-user interfaces. Another > alternative is to exploit (industry standard) Web Services computing > techniques against the existing integrated library system. In this > way you get XML data (information without presentation) back and you > can begin to do the same things. > > -- > Eric Lease Morgan > University Libraries of Notre Dame > -- John Durno Project Coordinator BC Electronic Library Network ~~~~~~~~~~~~~~~~~~~~ Phone: 604-268-7002 Fax: 604-291-3023 Email: [log in to unmask] Web: http://www.eln.bc.ca