On 5/22/06, Houghton,Andrew <[log in to unmask]> wrote: > > From: Code for Libraries [mailto:[log in to unmask]] On Behalf > > Of Ross Singer > > Sent: 22 May, 2006 11:58 > > To: [log in to unmask] > > Subject: Re: [CODE4LIB] Musings on using xISBN in our Horizon > > catalog > > > > Given the size of both datasets, we are still talking about a few > > seconds per query with most currently available SQL databases. But > > what happens when the size of the catalog doubles, triples, quadruples, etc.? > > Wait. A few seconds? You're kidding, right? An indexed look up of the group id Oops, flagrant typo revealed. That should have read "... We are still talking about a few milliseconds per query..." If only my fingers could read my mind... > Just so we're clear, we are talking about, at most, .5M rows that are perhaps 50 bytes > wide, with internal data. That translates to ~25 megabytes for the table, and the > indexes would be /maybe/ 5 megs each. There is more than one way to solve a problem. You are free to use an unnormalized database design if that is adequate for your needs. But why would you want a database ~25 MB with 10MB of indexes, when you could get the same results from a normalized database design whose storage requirements were *roughly* 4 times less and get better scalability should your requirements change in the future? Yes, I can hear the come-back, "disk space is cheap", but someone pays for it at some point in time. I can agree to disagree with using an unnormalized database design and as I indicated in my first post, "If I were setting up the SQL database, I would probably do the following..." > > A good database design will go a long way... Andy.