We use two products because each has its strengths. a. Ancient (4.x) version of WebTrends. Pluses: - Since we have been using it since the last millennium, it's the closest we can get to accurate trending over time on the usage of our server. - Unlike later versions of the same product, the license doesn't max out after processing X number of page views. So at this point it's free. - Counts everything in the log files, so if I edit out certain lines in the log files and rerun the reports, I can know % of usage from in-house, Googlebot, Slurp, etc. Minuses: - Can no longer report on things like referring search engines (since it's so old it doesn't know Google is a search engine). - Reporting capabilities are not as slick as Google Analytics. - Counts everything in the log files, so it's got a lot of noise data if you just want to see what real non-bot users from outside the library are doing. b. Google Analytics Pluses: - Very nice reporting. - Easy setup, including configuring it to exclude internal use. - Free. - Requires JavaScript, which more or less automatically excludes non-human bot hits from the reports. Minuses: - Someone else has our log data. - Haven't found any feature to combine multiple logs and report on all 3 public servers in a single report. - Requires JavaScript, which excludes reporting on usage from browsers with JavaScript turned off/unsupported. I looked around a number of other tools and found many of the exact same pluses and minuses, as well as the rather staggering price of some of the commercial offerings, many of which now also use the JavaScript-ping-to-remote-server method. Currently, I actually have somewhat more faith in Google to keep our users' data private than I do in the pricey analytics vendors, but this may be naive on my part. At any rate, using one log-based tool and one JavaScript-based tool is working OK for us for now. There's an archived webcast about analytics options specifically for measuring web "visits" from outside the library here: http://www.infopeople.org/training/webcasts/webcast_data/255/index.html Genny Engel Sonoma County Library [log in to unmask] 707 545-0831 x581 www.sonomalibrary.org >>> [log in to unmask] 11/11/08 12:52PM >>> Library Code People: 1 - What do you use for your web statistics package? Are you happy with it? Pros/Cons? 2 - What do you wish you used or had access to? 3 - Opinions on Specific Projects: 3.1 Piwiki/Mint Piwik and Mint both seem pretty interesting to me because they solve some of the problems of "traditional" log file analysis (see http://www.boxesandarrows.com/view/the-limitations-of#comments) , while of course introducing their own set of problems: given their reliance on a RDBMS to store each page load, there are some obvious scaling concerns for very high traffic sites, for example. I wonder if anyone here has put either of these or similar systems to the test on high traffic (define in your own terms) sites. 3.2 Google Analytics and/or Urchin Some libraries have incorporated Google Analytics into their privacy policies: http://www.google.com/search?q=google+analytics+libraries+privacy&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla:en-US:official&client=firefox-a . So, anyone here passionate one way or the other? Other Pros/Cons? Of course, favorite resources, questions I should be asking and the like are welcomed and appreciated as well :). In advance: thanks! Cheers, -Chad