Thanks, I added this as a comment on the code4lib talk page from the conf. If anyone else happens to be looking for a video and finds it, and you want to add it to the code4lib talk page in question, it would probably be useful for findability. In the past I think someone bulk added the URLs to all the talk pages, but I guess that didn't happen this time, I guess actually cause there aren't split videos of each talk but just video of the whole session? Hmm, I guess that makes it harder to figure out what the URL to the right minute of the talk should be, unless you're Jason. Oh well. Thanks Jason! On 9/14/2011 1:49 PM, Jason Stirnaman wrote: > PS, Here's the link for jumping to Thomas Browning's Metridoc talk: http://www.indiana.edu/~video/stream/launchflash.html?format=MP4&folder=vic&filename=C4L2011_session_2_20110208.mp4&starttime=3600 > > > Jason Stirnaman > Biomedical Librarian, Digital Projects > A.R. Dykes Library, University of Kansas Medical Center > [log in to unmask] > 913-588-7319 > > >>>> On 9/14/2011 at 11:13 AM, in message<[log in to unmask]>, Jonathan Rochkind<[log in to unmask]> wrote: > > Yeah, I think it ends up being pretty hard to create general-purpose > solutions to this sort of thing that are both not-monstrous-to-use and > flexible enough to do what everyone wants. Which is why most of the > 'data warehouse' solutions you see end up being so terrible, in my > analysis. > > I am not sure if there is any product specifically focused on library > usage/financial data -- that might end up being somewhat less monstrous, > it seems the more you focus your use case (instead of trying to provide > for general "data warehouse and analysis"), the more likely a software > provider can come up with something that isn't insane. > > At the 2011 Code4Lib Conf, Thomas Barker from UPenn presented on some > open source software they were developing (based on putting together > existing open source packages to be used together) to provide > library-oriented 'data warehousing'. I was interested that he talked > about how their _first_ attempt at this ended up being the sort of > monstrous flexible-but-impossible-to-use sort of solution we're talking > about, but they tried to learn from their experience and start over, > thinking they could do better. I'm not sure what the current status of > that project is. I'm not sure if any 2011 code4lib conf video is > available online? If it is, it doesn't seem to be linked to from the > conf presentation pages like it was in past years: > > http://code4lib.org/conference/2011/barker > > On 9/13/2011 5:37 PM, Jason Stirnaman wrote: >> Thanks, Shirley! I remember seeing that before but I'll look more closely now. >> I know what I'm describing is also known, typically, as a data warehouse. I guess I'm trying to steer around the usual solutions in that space. We do have an Oracle-driven data warehouse on campus, but the project is in heavy transition right now and we still had to do a fair amount of work ourselves just to get a few data sources into it. >> >> >> Jason Stirnaman >> Biomedical Librarian, Digital Projects >> A.R. Dykes Library, University of Kansas Medical Center >> [log in to unmask] >> 913-588-7319 >> >> >>>>> On 9/13/2011 at 04:25 PM, in message<[log in to unmask]>, Shirley Lincicum<[log in to unmask]> wrote: >> Jason, >> >> Check out: http://www.needlebase.com/ >> >> It was not developed specifically for libraries, but it supports data >> aggregation, analysis, web scraping, and does not require programming >> skills to use. >> >> Shirley >> >> Shirley Lincicum >> Librarian, Western Oregon University >> [log in to unmask] >> >> On Tue, Sep 13, 2011 at 2:08 PM, Jason Stirnaman<[log in to unmask]> wrote: >>> Does anyone have suggestions or recommendations for platforms that can aggregate usage data from multiple sources, combine it with financial data, and then provide some analysis, graphing, data views, etc? >>> From what I can tell, something like Ex Libris' Alma would require all "fulfillment" transactions to occur within the system. >>> I'm looking instead for something like Splunk that would accept log data, circulation data, usage reports, costs, and Sherpa/Romeo authority data but then schematize it for data analysis and maybe push out reporting dashboards<nods to Brown Library http://library.brown.edu/dashboard/widgets/all/> >>> I'd also want to automate the data retrieval, so that might consist of scraping, web services, and FTP, but that could easily be handled separately. >>> I'm aware there are many challenges, such as comparing usage stats, shifts in journal aggregators, etc. >>> Does anyone have any cool homegrown examples or ideas they've cooked up for this? Pie in the sky? >>> >>> >>> Jason >>> Jason Stirnaman >>> Biomedical Librarian, Digital Projects >>> A.R. Dykes Library, University of Kansas Medical Center >>> [log in to unmask] >>> 913-588-7319 >>>