I'll try to update a few more. You can adjust the starttime parameter (in seconds) in the URL accordingly for each talk. Of course, you have to watch to figure out where they start.
Biomedical Librarian, Digital Projects
A.R. Dykes Library, University of Kansas Medical Center
[log in to unmask]
>>> On 9/14/2011 at 12:58 PM, in message <[log in to unmask]>, Jonathan Rochkind <[log in to unmask]> wrote:
Thanks, I added this as a comment on the code4lib talk page from the conf.
If anyone else happens to be looking for a video and finds it, and you
want to add it to the code4lib talk page in question, it would probably
be useful for findability.
In the past I think someone bulk added the URLs to all the talk pages,
but I guess that didn't happen this time, I guess actually cause there
aren't split videos of each talk but just video of the whole session?
Hmm, I guess that makes it harder to figure out what the URL to the
right minute of the talk should be, unless you're Jason. Oh well. Thanks
On 9/14/2011 1:49 PM, Jason Stirnaman wrote:
> PS, Here's the link for jumping to Thomas Browning's Metridoc talk: http://www.indiana.edu/~video/stream/launchflash.html?format=MP4&folder=vic&filename=C4L2011_session_2_20110208.mp4&starttime=3600
> Jason Stirnaman
> Biomedical Librarian, Digital Projects
> A.R. Dykes Library, University of Kansas Medical Center
> [log in to unmask]
>>>> On 9/14/2011 at 11:13 AM, in message<[log in to unmask]>, Jonathan Rochkind<[log in to unmask]> wrote:
> Yeah, I think it ends up being pretty hard to create general-purpose
> solutions to this sort of thing that are both not-monstrous-to-use and
> flexible enough to do what everyone wants. Which is why most of the
> 'data warehouse' solutions you see end up being so terrible, in my
> I am not sure if there is any product specifically focused on library
> usage/financial data -- that might end up being somewhat less monstrous,
> it seems the more you focus your use case (instead of trying to provide
> for general "data warehouse and analysis"), the more likely a software
> provider can come up with something that isn't insane.
> At the 2011 Code4Lib Conf, Thomas Barker from UPenn presented on some
> open source software they were developing (based on putting together
> existing open source packages to be used together) to provide
> library-oriented 'data warehousing'. I was interested that he talked
> about how their _first_ attempt at this ended up being the sort of
> monstrous flexible-but-impossible-to-use sort of solution we're talking
> about, but they tried to learn from their experience and start over,
> thinking they could do better. I'm not sure what the current status of
> that project is. I'm not sure if any 2011 code4lib conf video is
> available online? If it is, it doesn't seem to be linked to from the
> conf presentation pages like it was in past years:
> On 9/13/2011 5:37 PM, Jason Stirnaman wrote:
>> Thanks, Shirley! I remember seeing that before but I'll look more closely now.
>> I know what I'm describing is also known, typically, as a data warehouse. I guess I'm trying to steer around the usual solutions in that space. We do have an Oracle-driven data warehouse on campus, but the project is in heavy transition right now and we still had to do a fair amount of work ourselves just to get a few data sources into it.
>> Jason Stirnaman
>> Biomedical Librarian, Digital Projects
>> A.R. Dykes Library, University of Kansas Medical Center
>> [log in to unmask]
>>>>> On 9/13/2011 at 04:25 PM, in message<[log in to unmask]>, Shirley Lincicum<[log in to unmask]> wrote:
>> Check out: http://www.needlebase.com/
>> It was not developed specifically for libraries, but it supports data
>> aggregation, analysis, web scraping, and does not require programming
>> skills to use.
>> Shirley Lincicum
>> Librarian, Western Oregon University
>> [log in to unmask]
>> On Tue, Sep 13, 2011 at 2:08 PM, Jason Stirnaman<[log in to unmask]> wrote:
>>> Does anyone have suggestions or recommendations for platforms that can aggregate usage data from multiple sources, combine it with financial data, and then provide some analysis, graphing, data views, etc?
>>> From what I can tell, something like Ex Libris' Alma would require all "fulfillment" transactions to occur within the system.
>>> I'm looking instead for something like Splunk that would accept log data, circulation data, usage reports, costs, and Sherpa/Romeo authority data but then schematize it for data analysis and maybe push out reporting dashboards<nods to Brown Library http://library.brown.edu/dashboard/widgets/all/>
>>> I'd also want to automate the data retrieval, so that might consist of scraping, web services, and FTP, but that could easily be handled separately.
>>> I'm aware there are many challenges, such as comparing usage stats, shifts in journal aggregators, etc.
>>> Does anyone have any cool homegrown examples or ideas they've cooked up for this? Pie in the sky?
>>> Jason Stirnaman
>>> Biomedical Librarian, Digital Projects
>>> A.R. Dykes Library, University of Kansas Medical Center
>>> [log in to unmask]