Print

Print


Thanks Eric for all the links (I am exploring these and still these are
accessible), and the AI4LAM Zoom meeting.

Best regards



On Fri, May 10, 2024 at 11:05 PM Eric Lease Morgan <
[log in to unmask]> wrote:

> Parthasarathi Mukhopadhyay <[log in to unmask]> wrote:
>
> > ...However, one interesting point to be mentioned here is the effect of
> prompt
> > engineering on a RAG pipeline. When I ask the same questions as Simon did
> > on the same set of documents in a similar kind of pipeline with prompt
> > engineering, the result shows some differences (see additional system
> > prompt in the snapshot):
> >
> > --
> > Parthasarathi Mukhopadhyay
>
>
> Yes, when it comes to generative-AI, prompt engineering is a real thing.
> Prompts are akin to commands given to a large-language model, and different
> large-language models have different prompts. The prompt I have been using
> in my proof-of-concept applications have this form:
>
>   Context information is below.
>   ---------------------
>   {context_str}
>   ---------------------
>   Given the context information and not prior knowledge, answer the query
>   Write the answer in the style of {speaker} and intended for {audience}.
>   Query: {query_str}
>   Answer:
>
> Where the placeholders (the things in curly braces) are replaced with
> values from the interface. For example, {context_str} is the content of
> documents pointing in the same vectored direction as the query. The
> {speaker} placeholder might be "a second grader", "a librarian", "a
> professor emeritus", etc. The same thing is true for {audience}. The value
> of {query_str} is whatever the user (I hate that word) entered in the
> interface.  The prompt is where one inserts things like the results of the
> previous interaction, what to do if there is very little context, etc.
> Prompt engineering is a catch-as-catch-can. Once a prompt is completed, it
> given as input to the large-languagae model for processing -- text
> generation.
>
> Over the past week, I have created five different generative-AI chatbot
> interfaces, each with their own different strengths and weaknesses:
>
>   * climate change - https://5c0af9ffadb4b3d2ba.gradio.live
>   * library cataloging - https://6a147d360a3fc1d7df.gradio.live
>   * Jane Austen - https://e7053a831a40f92a86.gradio.live
>   * children's literature - https://a10e1d2687be735f40.gradio.live
>   * What's Eric Reading - https://e462cd2ac6d1e35d1c.gradio.live
>
> These interfaces use a thing called Gradio (https://www.gradio.app/) for
> I/O, and they are supposed to last 72 hours, but all of them still seem to
> be active. Go figure.
>
> Finally, today I saw an announcment for a AI4LAM Zoom meeting on the topic
> of RAG where three different investigations will be presented:
>
>   * Kristi Mukk and Matteo Cargnelutti (Harvard Library Innovation Lab),
> Warc-GPT
>   * Daniel Hutchinson (Belmont Abbey College), the Nicolay project
>   * Antoine de Sacy and Adam Faci (HumaNum Lab), the Isidore project
>
> See the Google Doc for details: https://bit.ly/3QEzQ6f
>
> --
> Eric Morgan
> University of Notre Dame
>