Skip to content

shivaypiece/react-workshop

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

React Ollama Workshop

Prerequisites

Here are the requirements :

Inside Ollama:

Installation

The application we'll be building today is using Vite, a build tool for modern JavaScript (and TypeScript) applications.

We need to set up the initial, bootstrapped application for this workshop. Run the following commands to set it up:

cd react-workshop
npm install
npm run dev

Go the link displayed in your terminal, you should be seeing the initial application.

Excercises

Excercise 1

To interface with the LLMs from OpenAI, we need to install a library called LangChain:

npm install langchain

After the installation is complete, you should add a new file called .env in the root of your Vite application and add the following environment variable:

Next, we'll create a new file called src/utils/langchain.ts and add the following code:

src/utils/langchain.ts
```ts
import { OpenAI } from "@langchain/openai";
const llm = new OpenAI({
  openAIApiKey: import.meta.env.VITE_OPENAI_KEY
  });
```

This will initialize a connection to OpenAI using LangChain and let us access the models.

We'll create our first function that can be used to generate an answer for a question, add the following to the bottom of the file:

src/utils/langchain.ts
export async function generateAnswer(question: string) {
    let answer = '';

    try {
        answer = await llm.invoke(question);
    } catch (e) {
        return 'Something went wrong';
    }

    return answer;
}

To test if what we've done is working, create new file called src/utils/langchain.test.ts and write a test for the function generateAnswer.

Take the following code and modify it so the test will succeed:

src/utils/langchain.test.ts
import { describe, it, assert } from 'vitest';
import { generateAnswer } from './langchain';

describe('LangChain', () => {
    it('Answers a question', async () => {
        // 1. Add your own question here
        const answer = await generateAnswer('YOUR QUESTION');

        // 2. Match the answer from the LLM to a predicted value
        assert.equal(answer.trim(), "THE ANSWER");
    });
});

Run npm run test to run the above test. Make sure your test is succeeding.

Hint: Be explicit of what you expect the LLM to return.

Excercise 2

We want to be able to use the messagebox in the application to send the question to the LLM and show the answer in the screen.

From our App component in src/App.tsx, we need to call the generateAnswer function we created in the previous excercise. First, let's create some state variables and import the function:

 async function handleSubmitQuestion(input: string) {
    try {
      const response = await generateAnswer(input)
      console.log("Question:", input)
      console.log("Answer:", response)
    } catch(e) {
      console.error(e)
    }
  }

Excercise 3

When you submit the form, you want to see the question and the answer displayed in the screen. Create a new component called Message in src/components/Message/Message.tsx, we'll use this component to render the question and the answer:

src/components/Message/Message.tsx
type MessageProps = {
    sender: string
    title: string,
    message: string,
    timestamp?: string
}

export default function Message({ sender, title, message, timestamp = "" }: MessageProps) {
    return (
        <div className="flex items-start gap-2.5 mx-8 mb-4">
            <div className="w-8 h-8 rounded-full bg-red-300">
                <span className="w-8 h-8 flex justify-center items-center">{sender}</span>
            </div>
            <div className="flex flex-col w-full leading-1.5 p-4 border-gray-200 bg-gray-100 rounded-e-xl rounded-es-xl dark:bg-gray-700">
                <div className="flex items-center space-x-2 rtl:space-x-reverse">
                    <span className="text-sm font-semibold text-gray-900 dark:text-white">{title}</span>
                    {timestamp && <span className="text-sm font-normal text-gray-500 dark:text-gray-400">{timestamp}</span>}
                </div>
                <p className="text-sm font-normal py-2.5 text-gray-900 dark:text-white">{message}</p>
            </div>
        </div>
    );
}

Render this component from src/App.tsx so it shows the question and the answer. You can use a name like "Me" for the question, and "GPT (or "AI") for the answer:

src/App.tsx
import { useState } from "react";
import { generateAnswer } from "./utils/langchain";
// 1. import `Message` component

export default function App() {
    // ...

    return (
        // ...

        <div className="h-full ">
              <div className="h-full flex flex-col items-center text-sm dark:bg-gray-800">
                  // 2. Render the Message component for the question and answer
              </div>
          </div>

        // ...
    );
}

When you complete this excercise you should be able to type a question, submit the form and see both the answer and question displayed on the screen.

Excercise 4

The response from OpenAI might take some time to be delivered. That's why adding a loading indicator is a nice touch. You can use the following code block to create a new file called src/components/Loader/Loader.tsx:

src/components/Loader/Loader.tsx
export default function Loader() {
    return (
        <div role="status">
            <svg aria-hidden="true" className="w-8 h-8 text-gray-200 animate-spin dark:text-gray-600 fill-blue-600" viewBox="0 0 100 101" fill="none" xmlns="http://www.w3.org/2000/svg">
                <path d="M100 50.5908C100 78.2051 77.6142 100.591 50 100.591C22.3858 100.591 0 78.2051 0 50.5908C0 22.9766 22.3858 0.59082 50 0.59082C77.6142 0.59082 100 22.9766 100 50.5908ZM9.08144 50.5908C9.08144 73.1895 27.4013 91.5094 50 91.5094C72.5987 91.5094 90.9186 73.1895 90.9186 50.5908C90.9186 27.9921 72.5987 9.67226 50 9.67226C27.4013 9.67226 9.08144 27.9921 9.08144 50.5908Z" fill="currentColor"/>
                <path d="M93.9676 39.0409C96.393 38.4038 97.8624 35.9116 97.0079 33.5539C95.2932 28.8227 92.871 24.3692 89.8167 20.348C85.8452 15.1192 80.8826 10.7238 75.2124 7.41289C69.5422 4.10194 63.2754 1.94025 56.7698 1.05124C51.7666 0.367541 46.6976 0.446843 41.7345 1.27873C39.2613 1.69328 37.813 4.19778 38.4501 6.62326C39.0873 9.04874 41.5694 10.4717 44.0505 10.1071C47.8511 9.54855 51.7191 9.52689 55.5402 10.0491C60.8642 10.7766 65.9928 12.5457 70.6331 15.2552C75.2735 17.9648 79.3347 21.5619 82.5849 25.841C84.9175 28.9121 86.7997 32.2913 88.1811 35.8758C89.083 38.2158 91.5421 39.6781 93.9676 39.0409Z" fill="currentFill"/>
            </svg>
            <span className="sr-only">Loading...</span>
        </div>
    );
}

You can use this component in src/App.tsx to show a loading indicator when you're waiting for the request to OpenAI to resolve.

BONUS: Also add an error state.

The application will now have both a way to ask questions and shows a loading state when the answer is being fetched from OpenAI.

Excercise 5

The way you ask your question makes a huge difference in the response you're getting, maybe you've wondered why our answer is short and snappy rather than a blurb of text. Prompt engineering is a common way to change the format or style of the answer.

By giving the LLM a prompt template together with your question, you can control the format or sentiment of the answer. You don't always want to expose the prompt to the user of the application too.

Imagine we're building a GPT for a travel office, let's add the following prompt template:

Take the role of a personal travel assistant, and answer the following question in detail: {question}

We're going to test this prompt template in the OpenAI Playground first.

Have a look at the LangChainJS docs to see how to implement a prompt template for the generateAnswer function in src/utils/langchain.ts.

Try out the impact of the prompt template on the answer from the LLM. Make sure to update the test case in src/utils/langchain.test.ts too.

Hint: How can you overwrite the prompt instruction in the test case too without mocking?

Excercise 6

The way you prompt the LLM isn't the only way to change the answer of the LLM, another thing we can do is changing the temperature or by using a different model.

You can modify these values in src/utils/langchain.ts:

src/utils/langchain.ts
const llm = new OpenAI({
    openAIApiKey: import.meta.env.VITE_OPENAI_KEY,
    temperature: 0.9, // Can be between 0 and 1
    modelName: "gpt-4-0125-preview", // Default. Other options: https://platform.openai.com/docs/models/
    maxTokens: 300 // length of response, tokens !== characters
});

Play around with different values, both in your code and the OpenAI playground. How does this impact the quality or style of the answer?

Excercise 7

The above is an example of a "zero shot" prompt. We didn't provide the LLM with any context besides what role to take. Therefore we assumed the LLM knows what a travel agent is, but sometimes the model has no information on your question or needs additional context.

Before implementing a new type of prompting, we'll need to implement a chat model:

src/utils/langchain.ts
```ts
import { ChatOpenAI } from "@langchain/openai";
import { ChatPromptTemplate } from "@langchain/core/prompts";

const llm = new ChatOpenAI({
    openAIApiKey: import.meta.env.VITE_OPENAI_KEY,
    temperature: 1,
    modelName: "gpt-4-0125-preview",
});

export async function generateAnswer(
    question: string,
    promptTemplate: string = "Take the role of a {role}, that answers questions in a {style} way.",
    role: string = "Personal travel assistant",
    style: string = "consistent" // What happens if you change this to detailed?
) {
    let answer = ''

    const chatPrompt = ChatPromptTemplate.fromMessages([
        ["system", promptTemplate],
        ["human", "{question}"],
    ]);

    const formattedPrompt = await chatPrompt.formatMessages({
        role,
        style,
        question
    });

    try {
        const result = await llm.invoke(formattedPrompt);
        answer = result?.content as string;
    } catch (e) {
        return 'Something went wrong';
    }

    return answer;
}

```

In the above setup we made it easier to change the input variables, and by using a Chat model instead of LLM model we can start implementing different prompting techniques. You might see there's a human and system template, as in the Chat model subsequent messages are being used as context.

Fix your test so it will continue to run.

Excercise 8

We can also try "few shot prompting" where we give the LLM some examples before asking our question, try this example in the OpenAI playground.

Let's start by adding a few shot prompting technique:

src/utils/langchain.ts
import { ChatOpenAI } from "@langchain/openai";
import { ChatPromptTemplate, FewShotChatMessagePromptTemplate } from "@langchain/core/prompts";

const llm = new ChatOpenAI({
    openAIApiKey: import.meta.env.VITE_OPENAI_KEY,
    temperature: 1,
    modelName: "gpt-4-0125-preview",
});

export async function generateAnswer(
    question: string,
    promptTemplate: string = "Take the role of a Personal travel assistant, that answers questions in a consistent way."
) {
    let answer = '';

    const examples = [
        {
            input: "What are the best restaurants in Amsterdam?",
            output: "The highest rated restaurants in Amsterdam are (1), (2), (3)",
        },
        {
            input: "What is the best time of the year to visit The Netherlands?",
            output: "Summer",
        },
    ];

    const examplePrompt = ChatPromptTemplate.fromTemplate(`User: {input}
Assistant: {output}`);

    const fewShotPrompt = new FewShotChatMessagePromptTemplate({
        prefix: promptTemplate,
        suffix: "User: {input} Assistant:",
        examplePrompt,
        examples,
        inputVariables: ["input"],
    });

    const formattedPrompt = await fewShotPrompt.format({
        input: question,
    });

    try {
        const result = await llm.invoke(formattedPrompt);
        answer = result?.content as string;
    } catch (e) {
        console.log(e);
        return 'Something went wrong';
    }

    return answer;
}

Ask a question like "What are the best museums in amsterdam?" and the response should match the format of the examples. Try for yourself, see the difference when you change the provided examples.

BONUS: Edit the application to allow follow-up questions by passing the chat history.

BONUS: You can also implement few shot prompting without using chat, for this you can use prompt pipelines.

Excercise 9

Next to few-shot prompts or adding the chat history as context, you can also load data from external sources and pass it to the LLM.

In the directory src/public you can find a file called data.txt that contains a blog post about the best highlights in Amsterdam.

We don't want to pass the entire file to the LLM, as this can lead to overload when you have a lot of data. Instead, we need to take the most important parts of our data for whihch we would need a vector database.

We'll use a local in-memory vectorstore as this is a demo environment, by making these changes in src/utils/langchain.ts:

src/utils/langchain.ts
// ...
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai";
import { ChatPromptTemplate, FewShotChatMessagePromptTemplate } from "@langchain/core/prompts";
import { RecursiveCharacterTextSplitter } from "langchain/text_splitter";
import { MemoryVectorStore } from "langchain/vectorstores/memory";

// ...

let vectorStore: MemoryVectorStore;

export async function generateAndStoreEmbeddings() {
    const trainingText = await fetch("/data.txt")
        .then((response) => response.text())
        .then((text) => text);

    const textSplitter = new RecursiveCharacterTextSplitter({
        chunkSize: 1000,
    });

    const docs = await textSplitter.createDocuments([trainingText]);

    vectorStore = await MemoryVectorStore.fromDocuments(
        docs,
        new OpenAIEmbeddings({ openAIApiKey: import.meta.env.VITE_OPENAI_KEY }),
    );
}

// Everything else ...

In src/App.tsx we need to load this data on the first render:

src/App.tsx
// src/App.tsx
import { useEffect, useState } from "react";
import { generateAnswer, generateAndStoreEmbeddings } from "./utils/langchain";
import Message from "./components/Message/Message";
import Loader from "./components/Loader/Loader";

export default function App() {
    const [question, setQuestion] = useState("");
    const [result, setResult] = useState({ question: "", answer: "" });
    const [loading, setLoading] = useState(false);

    // 1. Load data into vector store

    // Everything else ...

}

The next step is to create a new function for the generateAnswer function to use the data stored in the vectorstore:

src/utils/langchain.ts
// ...
import { createRetrievalChain } from "langchain/chains/retrieval";
import { createStuffDocumentsChain } from "langchain/chains/combine_documents";

// ...

export async function generateAnswerRAG(question: string) {
    let answer = '';

    const prompt = ChatPromptTemplate.fromTemplate(`
Answer the following question based only on the provided context:

<context>
{context}
</context>

Question: {input}`
    );

    const documentChain = await createStuffDocumentsChain({
        llm,
        prompt,
    });

    const retriever = vectorStore.asRetriever();

    const retrievalChain = await createRetrievalChain({
        combineDocsChain: documentChain,
        retriever,
    });

    try {
        const result = await retrievalChain.invoke({
            input: question,
        });
        answer = result?.answer;
    } catch (e) {
        console.log(e);
        return 'Something went wrong';
    }

    return answer;
}

And finally, use this new function in src/App.tsx.

If you ask a question now, it will inject the data from the document. Try this out with multiple (follow-up) questions.

BONUS: Add the prompt template for few shot prompting back into this new function.

Where to go from here

  • Document Loaders
  • RAG (Retrieval Augmented Generation)
  • Agents

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published