HW: LLMs CPU+GPU platform

66 阅读7分钟

HW: LLMs!
1/16
HW: LLMs, vectors, RAG :)
Summary
In this fi nal HW, you will:
use Weaviate [ ], which is a vector DB - stores data as vectors after vectorizing, and computes a search query by vectorizing it and does
similarity search with existing vectors
crawl the web using a Node package, to compile a 'knowledge base' [to use subsequently (not part of the hw) as input to build a (!)]
using a Python module, perform RAG [retrieval augmentation] on a 'small', locally-hosted LLM [make that an 'S'LM :)]
use to run RAG on their CPU+GPU platform
These are cutting-edge techniques to know, from a POV :) Plus, they are simply, FUN!!
Please make sure you have these installed, before starting: git, Docker, Node, Python (or Conda/Anaconda), [with 'Desktop development with C++'
checked].
Note: you need to do all four, Q1..Q4 (not pick just one!) :)
Q1.
Description
We are going to use vector-based similarity search, to retrieve search results that are not keyword-driven.
The (three) steps we need are really simple:
install Weaviate plus vectorizer via Docker as images, run them as containers
specify a schema for data, upload data/knowledge (in .json format) to have it be vectorized
run a query 代 写HW: LLMs CPU+GPU platform (which also gets vectorized and then sim-searched), get back results (as JSON)
The following sections describe the above steps.

  1. Installing Weaviate and a vectorizer module
    After installing Docker, bring it up (eg. on Windows, run Docker Desktop). Then, in your (ana)conda shell, run this docker-compose
    command that uses this 'docker-compose.yml' confi g fi le to pull in two images: the 'weaviate' one, and a text2vec transformer
    called 't2v-transformers':
    docker-compose up -d
    weaviate.io/
    custom GPT
    lightning.ai
    future/career
    VS 2022
    yaml
    HW: LLMs!
    2/16
    These screenshots show the progress, completion, and subsequently, two containers automatically being started (one for weaviate,
    one for t2v-transformers):
    HW: LLMs!
    3/16
    Yeay! Now we have the vectorizer transformer (to convert sentences to vectors), and weaviate (our vector DB search engine)
    running! On to data handling :)
  2. Loading data to search for
    This is the data (knowledge, aka external memory, ie. prompt augmentation source) that we'd like searched, part of which will get
    returned to us as results. The data is represented as an array of JSON documents. is our data file, conveniently named
    data.json (you can rename it if you like) [you can visualize it better using ] - place it in the 'root' directory of
    your webserver (see below). As you can see, each datum/'row'/JSON contains three k:v pairs, with 'Category', 'Question', 'Answer' as
    keys - as you might guess, it seems to be in Jeopardy(TM) answer-question (reversed) format :) The fi le is actually called
    , I simply made a local copy called data.json.
    The overall idea is this: we'd get the 10 documents vectorized, then specify a query word, eg. 'biology', and automagically have that
    pull up related docs, eg. the 'DNA' one (even if the search result doesn't contain 'biology' in it)! This is a really useful semantic search
    feature where we don't need to specify exact keywords to search for.
    Start by installing the weaviate Python client:
    pip install weaviate-client
    So, how to submit our JSON data, to get it vectorized? Simply use Python script, do:
    python weave-loadData.py
    Here
    jsoncrack.com
    jeopardy?tiny.json
    this
    HW: LLMs!
    4/16
    You will see this:
    If you look in the script, you'll see that we are creating a schema - we create a class called 'SimSearch' (you can call it something else
    if you like). The data we load into the DB, will be associated with this class (the last line in the script does this via add_data_object()).
    NOTE - you NEED to run a local webserver [in a separate ana/conda (or other) shell], eg. via python - it's what will 'serve'
    data.json to weaviate :)
    Great! Now we have specifi ed our searchable data, which has been fi rst vectorized (by 't2v-transformers'), then stored as vectors (in
    weaviate).
    Only one thing left: querying!
  3. Querying our vectorized data
    To query, use this simple shell script called , and run this:
    sh weave-doQuery.sh
    As you can see in the script, we search for 'physics'-related docs, and sure enough, that's what we get:
    'serveit.py'
    weave-doQuery.sh
    HW: LLMs!
    5/16
    Why is this exciting? Because the word 'physics' isn't in any of our results!
    Now it's your turn:
    ? fi rst, MODIFY the contents of data.json, to replace the 10 docs in it, with your own data, where you'd replace ("Category","Question","Answer") with
    ANYTHING you like, eg. ("Author","Book","Summary"), ("MusicGenre","SongTitle","Artist"), ("School","CourseName","CourseDesc"), etc, etc - HAVE fun
    coming up with this! You can certainly add more docs, eg. have 20 of them instead of 10
    ? next, MODIFY the query keyword(s) in the query .sh fi le - eg. you can query for 'computer science' courses, 'female' singer, 'American' books,
    ['Indian','Chinese'] food dishes (the query list can contain multiple items), etc. Like in the above screenshot, 'cat' the query, then run it, and get a
    screenshot to submit. BE SURE to also modify the data loader .py script, to put in your keys (instead of ("Category","Question","Answer"))
    That's it, you're done :) In RL you will have a .json or fi le (or data in other formats) with BILLIONS of items! Later, do feel free to
    play with bigger JSON fi les, eg. this Jeopardy JSON fi le :)
    FYI/'extras'
    Here are two more things you can do, via 'curl':
    [you can also do ' ' in your browser]
    [you can also do ' ' in your browser]
    .csv
    200K
    http://localhost:8080/v1/meta
    http://localhost:8080/v1/schema
    HW: LLMs!
    6/16
    Weaviate has a cloud version too, called - you can try that as an alternative to using the Dockerized version:
    Run :)
    Also, for fun, see if you can print the raw vectors for the data (the 10 docs)...
    More info:
    ?
    ?
    ?
    ?
    Q2.
    You are going to run a crawler on a set of pages that you know contain 'good' data - that could be used by an LLM to answer
    questions 'intelligently' (ie. not confabulate, ie not 'hallucinate', ie. not make up BS based on its core, general-purpose pre-training!).
    The crawled results get conveniently packaged into a single output.json fi le. For this qn, please specify what group of pages you
    WCS
    this
    weaviate.io/developers/…
    weaviate.io/developers/…
    medium.com/semi-techno…
    weaviate.io/developers/…
    HW: LLMs!
    7/16
    crawled [you can pick any that you like], and, submit your output.json (see below for how to generate it).
    Take a look:
    You'll need to git-clone 'gpt-crawler' from . Then do 'npm install' to download the needed
    Node packages. Then edit confi g.ts [ ] to specify your crawl path,
    then simply run the crawler via npm.start! Voila - a resulting output.json, after the crawling is completed.
    For this hw, you'll simply submit your output.json - but its true purpose is to serve as input for a cstom GPT :)
    From builder.io's GitHub page:
    Amazing! You can use this to create all sorts of SMEs [subject matter experts] in the future, by simply scraping existing docs on the
    web.
    Q3.
    For this question, you are going to download a small (3.56G) model (with 7B parameters, compared to GPT-4's 1T for ex!), and use it
    along with an external knowledge source (a simple text fi le) vectorized using Chroma (a popular vector DB), and ask questions
    github.com/BuilderIO/g…
    github.com/BuilderIO/g… g.ts
    HW: LLMs!
    8/16
    whose answers would be found in the text fi le :) Fun!
    git clone this: - and cd into it. You'll see a Python script (app.py) and a requirements.txt fi le.
    Install pipenv:
    github.com/afaqueumer/…
    HW: LLMs!
    9/16
    Install the required components (Chroma, LangChain etc) like so:
    Turns out we need a newer version of llama-cpp-python, one of the modules we just installed - so do this:
    HW: LLMs!
    10/16
    Next, let's grab this LLM: huggingface.co/TheBloke/Ll… - and save it to a
    HW: LLMs!
    11/16
    models/folder inside your DocQA one:
    Modify app.py to specify this LLM:
    If you are curious about the .gguf format used to specify the LLM, read
    Now we have all the pieces! These include the req'd Python modules, the LLM, and an app.py that will launch a UI via . Run
    this.
    'Streamlit'
    HW: LLMs!
    12/16
    this [pipenv run streamlit run app.py]:
    OMG - our chat UI in a browser, via a local webserver [the console prints info about the LLM]:
    Now we need a simple text fi le to use for asking questions from (ie. 'external memory' for the LLM). I used
    page, to make fi le, for ex.
    We are now ready to chat with our doc! Upload the .txt, wait a few minutes for the contents to get vectorized and indexed :) When
    www.coursera.org/articles/st… this
    HW: LLMs!
    13/16
    that is done, ask a question - and get an answer! Like so:
    That's quite impressive!
    You would need to create a text fi le of your own [you could even type in your own text, about anything!], upload, ask a question,
    then get a screenshot of the Q and A. You'd submit the text fi le and the screenshot.
    The above is what the new 'magic' (ChatGPT etc) is about!! Later, you can try out , other language tasks, reading
    PDF, etc. Such custom 'agents' are sure to become commonplace, serving/dispensing expertise/advice in myriad areas of life.
    is more, related to Q3.
    Q4.
    This is a quick, easy and useful one!
    Go to and sign up for a free account. Then read these: and
    Browse through their vast collection of 'Studio' templates: - when you create (instantiate) one, you get
    your own sandboxed environment [a 'cloud supercomputer'] that runs on lightning.ai's servers. You get unlimited CPU use, and 22
    hours of GPU use per month (PLENTY, for beginner projects).
    Create this Studio: - you are going to use this
    to do RAG using your own PDF :)
    Upload (drag and drop) your PDF [can be on ANY topic - coding, cooking, crafting, canoeing, cattle-ranching, catfi shing... (lol!)].
    many other models
    Here

 HW: LLMs!
14/16
Eg. this shows the pdf I uploaded:
Next, edit run.ipynb, modify the 'fi les' variable to point to your pdf:
HW: LLMs!
15/16
Modify the 'query' var, and the 'queries' var, to each contain a QUESTION on which you'd like to do RAG, ie. get the answers from the
pdf you uploaded! THIS is the cool part - to be able to ask questions in natural language, rather than search by keyword, or look up
words in the index [if the pdf has an index].
Read through the notebook to understand what the code does and what the RAG architecture is, then run the code! You'll see the
two answers printed out. Submit screenshots of your pdf fi le upload, the two questions, and the two answers. The answers might
not be what you expected (ie might be imprecise, EVEN though it's RAG!) but that's ok - there are techniques to improve the quality
of the retrieval, you can dive into them later.
After the course, DO make sure to run (almost) all the templates! It's painless (zero installation!), fast (GPU execution!) and
useful/informative (well-documented!). It doesn't get more cutting edge than these, for 'IR'. You can even write and run your own
code in a Studio, and publish it as a template for others to use :)
HW: LLMs!
16/16
Getting help
There is a hw4 'forum' on Piazza, for you to post questions/answers. You can also meet w/the TAs, CPs, or me.
Have fun! This is a really useful piece of tech to know. Vector DBs are sure be used more and more in the near future, as a way to
provide 'infi nite external runtime memory' (augmentation) for pretrained LLMs. Read this too:
​​​​​​​WX:codinghelp