Hi, I’m not too informed about LLMs so I’ll appreciate any correction to what I might be getting wrong. I have a collection of books I would like to train an LLM on so I could use it as a quick source of information on the topics covered by the books. Is this feasible?

  • ryedaft@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    5 days ago

    Umm, fine-tuning the model that makes the embeddings, right? Or is there an API for messing with the generative AI somewhere? Or are we assuming that newbie has a lot of compute resources? And they would have to use the generative model to create queries for their passages as well, right?

    I would try something like

    Guides | RAGFlow - https://ragflow.io/docs/dev/category/guides

    or a similar tool.

    Edit: not for fine-tuning, just to get started. Local models, RAG, your books are your knowledge base

    • will@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      5 days ago

      Making your own embeddings is for RAG. Most base model providers have standardized on OpenAIs embeddings scheme, but there are many ways. Typically you embed a few tokens worth of data at a time and store that in your vector database. This lets your AI later do some vector math (usually cosine similarity search) to see how similar (related) the embeddings are to each other and to what you asked about. There are fine tuning schemes where you make embeddings before the tuning as well but most people today use whatever fine tuning services their base model provider offers, which usually has some layers of abstraction.