Code for the /wAIfu/ Collective project.
Go to file
11b a076746f9d fixup! feat: initial commit 2022-12-18 17:29:15 -03:00
data feat: initial commit 2022-12-17 17:52:34 -03:00
waifu fixup! feat: initial commit 2022-12-18 17:29:15 -03:00
.gitignore chore: enable strict mode on mypy, add isort check 2022-12-17 21:31:41 -03:00
.tool-versions feat: initial commit 2022-12-17 17:52:34 -03:00
LICENSE.md feat: initial commit 2022-12-17 17:52:34 -03:00
README.md feat: initial commit 2022-12-17 17:52:34 -03:00
pdm.lock feat: initial commit 2022-12-17 17:52:34 -03:00
pyproject.toml chore: enable strict mode on mypy, add isort check 2022-12-17 21:31:41 -03:00

README.md

11b's /wAIfu/ Toolbox

Note: This is a very early work-in-progress. Expect the unexpected.

As of the moment I'm writing this, the roadmap for the project's prototype model is basically:

  • Build a dataset
  • Fine-tune a pre-trained language model on that dataset
  • Play around, observe behavior and identify what's subpar
  • Adjust dataset accordingly as to try and address the relevant shortcomings
  • Repeat.

This repository is where I'm versioning all the code I've written to accomplish the above.

In short, here's how it works:

  • We start off with raw datasets (see /waifu/datasets/).
    • These are basically classes reponsible for giving us raw data. They might, for example, download a .zip off the internet, unzip it, read a .json file from in there and then return its contents.
  • Modules then make use of these datasets (/waifu/modules/).
    • These are heavily inspired by the papers that introduced LaMDA and BlenderBot3 (and their relevant supporting papers as well).
    • In general, each module is responsible for using a dataset as an input, and processing that data down into text that will be used in the fine-tuning process.
  • A final data file is produced by concatenating the outputs of all the modules. This file is used as an input for the fine-tuning process.