Skip to content

peterhj/cacti

cacti

cacti is a library for experimenting with computation graphs (or computation "spines"). cacti is written in the Rust language, and uses the Futhark language to implement computational kernels targeting GPUs and multicore CPUs.

The current pre-release of cacti is capable of larger-than-VRAM training or fine-tuning of LLaMA-style language models, using the full-precision gradient (e.g. fp16). In other words, using cacti, there is no need to invoke any specialized optimizer to enable larger-than-VRAM training or fine-tuning; the underlying dataflow system of cacti will do its best to utilize the available hardware resources, based on your system's GPU memory and host CPU memory capacities. cacti achieves this through an out-of-memory policy that aggressively garbage-collects those dataflow cells (i.e. "tensors") that are determined to be unreachable via a static analysis, and opportunistically spills other cells (that cannot be garbage-collected) from the GPU to the host CPU memory. (Note that you still need enough host CPU memory capacity; otherwise it is possible to observe OOM on the host CPU side.)

cacti is oriented toward developing memory-safe AI systems, and so the cacti system code is entirely implemented in Rust. The main non-Rust component is the Futhark language and compiler, which is implemented in Haskell, and which is used extensively in cacti to implement the underlying computation kernels of the computation graph ("spine"). In fact, Futhark is essential to the design of cacti, and some design choices in Futhark are likewise visible in cacti.

Design Priorities

  • Purely functional/monotone update semantics
  • Operational semantics specifying safe dataflow and autodiff
  • Programming model: cyclic dataflow + coroutine (reset, compile, resume, yield_)
  • Computational kernels: first, write it in Futhark
  • Written for Rust

Tradeoffs and Limitations

As this is a pre-release of cacti, there are a number of known limitations due to tradeoffs made in prioritizing what to implement first. Given finite development resources, maintaining the design priorities listed earlier took precedence over significant extensions to the system architecture, or other potential new features, such as the following:

  • Implementations for GPU targets other than CUDA
  • Multi-GPU
  • Quantization
  • Python inter-op

Installation

Prerequisites

Requirements:

  • git
  • make
  • ghc and cabal-install (minimum: GHC >= 9.0)
  • rustc and cargo (tested with Rust >= 1.62; some older versions should also compile)
  • gcc or clang, for Futhark dynamic compilation (tested with gcc)

Additional requirements for sentencepiece support (the default configuration):

  • cmake
  • g++ or clang++

For cacti GPU support, the following CUDA configurations are known to work:

  • CUDA 11.6 update 2, driver 510
  • CUDA 11.5 update 2, driver 495

Thus far cacti has also been built on the following system configurations:

  • Debian Bookworm (x86_64-gnu-linux, GCC 12, GHC 9.0, Rust 1.63)
  • Debian Bullseye (x86_64-gnu-linux, GCC 10)
    • GHC 9.0.2 installed via ghcup
    • Rust 1.71.1 installed via rustup

Bootstrapping and installing from source (recommended method)

It is recommended to use the cacti-bootstrap git repository to bootstrap sources, as that repo vendors git submodules of all Rust dependencies, as well as our patched branch of the Futhark compiler. However, please note that Futhark itself depends on many Haskell packages which we do not vendor, and are instead downloaded by cabal during bootstrapping.

git clone 'https://github.com/peterhj/cacti-bootstrap'
cd cacti-bootstrap
./bootstrap.sh
cd cacti
make

Bootstrapping and installing from source (alternative method)

An alternative bootstrapping method, for developing on cacti itself, uses the provided bootstrap.sh script to git clone HEAD on all of the vendored dependencies. Please note that this bootstrapping method will create a bunch of cloned repo directories outside the cacti repo directory, thus it is recommended to perform this in a dedicated workspace directory.

mkdir <your-workspace-dir>
cd <your-workspace-dir>
git clone 'https://github.com/peterhj/cacti'
cd cacti
./bootstrap.sh
make

Examples

In the "examples" directory, you will find provided code for both fine-tuning and inference based on OpenLLaMA-3B. The two example files, "open_llama_3b_deploy.rs" and "open_llama_3b_train.rs", use cacti as a library and are otherwise self-contained examples; the first an example of inference, and the second an example of full-precision, full-gradient fine-tuning.

Please note that the fine-tuning example ("open_llama_3b_train.rs") may require 64 GB of host CPU RAM to run using the malloc allocator, and up to 96-128 GB when using the pagelocked allocator.

It is recommended to read and understand the examples, and to use them as starting points for your own experiments.

Documentation

Environment variables

cacti will inspect the following environment variables to control its run-time behavior.

  • CACTI_BIN_PATH: This is the path to a directory in which to search for the cacti-futhark binary, which was installed by cabal-install when bootstrapping from source. If this variable was not specified, the default value is ${HOME}/.cabal/bin where ${HOME} is the current user's home directory.
  • CACTI_CACHE_PATH: This is the path to a directory in which cacti will store run-time build artifacts of the Futhark compiler, which is used in cacti to compile and run computation kernels. If this variable was not specified, the default value is ${HOME}/.cacti/cache where ${HOME} is the current user's home directory.
  • CACTI_CUDA_PREFIX: This is a colon-delimited list of paths to search for a CUDA installation. If this variable was not specified, the default value is /usr/local/cuda.
  • CACTI_VMEM_SOFT_LIMIT: Set this to either a specific size (bytes/GB/GiB/etc.) or a fraction (of the total GPU VRAM). Then, the GPU subsystem will pretend as if that were the physical limit of GPU VRAM, and make garbage-collection/OOM decisions accordingly. Note that setting CACTI_VMEM_SOFT_LIMIT=1 will attempt to reserve all of the available GPU VRAM.
  • CACTI_NVGPU_MEM_ALLOC: This specifies which CUDA-aware allocator is used for host CPU memory. Allowed values are malloc and pagelocked (the latter corresponding to cuMemAllocHost). Note that the CUDA page-locked memory limit seems to be capped at a percentage of the total system memory capacity, so using it may cause surprising host CPU memory OOMs.
  • CACTI_VERBOSE: Setting this will increase the verbosity of the stdout logging.

Reference (todo)

Please check back soon; this is a work in progress.

About

No description, website, or topics provided.

Resources

License

Apache-2.0 and 2 other licenses found

Licenses found

Apache-2.0
LICENSE-APACHE-2.0
ISC
LICENSE-ISC
ISC
LICENSE-futhark

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages