AirLLM enables 70B large language model inference on a single 4GB GPU, making large model inference accessible without expensive hardware.

Fund this project

Unverified URL

The funding manifest has not provided proof via wellKnown that this link is associated with it. Learn more.

Continue