A (simple) command-line to estimate inference memory requirements on Hugging Face
cargo install hf-mem
And then:
hf-mem --model-id meta-llama/Llama-3.1-8B-Instruct --token ...
- Fast and light command-line, with a single installable binary
- Fetches just the required bytes from the
safetensors
files on the Hugging Face Hub that contain the metadata - Provides an estimation based on the count of the parameters on the different dtypes
- Supports both sharded i.e.
model-00000-of-00000.safetensors
and not sharded i.e.model.safetensors
files
- Add tracing and progress bars when fetching from the Hub
- Support other file types as e.g.
gguf
- Read metadata from local files if existing, instead of just fetching from the Hub every single time
- Add more flags to support estimations assuming quantization, extended context lengths, any added memory overhead, etc.
This project is licensed under either of the following licenses, at your option:
Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in this project by you, as defined in the Apache-2.0 license, shall be dual licensed as above, without any additional terms or conditions.