Skip to content

llama-fit-params: free memory target per device#18679

Merged
JohannesGaessler merged 1 commit intoggml-org:masterfrom
JohannesGaessler:llama-fp-targets-per-gpu
Jan 8, 2026
Merged

llama-fit-params: free memory target per device#18679
JohannesGaessler merged 1 commit intoggml-org:masterfrom
JohannesGaessler:llama-fp-targets-per-gpu

Conversation

@JohannesGaessler
Copy link
Copy Markdown
Contributor

Fixes #18390 .

This PR extends llama_params_fit with the option to specify a memory target per device instead of only a single one. If a user specifies only a single number this number is broadcast across all devices.

@JohannesGaessler JohannesGaessler merged commit 64848de into ggml-org:master Jan 8, 2026
75 checks passed
gary149 pushed a commit to gary149/llama-agent that referenced this pull request Jan 8, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Feature Request: Allow per-device memory margin for --fit-target

2 participants