Skip to content

feat: NUMA-aware model sharding package for POWER8 llama.cpp (issue #2277)#1799

Merged
Scottcjn merged 1 commit intoScottcjn:mainfrom
createkr:feat/issue2277-power8-numa-sharding
Mar 25, 2026
Merged

feat: NUMA-aware model sharding package for POWER8 llama.cpp (issue #2277)#1799
Scottcjn merged 1 commit intoScottcjn:mainfrom
createkr:feat/issue2277-power8-numa-sharding

Conversation

@createkr
Copy link
Contributor

Implements a NUMA-aware sharding package for llama.cpp with layer routing, benchmark harness, reproducible presets, validation reports, and integration docs.\n\nThis submission is the working sharding package intended for the partial payout track while waiting for direct POWER8 hardware benchmark execution.\n\nValidation:\n- artifact package + benchmark harness included\n- tuning presets included\n- integration docs included\n\nCloses #2277

@github-actions github-actions bot added documentation Improvements or additions to documentation BCOS-L1 Beacon Certified Open Source tier BCOS-L1 (required for non-doc PRs) labels Mar 23, 2026
@createkr
Copy link
Contributor Author

RTC1d48d848a5aa5ecf2c5f01aa5fb64837daaf2f35

@github-actions github-actions bot added the size/XL PR: 500+ lines label Mar 23, 2026
@Scottcjn Scottcjn merged commit c296d67 into Scottcjn:main Mar 25, 2026
4 of 6 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

BCOS-L1 Beacon Certified Open Source tier BCOS-L1 (required for non-doc PRs) documentation Improvements or additions to documentation size/XL PR: 500+ lines

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants