SHA-1, SHA-256 and SHA-512 compression functions using Intel, ARMv8 and Power8 SHA intrinsics
-
Updated
Jan 8, 2026 - C
SHA-1, SHA-256 and SHA-512 compression functions using Intel, ARMv8 and Power8 SHA intrinsics
NUMA-distributed weight banking for LLM inference on IBM POWER8. 147 t/s (8.8x stock). Part of the Proof of Physical AI stack.
llama.cpp optimizations for IBM POWER8: vec_perm non-bijunctive collapse, PSE hardware entropy, DCBT prefetch. Sovereign inference. Part of the Proof of Physical AI stack.
AES encryption function using Intel, ARMv8 and Power8 intrinsics
Native ppc64le modules for llama.cpp webui (lightningcss, tailwindcss-oxide) - Built on IBM POWER8
Non-bijunctive attention collapse for LLM inference — POWER8 hardware AES (vcipher) + AltiVec vec_perm. Hebbian path selection, cross-head diffusion, O(1) KV prefiltering.
Accelerate LLM inference by collapsing attention paths with hardware-optimized selective pruning using POWER8 vector instructions and crypto operators.
Add a description, image, and links to the power8 topic page so that developers can more easily learn about it.
To associate your repository with the power8 topic, visit your repo's landing page and select "manage topics."