A developer wrote a Llama 2 inference engine using less than 1,500 bytes of x86 assembly. This project, sectorllm, strips away all high-level abstractions to execute model weights directly. It serves as a technical curiosity rather than a production tool. Practitioners can use it to study the absolute minimum requirements for LLM execution.