-
A Quick Mental Model for Estimating LLM GPU Memory Use
Before downloading a large model or spinning up a container, it’s useful to know whether an open-source LLM will actually fit on your GPU.
-
Designing a Maintainable Replay Buffer in RL Systems
A structured and engineering-focused reflection on replay buffer design in RL, emphasizing clarity, extensibility, and long-term maintainability.
-
Tracing the Root Cause of Missing GPUs in Docker Containers
A debugging record of why Docker refused to expose GPUs inside a container even though the host recognized them perfectly, and how every layer of the system contributed a small piece to the failure.
-
A Measure of Range Compression on Different Board Textures
A mathematical approach to quantifying how different flop textures compress or expand pre-flop ranges in No-Limit Hold’em.
-
Running dm-control on a Headless Server: A Complete Debugging Log
A practical record of configuring dm-control with Mujoco on a headless Ubuntu server, covering rendering failures, version mismatches, and the final workable setup.