<ul data-eligibleForWebStory="true">Hyperdimensional Computing (HDC) is a computing paradigm using high-dimensional hypervectors.Recent HDC methods focus on iterative training for improved accuracy, accelerated on GPUs.Efficient HDC inference has mostly been on specialized hardware, not multi-core CPUs.ScalableHD is proposed for high-throughput HDC inference on multi-core CPUs.ScalableHD uses a two-stage pipelined execution model parallelized across cores.Intermediate results are streamed between stages to enhance cache locality.Features like memory tiling and NUMA-aware worker-to-core binding are integrated for performance.ScalableHD has variants for small and large batch sizes to exploit compute parallelism.It achieves up to 10x speedup over TorchHD, maintaining accuracy for tasks like image classification.ScalableHD shows robust scalability with throughput improvements as cores increase.