Which H100 instance to train Nanochat
Benchmarking H100 PCIe vs SXM vs NVL on training cost, step times, and NCCL profiling to find the cheapest GPU configuration for Nanochat
Latest thoughts on LLMs, engineering, and data science
Benchmarking H100 PCIe vs SXM vs NVL on training cost, step times, and NCCL profiling to find the cheapest GPU configuration for Nanochat
Enabling endless capabilities for LLMs
Validate the data as it streams, don't make your users wait.