Optimizing for Low-Latency Communication in Inference Workloads with JAX and XLA