“Optimizing LLM Deployments through Inference Backends” (2024) Journal of Artificial Intelligence & Cloud Computing, 3(4), pp. 1–4. doi:10.47363/JAICC/2024(3)E128.