“Optimizing LLM Deployments through Inference Backends”. Journal of Artificial Intelligence & Cloud Computing, vol. 3, no. 4, July 2024, pp. 1-4, https://doi.org/10.47363/JAICC/2024(3)E128.