Maximize LLM Throughput: vLLM vs TGI with Batching