Effectively benchmarking OCI Compute Shapes for LLM inference serving