{
  "status": "completed",
  "message": "vLLM server is running and the probe request succeeded",
  "updatedAt": "2026-03-18T21:59:18Z",
  "template": "llm-inference",
  "model": "Qwen/Qwen2.5-0.5B-Instruct",
  "port": "8000"
}
