{
  "autoDestroy": false,
  "bundlePath": "/home/mnm/workspaces/runpod-devenv-module/templates/llm-inference/.devenv/state/runpod/bundles/llm-inference-e2e-2.tar.gz",
  "createdAt": "2026-03-19T00:09:15.899574+00:00",
  "destroyResponse": {
    "podId": "1m801dvtfux2j2",
    "terminated": true,
    "terminatedAt": "2026-03-19T00:13:59.333390+00:00"
  },
  "destroyedAt": "2026-03-19T00:13:59.333412+00:00",
  "diskGb": 160,
  "gpuTypes": [
    "NVIDIA RTX A5000"
  ],
  "image": "runpod/pytorch:1.0.2-cu1281-torch280-ubuntu2404",
  "jobId": "llm-inference-e2e-2",
  "podCreateResponse": {
    "costPerHr": 0.27,
    "desiredStatus": "RUNNING",
    "id": "1m801dvtfux2j2",
    "imageName": "runpod/pytorch:1.0.2-cu1281-torch280-ubuntu2404",
    "name": "runpod-runner-llm-inference-e2e-2",
    "podType": "RESERVED",
    "runtime": null
  },
  "podId": "1m801dvtfux2j2",
  "podType": "RESERVED",
  "ports": "22/tcp,8000/http",
  "profilePath": "/nix/store/c5w0sqm0vqcgds7kfl5cng5q3xl2k4sw-runpod-profile-serve.json",
  "remoteBundlePath": "/workspace/llm-inference-e2e-2.tar.gz",
  "remoteEnvKeys": [
    "MODEL_NAME",
    "PROBE_PROMPT",
    "PYTHONUNBUFFERED",
    "SERVE_COMMAND",
    "SIMULATE_ONLY",
    "VLLM_GPU_MEMORY_UTILIZATION",
    "VLLM_HOST",
    "VLLM_INSTALL_COMMAND",
    "VLLM_MAX_MODEL_LEN",
    "VLLM_PORT",
    "VLLM_READINESS_TIMEOUT_SECONDS",
    "VLLM_VENV_DIR"
  ],
  "remoteWorkspaceRoot": "/workspace/llm-inference-e2e-2",
  "runId": "runpod-20260319T000915Z-d36dcaa5",
  "sshHost": "69.30.85.240",
  "sshPort": 22182,
  "sshReadyAt": "2026-03-19T00:09:57.772477+00:00",
  "sshUser": "root",
  "timeoutMinutes": 90,
  "workspaceDir": "/home/mnm/workspaces/runpod-devenv-module/templates/llm-inference/workspace"
}
