{
  "autoDestroy": false,
  "bundlePath": "/home/mnm/workspaces/runpod-devenv-module/templates/llm-inference/.devenv/state/runpod/bundles/llm-inference-e2e-3.tar.gz",
  "bundleUploadedAt": "2026-03-19T00:14:25.495537+00:00",
  "createdAt": "2026-03-19T00:14:00.539029+00:00",
  "destroyResponse": {
    "podId": "cj2prf6elif88k",
    "terminated": true,
    "terminatedAt": "2026-03-19T00:22:12.430306+00:00"
  },
  "destroyedAt": "2026-03-19T00:22:12.430324+00:00",
  "diskGb": 160,
  "downloadedAt": "2026-03-19T00:22:11.559913+00:00",
  "gpuTypes": [
    "NVIDIA RTX A5000"
  ],
  "image": "runpod/pytorch:1.0.2-cu1281-torch280-ubuntu2404",
  "jobId": "llm-inference-e2e-3",
  "localOutputDir": "/home/mnm/workspaces/runpod-devenv-module/templates/llm-inference/.devenv/state/runpod/outputs/runpod-20260319T001359Z-14e60836",
  "podCreateResponse": {
    "costPerHr": 0.27,
    "desiredStatus": "RUNNING",
    "id": "cj2prf6elif88k",
    "imageName": "runpod/pytorch:1.0.2-cu1281-torch280-ubuntu2404",
    "name": "runpod-runner-llm-inference-e2e-3",
    "podType": "RESERVED",
    "runtime": null
  },
  "podId": "cj2prf6elif88k",
  "podType": "RESERVED",
  "ports": "22/tcp,8000/http",
  "profilePath": "/nix/store/cv5fyaspabl31j6xjcsyzx4k96y88cw7-runpod-profile-serve.json",
  "remoteBundlePath": "/workspace/llm-inference-e2e-3.tar.gz",
  "remoteEnvKeys": [
    "MODEL_NAME",
    "PROBE_PROMPT",
    "PYTHONUNBUFFERED",
    "SERVE_COMMAND",
    "SIMULATE_ONLY",
    "VLLM_GPU_MEMORY_UTILIZATION",
    "VLLM_HOST",
    "VLLM_INSTALL_COMMAND",
    "VLLM_MAX_MODEL_LEN",
    "VLLM_PORT",
    "VLLM_READINESS_TIMEOUT_SECONDS",
    "VLLM_VENV_DIR"
  ],
  "remoteWorkspaceRoot": "/workspace/llm-inference-e2e-3",
  "runId": "runpod-20260319T001359Z-14e60836",
  "sshHost": "69.30.85.240",
  "sshPort": 22043,
  "sshReadyAt": "2026-03-19T00:14:21.760547+00:00",
  "sshUser": "root",
  "timeoutMinutes": 90,
  "workspaceDir": "/home/mnm/workspaces/runpod-devenv-module/templates/llm-inference/workspace"
}
