{
  "repodata": {
    "build": "gpu_cuda128_py310hb9f6058_303",
    "build_number": 303,
    "depends": [
      "__glibc >=2.28,<3.0.a0",
      "_openmp_mutex >=4.5",
      "_openmp_mutex >=5.1",
      "blas 1.0 mkl",
      "cuda-cudart >=12.8.90,<13.0a0",
      "cuda-cupti >=12.8.90,<13.0a0",
      "cuda-nvrtc >=12.8.93,<13.0a0",
      "cuda-nvtx >=12.8.90,<13.0a0",
      "cuda-version >=12.8,<13",
      "cudnn >=9.15.1.9,<10.0a0",
      "filelock",
      "fsspec",
      "intel-openmp >=2025.0.0,<2026.0a0",
      "jinja2",
      "libabseil * cxx17*",
      "libabseil >=20260107.0,<20260108.0a0",
      "libcublas >=12.8.4.1,<13.0a0",
      "libcudnn >=9.15.1.9,<10.0a0",
      "libcufft >=11.3.3.83,<12.0a0",
      "libcurand >=10.3.9.90,<11.0a0",
      "libcusolver >=11.7.3.90,<12.0a0",
      "libcusparse >=12.5.8.93,<13.0a0",
      "libgcc >=14",
      "libprotobuf >=6.33.5,<6.33.6.0a0",
      "libstdcxx >=14",
      "libtorch 2.8.0.*",
      "libuv >=1.52.0,<2.0a0",
      "magma >=2.9.0,<3.0a0",
      "mkl >=2025.0.0,<2026.0a0",
      "mkl-service >=2.3.0,<3.0a0",
      "nccl >=2.21.5.1,<3.0a0",
      "networkx",
      "numpy >=1.21,<3",
      "numpy >=1.24.0,<3.0.0",
      "opentelemetry-api",
      "python >=3.10,<3.11.0a0",
      "setuptools",
      "sleef >=3.5.1,<4.0a0",
      "sympy >=1.13.3",
      "triton 3.4.0.*",
      "typing_extensions"
    ],
    "license": "BSD-3-Clause",
    "license_family": "BSD",
    "md5": "3a78fd3f039051cc0f40b937f0470bc5",
    "name": "pytorch",
    "sha256": "3556aa37bcc3cb5467d37f0342658d7c6d78d1346fb06b8cd3ac7c163f206173",
    "size": 35005138,
    "subdir": "linux-64",
    "timestamp": 1772926234156,
    "version": "2.8.0"
  },
  "s3": "builds/ci/prefect/Yhb51jdfTWeikYoKTMRCtw/1772928085/linux-64/pytorch-2.8.0-gpu_cuda128_py310hb9f6058_303.tar.bz2",
  "signatures": {
    "7e3910a4b96ef2fe7242b10587a47039c8924fadf98a69503d63445e88b984b3": "bb7a71f85239986b18ca28cd5ba020cdd3771edf0ac2dc1ad5f7353f3a118e58c2c9169e88ee79bb9a7ae2820a8b0eb10e5a19262e4cf953ddeb5309a9875605"
  }
}