{
  "info": {
    "subdir": "linux-64"
  },
  "packages": {
    "flash-attn-2.8.3-py310hb4df0bb_1.tar.bz2": {
      "build": "py310hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "bb3a18744eaa2e32229528cb2e5e0cf8",
      "name": "flash-attn",
      "sha256": "8fedc62fb006f6eb217b9118d5454482d0a52032db5e3b2534c053471fd9b8f0",
      "size": 295387567,
      "subdir": "linux-64",
      "timestamp": 1777350709319,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py310hdb63472_1.tar.bz2": {
      "build": "py310hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "68ce1de1303d3005e51d88679a80a3c4",
      "name": "flash-attn",
      "sha256": "99abd360cbc14906885c89d569353be8bf40993a19ee924b39c3ee863dc7265b",
      "size": 312605416,
      "subdir": "linux-64",
      "timestamp": 1777350997582,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hb4df0bb_1.tar.bz2": {
      "build": "py311hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "2474a04b453587024353bda92c7f3786",
      "name": "flash-attn",
      "sha256": "393977e2fd4d568c7608d4a77c79183a7905962089ce32595bd10d631937a323",
      "size": 295716113,
      "subdir": "linux-64",
      "timestamp": 1777351794412,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hdb63472_1.tar.bz2": {
      "build": "py311hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "cdadc8dd00023d08eb14d006bfc1792d",
      "name": "flash-attn",
      "sha256": "5167a7648791f728f8607afff2e575ae58d6b52e94801cbb69103fe182513c38",
      "size": 312773131,
      "subdir": "linux-64",
      "timestamp": 1777351557088,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hb4df0bb_1.tar.bz2": {
      "build": "py312hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a8f818a061cfef4be3749e1837be8406",
      "name": "flash-attn",
      "sha256": "249005d99130cb8880bcb7fe6953e7afa7694527fbb937b00313ad1ab0438350",
      "size": 295624860,
      "subdir": "linux-64",
      "timestamp": 1777351071768,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hdb63472_1.tar.bz2": {
      "build": "py312hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "4c632570641d7fae61d3045f1453e9e5",
      "name": "flash-attn",
      "sha256": "b4ff24900d9d636fdf2f8aeb05777d67e68b8e705c98755c1042877d2d78bce2",
      "size": 312803460,
      "subdir": "linux-64",
      "timestamp": 1777351391083,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hb4df0bb_1.tar.bz2": {
      "build": "py313hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a324aefd099819ccb6c94b39892d1c41",
      "name": "flash-attn",
      "sha256": "f8ff557c5a0162759ded79afa3590ef71dcd16541b9fe001331c358fcae1685e",
      "size": 295793973,
      "subdir": "linux-64",
      "timestamp": 1777351302853,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hdb63472_1.tar.bz2": {
      "build": "py313hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "d15c6e8f789ac24796e2f7767c12194d",
      "name": "flash-attn",
      "sha256": "0319ea9ff099b6208f022290c62c3b26f4f7a561ba2d1c53db78f756904e8c59",
      "size": 312641381,
      "subdir": "linux-64",
      "timestamp": 1777350886152,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h214eb8d_1.tar.bz2": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "1bbfae0f6c149165dc0d44df0cb6e535",
      "name": "flash-attn-fused-dense",
      "sha256": "e116c029ca93f07d0f36b561c07a6d12d9e41a71ea31cbc45211a1cc235915d3",
      "size": 92596,
      "subdir": "linux-64",
      "timestamp": 1777351460059,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h86323b5_1.tar.bz2": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "fe643567133fc7f2f2b2ccaf4f0f7afa",
      "name": "flash-attn-fused-dense",
      "sha256": "11f76235ed946318cd8d14552589277a3286ac45ae34d093535cf7ed18785d1a",
      "size": 94647,
      "subdir": "linux-64",
      "timestamp": 1777351120618,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h214eb8d_1.tar.bz2": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "92da7d83371782471be668f80ec8a9a4",
      "name": "flash-attn-fused-dense",
      "sha256": "749f6276fc9fda4a494776158d639240b3ab954da93ee571e30495f55c090f98",
      "size": 94189,
      "subdir": "linux-64",
      "timestamp": 1777352062064,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h86323b5_1.tar.bz2": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "553ce08607ad5a4df02b7f0e78f54815",
      "name": "flash-attn-fused-dense",
      "sha256": "7d2223135dfefe5490831ea5b6e65318f1e8e295fc700b73dd3db43faf3f5857",
      "size": 96189,
      "subdir": "linux-64",
      "timestamp": 1777352223981,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h214eb8d_1.tar.bz2": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "c07eb89594ef0ebe4be7dcadbfde3fb7",
      "name": "flash-attn-fused-dense",
      "sha256": "38c6cb55ee9acf6c5b16c497b4125b5430d9c619ba48e660ffdbf4d1f1a7f234",
      "size": 94591,
      "subdir": "linux-64",
      "timestamp": 1777351884331,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h86323b5_1.tar.bz2": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "b41fc0e40c70a7d1fd1dbc817f0ae253",
      "name": "flash-attn-fused-dense",
      "sha256": "3611147a6fe29ad998efb71f3633270dbee50605f0d2d974ce071ce9b935cff4",
      "size": 96743,
      "subdir": "linux-64",
      "timestamp": 1777351495932,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h214eb8d_1.tar.bz2": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "0d75cdcc406fe872c8cf176da46c152e",
      "name": "flash-attn-fused-dense",
      "sha256": "ff388ac282cb594873632f9115102a54d238fc9b1eeda21bb108c5f611c877b7",
      "size": 94599,
      "subdir": "linux-64",
      "timestamp": 1777351318465,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h86323b5_1.tar.bz2": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "33831e601d1d0acd2e22fdc00d80014e",
      "name": "flash-attn-fused-dense",
      "sha256": "2d121c81e2e57c8be8903c719394ef8e9a0b79ef5754162ccd21ace19469825e",
      "size": 96674,
      "subdir": "linux-64",
      "timestamp": 1777351733148,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h214eb8d_1.tar.bz2": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "fae0cdc65cd20db819c76e3082fd2d9e",
      "name": "flash-attn-layer-norm",
      "sha256": "43a1c8b3d0d369b81c74be8d9717bcab7933fec702a9a275dee4daf57677c55b",
      "size": 353007055,
      "subdir": "linux-64",
      "timestamp": 1777351512268,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h86323b5_1.tar.bz2": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "f092eef0e191251d151aa49475822be8",
      "name": "flash-attn-layer-norm",
      "sha256": "f08880521e8ec6fe8bc06bd125fc81a4b56a199b5ea11db77a355d753edcf3b5",
      "size": 438773146,
      "subdir": "linux-64",
      "timestamp": 1777351171596,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h214eb8d_1.tar.bz2": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "413cd0fba46274fdb77550d203dbf14c",
      "name": "flash-attn-layer-norm",
      "sha256": "b31489c50d59b7bd05a8879c028cf8e124ef0ecdb65e165d1bd7896cf1b1cbd9",
      "size": 353015440,
      "subdir": "linux-64",
      "timestamp": 1777352118767,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h86323b5_1.tar.bz2": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "dd308a05bcb3d2a74294ff276f4995fa",
      "name": "flash-attn-layer-norm",
      "sha256": "59b8472f50f400a5590b3770e129c43df41740058de15cc5aff037297445255c",
      "size": 438768338,
      "subdir": "linux-64",
      "timestamp": 1777352275483,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h214eb8d_1.tar.bz2": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a01d9ea9ec6684c0ea6bfcc28f64fbf6",
      "name": "flash-attn-layer-norm",
      "sha256": "66a2a869d02714a8e87d3086635d35cd649afaa71628d92de291b9d5322f24f1",
      "size": 353009697,
      "subdir": "linux-64",
      "timestamp": 1777351939102,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h86323b5_1.tar.bz2": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "4e66725eb65ed9e90ded31512aaf15f2",
      "name": "flash-attn-layer-norm",
      "sha256": "005420f1a9a04d3787d496f776e407de1b2f3448bb9ea9cfc97c84d67d5ed63e",
      "size": 438777460,
      "subdir": "linux-64",
      "timestamp": 1777351548138,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h214eb8d_1.tar.bz2": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "43359fbe20cfb80e0478d39541a3eeef",
      "name": "flash-attn-layer-norm",
      "sha256": "50937064bdfbc0ab148e95c68b8ec67e017e4cafa37c8064c283fa7051cd756a",
      "size": 353006178,
      "subdir": "linux-64",
      "timestamp": 1777351374255,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h86323b5_1.tar.bz2": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "7d8e746f38ddb157b5445d7e857a4577",
      "name": "flash-attn-layer-norm",
      "sha256": "703659856e15b4ad0f549c439d9f160f57537d690d7392afe2511d0621f9e37d",
      "size": 438764337,
      "subdir": "linux-64",
      "timestamp": 1777351785145,
      "version": "2.8.3"
    }
  },
  "packages.conda": {
    "flash-attn-2.8.3-py310hb4df0bb_1.conda": {
      "build": "py310hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "82155d3354b0af280f8a360cbfa20367",
      "name": "flash-attn",
      "sha256": "f580bf3968285d013b9507002d43cbe0598dc4ea9410e70136741478c6d27430",
      "size": 196555605,
      "subdir": "linux-64",
      "timestamp": 1777350709319,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py310hdb63472_1.conda": {
      "build": "py310hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "039f1d8bc639bb2210cfe4c4109d7018",
      "name": "flash-attn",
      "sha256": "0445cd5e8cd31f18c9c9041d49e2d805dd2b11a0c22bcbf76f28b936fd70df3a",
      "size": 196173391,
      "subdir": "linux-64",
      "timestamp": 1777350997582,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hb4df0bb_1.conda": {
      "build": "py311hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "62b18fb37e38bad117306679dd5f5156",
      "name": "flash-attn",
      "sha256": "41159e2a59af943f02dab30196cbaa487224a20753720c783136669052ed087a",
      "size": 196759831,
      "subdir": "linux-64",
      "timestamp": 1777351794412,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hdb63472_1.conda": {
      "build": "py311hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "3eec1c4bdb63e22027c7c1c567c6cd1b",
      "name": "flash-attn",
      "sha256": "91c80e865be74c07d39cdea9227e86bc13ebd665ba84ac62b571bc6e32ac724d",
      "size": 196379826,
      "subdir": "linux-64",
      "timestamp": 1777351557088,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hb4df0bb_1.conda": {
      "build": "py312hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "643a61738ace217a50280d7df385e1e2",
      "name": "flash-attn",
      "sha256": "f3ee4c06162d28c26dfa4a850081c87ea7aa55aad54a6a46dadab6fa51bdd997",
      "size": 196744328,
      "subdir": "linux-64",
      "timestamp": 1777351071768,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hdb63472_1.conda": {
      "build": "py312hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5fc6102fd8c120db6707c22a1e969024",
      "name": "flash-attn",
      "sha256": "7932c2ed25bbef0df69e4e0c8a1857ebfe7fdbfd238796954080ac1734b9f213",
      "size": 196350259,
      "subdir": "linux-64",
      "timestamp": 1777351391083,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hb4df0bb_1.conda": {
      "build": "py313hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "688f06e527ee6de01733c1d91c8d0e0b",
      "name": "flash-attn",
      "sha256": "0d1ea52d1056800983877a1ba1fbfdfeff2f3f56afd1cb238f9998f59db12a49",
      "size": 196747547,
      "subdir": "linux-64",
      "timestamp": 1777351302853,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hdb63472_1.conda": {
      "build": "py313hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "43c68b35ce02075fb6d61b0cd049c836",
      "name": "flash-attn",
      "sha256": "1694e7edc8cfee5c197644a76dd91feaaa309fad81b7fc3614bddfe4fc35902f",
      "size": 196367661,
      "subdir": "linux-64",
      "timestamp": 1777350886152,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h214eb8d_1.conda": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "b4a0eb4aade12cb805b1dab8e415ae68",
      "name": "flash-attn-fused-dense",
      "sha256": "2e1633f755f61cc427028b416cb08ace6e24d56f48eb6521789cf524b2e0a187",
      "size": 84918,
      "subdir": "linux-64",
      "timestamp": 1777351460059,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h86323b5_1.conda": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "4acd8a028d272460f73f3e679731ab60",
      "name": "flash-attn-fused-dense",
      "sha256": "770b55c361afe896003550742c2d965a88230350795bb25d90d1632a6cc9e342",
      "size": 85958,
      "subdir": "linux-64",
      "timestamp": 1777351120618,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h214eb8d_1.conda": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "d661d60ce9e2e1ab38376d958ec030e2",
      "name": "flash-attn-fused-dense",
      "sha256": "9e90b28ddc9d434e71e85c88707e2b0317c8649614e75f844465cc4e50be5237",
      "size": 86316,
      "subdir": "linux-64",
      "timestamp": 1777352062064,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h86323b5_1.conda": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "e61d57fe207736464bc6671651bdcb4f",
      "name": "flash-attn-fused-dense",
      "sha256": "1be41d101d1e163f8311dcb7095e462710bef6fcb6f91972522e1df82ad39e9b",
      "size": 87488,
      "subdir": "linux-64",
      "timestamp": 1777352223981,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h214eb8d_1.conda": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "c9ea6ea60b0f253c04e8d1060ee0ae4c",
      "name": "flash-attn-fused-dense",
      "sha256": "eaa131928d4d7d4e70f98807235407eb8db8d270414fcf4dcc4c2346b762775b",
      "size": 86783,
      "subdir": "linux-64",
      "timestamp": 1777351884331,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h86323b5_1.conda": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "4d9f0dedf824c95de35fb38b8fa178cd",
      "name": "flash-attn-fused-dense",
      "sha256": "c95a1033cb645b51349227dbc85e1f2bc5fbe314bd062770ca723048d642037b",
      "size": 87717,
      "subdir": "linux-64",
      "timestamp": 1777351495932,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h214eb8d_1.conda": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "63276a1f154b0b43d088f709ee6ed701",
      "name": "flash-attn-fused-dense",
      "sha256": "3ce2a8bc92cb243cdaa9f9f064e7ef9567be14352f9b3d8fb19dd5e3142ee9e3",
      "size": 86398,
      "subdir": "linux-64",
      "timestamp": 1777351318465,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h86323b5_1.conda": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "946f67a56117dfe9b3af16322cb782b9",
      "name": "flash-attn-fused-dense",
      "sha256": "4b7e29c1cd06cc2b763e6d81bb60ca47b066cd3a8e4694500cc003ea9ce79c2a",
      "size": 87702,
      "subdir": "linux-64",
      "timestamp": 1777351733148,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h214eb8d_1.conda": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5e84e12b37693f239d42b38ac311ff5c",
      "name": "flash-attn-layer-norm",
      "sha256": "a16f1143168ccb8249d9f8f8f457db9fa0618c4b76331191ce3bb1ddc143a62c",
      "size": 210016746,
      "subdir": "linux-64",
      "timestamp": 1777351512268,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h86323b5_1.conda": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "1146215968df45a1b383c5a41ac2c9af",
      "name": "flash-attn-layer-norm",
      "sha256": "d871dd9e9585590620103c2cedf0e686b5c69391fcc8680b3eb0435650adfaac",
      "size": 334810482,
      "subdir": "linux-64",
      "timestamp": 1777351171596,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h214eb8d_1.conda": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "7272014e8e1195f3e9e78dd83e72c3e6",
      "name": "flash-attn-layer-norm",
      "sha256": "e338c3cbc932c63645d08eb51da00a268a0c1b785ef8cf8d170115de7cf4c50b",
      "size": 210346008,
      "subdir": "linux-64",
      "timestamp": 1777352118767,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h86323b5_1.conda": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "8f414b6086c944a587a457de0f9e9025",
      "name": "flash-attn-layer-norm",
      "sha256": "61c88a6bddf454dffdb662ca725818a588d2106e567fde15c23f60d7389d9b28",
      "size": 334831417,
      "subdir": "linux-64",
      "timestamp": 1777352275483,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h214eb8d_1.conda": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "f35ceb16f401c79a3a50652473cf86e8",
      "name": "flash-attn-layer-norm",
      "sha256": "e6ed12cd896e12d25c220602d82eac78499738a50501638fe9b6b0eaa1952b7e",
      "size": 210221005,
      "subdir": "linux-64",
      "timestamp": 1777351939102,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h86323b5_1.conda": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "05be57a1ce1949fcb1c0e74de0f63946",
      "name": "flash-attn-layer-norm",
      "sha256": "5ae13f78506ee6d679b0c748211717b1c250e1280b0c9a0cbc607fbb60eeb1a3",
      "size": 334863931,
      "subdir": "linux-64",
      "timestamp": 1777351548138,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h214eb8d_1.conda": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "e644a3dbbdf54feed1f6a90425f73e0a",
      "name": "flash-attn-layer-norm",
      "sha256": "c11e25ad8e1f7e70fbd4eb95aaa5fbbe724dcb555633fc35c5925aa1460f48d0",
      "size": 210220021,
      "subdir": "linux-64",
      "timestamp": 1777351374255,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h86323b5_1.conda": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "39c35ef26347988a079d6e48fe7b85a0",
      "name": "flash-attn-layer-norm",
      "sha256": "1c98b6687ddae9f9acf3e15c8ea2940e726dce6b63b4e9578a51dd27a8b4bf15",
      "size": 334713378,
      "subdir": "linux-64",
      "timestamp": 1777351785145,
      "version": "2.8.3"
    }
  },
  "removed": [],
  "repodata_version": 1
}