{
  "info": {
    "subdir": "linux-64"
  },
  "packages": {
    "flash-attn-2.8.3-py310hb4df0bb_1.tar.bz2": {
      "build": "py310hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5a7db802cbe842c4223b8c45bee0e403",
      "name": "flash-attn",
      "sha256": "724659118b294a51f328d6f0e77885a097d8e59f91176a034a02b3d89ccf01f6",
      "size": 295489268,
      "subdir": "linux-64",
      "timestamp": 1777360527024,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py310hdb63472_1.tar.bz2": {
      "build": "py310hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a66e01990638c4b0c3b5d059179c34a2",
      "name": "flash-attn",
      "sha256": "dff21b63a6bae844d23492696bb72f6be724110c2cdfde2349fd65c10a194214",
      "size": 312554351,
      "subdir": "linux-64",
      "timestamp": 1777361422617,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hb4df0bb_1.tar.bz2": {
      "build": "py311hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "1f6c98bfc93bda8d93198267a2fdba38",
      "name": "flash-attn",
      "sha256": "e1aed8d4297b962137b3997b8b14d380e8f8dc5f9348b6d6bca74e2f45e02899",
      "size": 295749351,
      "subdir": "linux-64",
      "timestamp": 1777364632157,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hdb63472_1.tar.bz2": {
      "build": "py311hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "2b8fa497c78b33aa76ffef4ae5708698",
      "name": "flash-attn",
      "sha256": "465faf49ab06741851a472f511d734ea4a9aa3044a8bf2d073d74d51ba869fe0",
      "size": 312737553,
      "subdir": "linux-64",
      "timestamp": 1777360851077,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hb4df0bb_1.tar.bz2": {
      "build": "py312hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "59f8d2bd8c8e067a1b596e41a5803ae9",
      "name": "flash-attn",
      "sha256": "67adcece58a3234d60c1b0905d105dbf6208e7664a19dc206972be83a53a39d3",
      "size": 295728270,
      "subdir": "linux-64",
      "timestamp": 1777360615905,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hdb63472_1.tar.bz2": {
      "build": "py312hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "7952ff2c5b0f0401076238af2fe69542",
      "name": "flash-attn",
      "sha256": "d16b46fd6a4702f109816fc58fff52d05547d0c4ef1ff5227e308cc012a54adf",
      "size": 312657411,
      "subdir": "linux-64",
      "timestamp": 1777361393297,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hb4df0bb_1.tar.bz2": {
      "build": "py313hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "1e9fc1d92b95fb8bbc8c1a2c3e29ceb7",
      "name": "flash-attn",
      "sha256": "456465359b04cc4102a3929a3a6f56c8938368e5499675b67fc97e11b8bb500e",
      "size": 295755073,
      "subdir": "linux-64",
      "timestamp": 1777361603008,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hdb63472_1.tar.bz2": {
      "build": "py313hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "97c94e46108dd6f872ad5845ccf67d73",
      "name": "flash-attn",
      "sha256": "c1d25c9b03b5c8069919870f02068c70fc7586e07a541ad7e968218bae919350",
      "size": 312324171,
      "subdir": "linux-64",
      "timestamp": 1777360917100,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h214eb8d_1.tar.bz2": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "b5bc77651ca5d2c170baa81def9828e1",
      "name": "flash-attn-fused-dense",
      "sha256": "03e857833efa25658b325407365e347417dca5de32ac8cd4d9a88d9f3ea739b6",
      "size": 92593,
      "subdir": "linux-64",
      "timestamp": 1777361914446,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h86323b5_1.tar.bz2": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "b97f93b0bbe020bd299ae47d696b130a",
      "name": "flash-attn-fused-dense",
      "sha256": "330696a44c713a8a05f5396c6cc7588eb9982f49bfe6163cbe7cea320f14d51e",
      "size": 94623,
      "subdir": "linux-64",
      "timestamp": 1777360952922,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h214eb8d_1.tar.bz2": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "d28f7d1b500871ed78867c9f995b6148",
      "name": "flash-attn-fused-dense",
      "sha256": "e5d3885f2d7aa57deea93f45e676e157d446bb3ef436e3815ef29a0e7b2a0a17",
      "size": 94224,
      "subdir": "linux-64",
      "timestamp": 1777361351905,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h86323b5_1.tar.bz2": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "220d89a0f9b6d3d35be11596052433e1",
      "name": "flash-attn-fused-dense",
      "sha256": "51f391eb66bd47975d9d4bf6ef550364ec14aa1a583bc7f762b4dfde73a9c2fb",
      "size": 96240,
      "subdir": "linux-64",
      "timestamp": 1777365134942,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h214eb8d_1.tar.bz2": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "6784aeb16c64bf383fb267a872cee75b",
      "name": "flash-attn-fused-dense",
      "sha256": "4a4fab49cb509ee84750b35a5d013119f85ff992b5bee6c0ff0999b3fdf02819",
      "size": 94614,
      "subdir": "linux-64",
      "timestamp": 1777361885434,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h86323b5_1.tar.bz2": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "33fafaf281f1f77d3646f08bff3d8bac",
      "name": "flash-attn-fused-dense",
      "sha256": "156c02ec642f6c3b133dac14199ad5691c3eea79f107feaebbc76c8c922d39a7",
      "size": 96706,
      "subdir": "linux-64",
      "timestamp": 1777361021801,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h214eb8d_1.tar.bz2": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a3fd3de20243ffd2ba79ceae873de18e",
      "name": "flash-attn-fused-dense",
      "sha256": "49423913939adccdeaf9d682bc3840435468165e4cc6aea8f1ab2579707ff4a9",
      "size": 94558,
      "subdir": "linux-64",
      "timestamp": 1777361349365,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h86323b5_1.tar.bz2": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "238cbc6210fede5202688d645978a22c",
      "name": "flash-attn-fused-dense",
      "sha256": "e67b04aa0466aa118d65732d4c1d96fb247c92d6723a4a0f670e3564223845ec",
      "size": 96767,
      "subdir": "linux-64",
      "timestamp": 1777362072403,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h214eb8d_1.tar.bz2": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "78e469873302d687dcc3c1f6ae896e64",
      "name": "flash-attn-layer-norm",
      "sha256": "d29d44352315e09d32c12eeea683005714e9e313f604f3d961eba4a514dd5f45",
      "size": 353016631,
      "subdir": "linux-64",
      "timestamp": 1777361969086,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h86323b5_1.tar.bz2": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "384573f8df1b2d13382c62c0d35d5646",
      "name": "flash-attn-layer-norm",
      "sha256": "5b8bfd937d19421b71c557b46ead3e980221164d79c3bd048c65e2829cb249ba",
      "size": 438771336,
      "subdir": "linux-64",
      "timestamp": 1777361005995,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h214eb8d_1.tar.bz2": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5553fde942a44d0ab1b10b61295f843d",
      "name": "flash-attn-layer-norm",
      "sha256": "46767c5d760265076cf2b23a04d8d5061d1e4777cded305519ae7fc040881695",
      "size": 353015954,
      "subdir": "linux-64",
      "timestamp": 1777361404493,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h86323b5_1.tar.bz2": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "27a5161a27380041b2abf53e3eb0d712",
      "name": "flash-attn-layer-norm",
      "sha256": "28ccf0bcd994cdfd158e9d3dcdb440cc567a060b361471071017769404eb2075",
      "size": 438776487,
      "subdir": "linux-64",
      "timestamp": 1777365188868,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h214eb8d_1.tar.bz2": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "c86122ab1c645214af8080fb9eb51885",
      "name": "flash-attn-layer-norm",
      "sha256": "b248837318b98264a300ccfa2f03eb98867c2167c29b0e22f8357a43c5878e72",
      "size": 353009892,
      "subdir": "linux-64",
      "timestamp": 1777361938950,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h86323b5_1.tar.bz2": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "068709465bc15b15bdf35b674a6dd9a6",
      "name": "flash-attn-layer-norm",
      "sha256": "886e65f58d101e3b5e14a95d919976a45f784e2a2a8902e9b0225266eb2d2c7a",
      "size": 438774419,
      "subdir": "linux-64",
      "timestamp": 1777361071387,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h214eb8d_1.tar.bz2": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5dad2ced7cdc7001f0881462de063973",
      "name": "flash-attn-layer-norm",
      "sha256": "40ecdea4489a13bd331dddafc9e0708be2ac5eaa514bebab788ea6f9a1db1f59",
      "size": 353010488,
      "subdir": "linux-64",
      "timestamp": 1777361401993,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h86323b5_1.tar.bz2": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "976699520e8f850349f37ab16ae95041",
      "name": "flash-attn-layer-norm",
      "sha256": "b397a988366bccee9502d4502f44e177c40248f069c050dcb8fff2ec9929311d",
      "size": 438779422,
      "subdir": "linux-64",
      "timestamp": 1777362124566,
      "version": "2.8.3"
    }
  },
  "packages.conda": {
    "flash-attn-2.8.3-py310hb4df0bb_1.conda": {
      "build": "py310hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "d4e685b92d761ad5f2bb8927480b2b38",
      "name": "flash-attn",
      "sha256": "a79088133d88a7712d71d5e4612f8defbfae7fe2a60106b32d452b304237746e",
      "size": 196556311,
      "subdir": "linux-64",
      "timestamp": 1777360527024,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py310hdb63472_1.conda": {
      "build": "py310hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "a235e312dc2c983e30daba5abf580ea2",
      "name": "flash-attn",
      "sha256": "fa9b8b567ca2d3a7e11bf95e46c7c5e56cf1c7fdd93cfb48f60b80e2df4337f7",
      "size": 196159264,
      "subdir": "linux-64",
      "timestamp": 1777361422617,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hb4df0bb_1.conda": {
      "build": "py311hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "d155cb7d2863c762929b565d7cac8f2e",
      "name": "flash-attn",
      "sha256": "5fb2a9a4bb0c53c63aa6d454ab28d4bfdf36c764d945900ab2c338fc5c9b315a",
      "size": 196736222,
      "subdir": "linux-64",
      "timestamp": 1777364632157,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py311hdb63472_1.conda": {
      "build": "py311hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "2002162a3a5373eef06e5929684a5f27",
      "name": "flash-attn",
      "sha256": "f706dc91c1c32e06f918d8a5d31fef809e2b688528a39c5a62fbfb8864bc7d77",
      "size": 196380045,
      "subdir": "linux-64",
      "timestamp": 1777360851077,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hb4df0bb_1.conda": {
      "build": "py312hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "f204794e3aa9210f34a503e357ff58cc",
      "name": "flash-attn",
      "sha256": "414d36e37c3aa97895e230a8687903383e81d30cc5c8af0c9a6562a198268daa",
      "size": 196742498,
      "subdir": "linux-64",
      "timestamp": 1777360615905,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py312hdb63472_1.conda": {
      "build": "py312hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "6ff8cc61d16e047b689a800616d29890",
      "name": "flash-attn",
      "sha256": "4f18cd32ce2daa865173deef3c58fd4dc06ec873a5433804344f78076580cb0b",
      "size": 196329246,
      "subdir": "linux-64",
      "timestamp": 1777361393297,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hb4df0bb_1.conda": {
      "build": "py313hb4df0bb_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "c4e47d93c0ca82e224acaaa54a4e966f",
      "name": "flash-attn",
      "sha256": "7b083fbed150c822dd0ddf98e32f01e671a11c8cb2fec84c97e28075844a8e2e",
      "size": 196763134,
      "subdir": "linux-64",
      "timestamp": 1777361603008,
      "version": "2.8.3"
    },
    "flash-attn-2.8.3-py313hdb63472_1.conda": {
      "build": "py313hdb63472_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "einops",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch * *cuda*",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "1a277440431b2ee345699d2fe277158c",
      "name": "flash-attn",
      "sha256": "ea03257232e3efe296259b600ace53cbf17f64e830a071a35c6ef7f294b70124",
      "size": 196360807,
      "subdir": "linux-64",
      "timestamp": 1777360917100,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h214eb8d_1.conda": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "4a24ea8e53190c046671c0577cde39db",
      "name": "flash-attn-fused-dense",
      "sha256": "2a517ec1fdbe48d2ee050bfa99b4dba8fd974b65f6306bfdf1b68b29c432bf13",
      "size": 84967,
      "subdir": "linux-64",
      "timestamp": 1777361914446,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py310h86323b5_1.conda": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "9d5e6dbd10667dfc50d02f83c86cb473",
      "name": "flash-attn-fused-dense",
      "sha256": "f57b66e3302ceb3074a299b41bcd1b3f7a37c6968b1cb3e711066c6f97577aa0",
      "size": 85973,
      "subdir": "linux-64",
      "timestamp": 1777360952922,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h214eb8d_1.conda": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "6ed31669149b61209dac495b360074b8",
      "name": "flash-attn-fused-dense",
      "sha256": "c87b28118a541ec4c3da5c05ce83bb819b1137ad8274762bca8ac009a680089e",
      "size": 86049,
      "subdir": "linux-64",
      "timestamp": 1777361351905,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py311h86323b5_1.conda": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "6e7136b5c9ba1ee8395b5db8804dd044",
      "name": "flash-attn-fused-dense",
      "sha256": "01a38a6726aeab21fdb0c3b0d46dc608261ee1a57caffd14028382591b9ed85b",
      "size": 87242,
      "subdir": "linux-64",
      "timestamp": 1777365134942,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h214eb8d_1.conda": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "0e594338710a9b03975a41da00a55809",
      "name": "flash-attn-fused-dense",
      "sha256": "20547054537cad04a6dc3ada684821ab3d558cee74f3d2993aed34a8d8146b13",
      "size": 86727,
      "subdir": "linux-64",
      "timestamp": 1777361885434,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py312h86323b5_1.conda": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "996cd367b4f234e1952ec44caf4a3460",
      "name": "flash-attn-fused-dense",
      "sha256": "73e2577ec0be7ac16f83e7093c88725a03a69588d394c4c6c5f8b8b95ebf18e9",
      "size": 87622,
      "subdir": "linux-64",
      "timestamp": 1777361021801,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h214eb8d_1.conda": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libcublas >=12.8.5.5,<13.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "c3c0e08af3ac1724d18da1218feeb97f",
      "name": "flash-attn-fused-dense",
      "sha256": "9d1e4f2ac81a50da51a9797e9169c4253f8020550b0139844c78ebf746a99c99",
      "size": 86447,
      "subdir": "linux-64",
      "timestamp": 1777361349365,
      "version": "2.8.3"
    },
    "flash-attn-fused-dense-2.8.3-py313h86323b5_1.conda": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libcublas >=13.1.1.3,<14.0a0",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "97ef495cc42cd983347ead48fce83860",
      "name": "flash-attn-fused-dense",
      "sha256": "b5a47d0ede2ac5e13d7b6043b7c2161cdc97e42ca5677878010bb8878f8711ef",
      "size": 87595,
      "subdir": "linux-64",
      "timestamp": 1777362072403,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h214eb8d_1.conda": {
      "build": "py310h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py310hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "b86c0d786a1f5abb2b83c1035688364d",
      "name": "flash-attn-layer-norm",
      "sha256": "1876a1a948cd44f626c7bae285affde73e7daa9d4691ab0c11c0514e0e8edca9",
      "size": 210461390,
      "subdir": "linux-64",
      "timestamp": 1777361969086,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py310h86323b5_1.conda": {
      "build": "py310h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py310hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.10,<3.11.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "19b4138b9aabfb40fe3926bc60e013a4",
      "name": "flash-attn-layer-norm",
      "sha256": "38cbeac7f5da9c545fef58ff6ac00023eb51e097a200ee06beb5ab5e3bc28714",
      "size": 334760771,
      "subdir": "linux-64",
      "timestamp": 1777361005995,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h214eb8d_1.conda": {
      "build": "py311h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py311hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "5510d06eceabab56f97d57b9765fdde7",
      "name": "flash-attn-layer-norm",
      "sha256": "113d776ab4188cf89824ed2aac46b12b15a112fd24bda0ad8a2246c030aaa145",
      "size": 210488348,
      "subdir": "linux-64",
      "timestamp": 1777361404493,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py311h86323b5_1.conda": {
      "build": "py311h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py311hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.11,<3.12.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "72114a3b7a8372d40125363bde46f2ad",
      "name": "flash-attn-layer-norm",
      "sha256": "28d6e3963e35ccdfd71bbf9f0e61aae4792193436779e8dd3516468a06003ac7",
      "size": 334782727,
      "subdir": "linux-64",
      "timestamp": 1777365188868,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h214eb8d_1.conda": {
      "build": "py312h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py312hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "3fba58b24aa29e08e4c7b4fcd804dcd9",
      "name": "flash-attn-layer-norm",
      "sha256": "96ec089fa28ddc745fb7296fb81a12d85a866319f3a6ebd9254148ea23cae234",
      "size": 210447130,
      "subdir": "linux-64",
      "timestamp": 1777361938950,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py312h86323b5_1.conda": {
      "build": "py312h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py312hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.12,<3.13.0a0",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "2f8f24c05e12d84cf5ccc4c5c08c9785",
      "name": "flash-attn-layer-norm",
      "sha256": "9d4d9882ef083a05ab9e96f41dda040fed6852a9ad8e397dac76fe988f296d6d",
      "size": 334857946,
      "subdir": "linux-64",
      "timestamp": 1777361071387,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h214eb8d_1.conda": {
      "build": "py313h214eb8d_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=12.8.90,<13.0a0",
        "cuda-version >=12.8,<13",
        "flash-attn 2.8.3 py313hdb63472_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "e0ab57b48eefbbe20445f79f8ca7925a",
      "name": "flash-attn-layer-norm",
      "sha256": "0590d50ee1d53710b2c93e3174a2497bc160cca77ef9dfc3d7e92e3035f569f6",
      "size": 210282457,
      "subdir": "linux-64",
      "timestamp": 1777361401993,
      "version": "2.8.3"
    },
    "flash-attn-layer-norm-2.8.3-py313h86323b5_1.conda": {
      "build": "py313h86323b5_1",
      "build_number": 1,
      "depends": [
        "__glibc >=2.28,<3.0.a0",
        "cuda-cudart >=13.0.96,<14.0a0",
        "cuda-version >=13.0,<14",
        "flash-attn 2.8.3 py313hb4df0bb_1",
        "libgcc >=14",
        "libstdcxx >=14",
        "libtorch >=2.10.0,<2.11.0a0",
        "python >=3.13,<3.14.0a0",
        "python_abi 3.13.* *_cp313",
        "pytorch >=2.10.0,<2.11.0a0"
      ],
      "license": "BSD-3-Clause",
      "license_family": "BSD",
      "md5": "7a820e8dc0dd0c266b11a89cebbdf094",
      "name": "flash-attn-layer-norm",
      "sha256": "79772ce6a299b81b07937cfccfb4d5333bf40b4e9ba1c094bdbbf5cf35719128",
      "size": 334759713,
      "subdir": "linux-64",
      "timestamp": 1777362124566,
      "version": "2.8.3"
    }
  },
  "removed": [],
  "repodata_version": 1
}