{
    "componentChunkName": "component---src-templates-mdx-post-js",
    "path": "/blog/compute_aware_arch_search",
    "result": {"data":{"mdx":{"body":"const _excluded = [\"components\"];\nfunction _extends() { return _extends = Object.assign ? Object.assign.bind() : function (n) { for (var e = 1; e < arguments.length; e++) { var t = arguments[e]; for (var r in t) ({}).hasOwnProperty.call(t, r) && (n[r] = t[r]); } return n; }, _extends.apply(null, arguments); }\nfunction _objectWithoutProperties(e, t) { if (null == e) return {}; var o, r, i = _objectWithoutPropertiesLoose(e, t); if (Object.getOwnPropertySymbols) { var n = Object.getOwnPropertySymbols(e); for (r = 0; r < n.length; r++) o = n[r], -1 === t.indexOf(o) && {}.propertyIsEnumerable.call(e, o) && (i[o] = e[o]); } return i; }\nfunction _objectWithoutPropertiesLoose(r, e) { if (null == r) return {}; var t = {}; for (var n in r) if ({}.hasOwnProperty.call(r, n)) { if (-1 !== e.indexOf(n)) continue; t[n] = r[n]; } return t; }\n/* @jsxRuntime classic */\n/* @jsx mdx */\n\nconst _frontmatter = {\n  \"title\": \"Compute-Aware Hybrid Attention Architecture Search\",\n  \"description\": \"Searching compute-aware hybrid attention architectures for large language models\",\n  \"date\": \"2025-12-10\",\n  \"slug\": \"/blog/compute_aware_arch_search\",\n  \"tags\": [\"llm\", \"architecture-search\", \"attention\"]\n};\nconst layoutProps = {\n  _frontmatter\n};\nconst MDXLayout = \"wrapper\";\nreturn function MDXContent(_ref) {\n  let {\n      components\n    } = _ref,\n    props = _objectWithoutProperties(_ref, _excluded);\n  return mdx(MDXLayout, _extends({}, layoutProps, props, {\n    components: components,\n    mdxType: \"MDXLayout\"\n  }), mdx(\"iframe\", {\n    src: \"https://sanjithudupa.github.io/compute-aware-arch-search/\",\n    width: \"100%\",\n    height: \"1200px\",\n    style: {\n      border: 'none',\n      marginTop: '2rem',\n      borderRadius: '8px'\n    },\n    title: \"Compute-Aware Hybrid Attention Architecture Search\",\n    allow: \"fullscreen\"\n  }), mdx(\"p\", {\n    style: {\n      marginTop: '1rem',\n      textAlign: 'center'\n    }\n  }, mdx(\"a\", {\n    href: \"https://sanjithudupa.github.io/compute-aware-arch-search/\",\n    target: \"_blank\",\n    rel: \"noopener noreferrer\"\n  }, \"Open in new tab\")), mdx(\"h1\", null, `Abstract`), mdx(\"p\", null, `This project is a deep dive into the compute-aware hybrid attention architecture search for large language models. We will be using the Llama 3.1 8B model and the Qwen 2.5 7B model to search for the best architecture.`), mdx(\"p\", null, `We will be using the compute-aware hybrid attention architecture search to search for the best architecture for the Llama 3.1 8B model and the Qwen 2.5 7B model.`), mdx(\"p\", null, `We will be using the compute-aware hybrid attention architecture search to search for the best architecture for the Llama 3.1 8B model and the Qwen 2.5 7B model.`), mdx(\"p\", null, `We will be using the compute-aware hybrid attention architecture search to search for the best architecture for the Llama 3.1 8B model and the Qwen 2.5 7B model.`));\n}\n;\nMDXContent.isMDXComponent = true;","frontmatter":{"title":"Compute-Aware Hybrid Attention Architecture Search","description":"Searching compute-aware hybrid attention architectures for large language models","date":"2025-12-10","slug":"/blog/compute_aware_arch_search","tags":["llm","architecture-search","attention"]}}},"pageContext":{"slug":"/blog/compute_aware_arch_search"}},
    "staticQueryHashes": ["1323149524","1761877533","1994492073","2885896977"]}