{"cache_data_version":"1.0.6","atoms":{"app-ai/llama-cpp-0.0.7388":{"atom":"app-ai/llama-cpp-0.0.7388","category":"app-ai","package":"llama-cpp","revision":"0","catpkg":"app-ai/llama-cpp","eclasses":[["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["xdg-utils","14d00d009167652b1fa363e55effe213"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake","x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"relations_by_kind":{"BDEPEND":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake"],"DEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"RDEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"]},"metadata":{"BDEPEND":"vulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install prepare test","DEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","DESCRIPTION":"LLM inference in C/C++","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://github.com/ggml-org/llama.cpp","INHERITED":"toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake","IUSE":"static blas cuda vulkan","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","REQUIRED_USE":"","RESTRICT":"","SLOT":"0","SRC_URI":"https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7388 -\u003e llama-cpp-0.0.7388-4ed2bae.tar.gz"},"metadata_out":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\ncuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\n0\nhttps://api.github.com/repos/ggml-org/llama.cpp/tarball/b7388 -\u003e llama-cpp-0.0.7388-4ed2bae.tar.gz\n\nhttps://github.com/ggml-org/llama.cpp\nMIT\nLLM inference in C/C++\n*\ntoolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake\nstatic blas cuda vulkan\n\n\nvulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake\n7\n\ncompile configure install prepare test\n\n\n\n\n\n","manifest_md5":"130d79604020aa3f4860c95d060d6252","md5":"bf69a4f8e46de05b23f9dd80ae5303a2","files":[{"src_uri":["https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7388"],"size":"28132487","hashes":{"blake2b":"02131e66cac092c5f0c23a86b100fab2a41b1a4c357fccedaa5c687200b6ad15a9644e4a95e7129f242f1474a82987dd3ab6197f2b82e8f2849d004edf78ee03","sha512":"84efa2b4441b1731647ab8b8d9e94626e9a348a755674ae5f3553efd564fd29ed3b7c0165093823e3147cb1d47abd7300871456882db8970eda43dd9305ae6d2"},"name":"llama-cpp-0.0.7388-4ed2bae.tar.gz"}]},"app-ai/llama-cpp-0.0.7492":{"atom":"app-ai/llama-cpp-0.0.7492","category":"app-ai","package":"llama-cpp","revision":"0","catpkg":"app-ai/llama-cpp","eclasses":[["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["xdg-utils","14d00d009167652b1fa363e55effe213"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake","x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"relations_by_kind":{"BDEPEND":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake"],"DEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"RDEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"]},"metadata":{"BDEPEND":"vulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install prepare test","DEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","DESCRIPTION":"LLM inference in C/C++","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://github.com/ggml-org/llama.cpp","INHERITED":"toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake","IUSE":"static blas cuda vulkan","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","REQUIRED_USE":"","RESTRICT":"","SLOT":"0","SRC_URI":"https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7492 -\u003e llama-cpp-0.0.7492-ddcb75d.tar.gz"},"metadata_out":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\ncuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\n0\nhttps://api.github.com/repos/ggml-org/llama.cpp/tarball/b7492 -\u003e llama-cpp-0.0.7492-ddcb75d.tar.gz\n\nhttps://github.com/ggml-org/llama.cpp\nMIT\nLLM inference in C/C++\n*\ntoolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake\nstatic blas cuda vulkan\n\n\nvulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake\n7\n\ncompile configure install prepare test\n\n\n\n\n\n","manifest_md5":"130d79604020aa3f4860c95d060d6252","md5":"012cabae5e173400679ca42a3631a137","files":[{"src_uri":["https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7492"],"size":"28579052","hashes":{"blake2b":"cf967832300c53e4c9e33866be436b3589327dd43aebc30b8c8c167ca63d0ec52bedf649d03643c3a1c57591d68c1bc94931e1584c03160d1fe13ed5825fd119","sha512":"072309c18d3efa8d4f983b7911a6dab61659e93771675ce0a4de97baaf0276f857c2e7e3dace4931e2790aceb6b96382fdb44b8f12e108635f7647247ca9ce12"},"name":"llama-cpp-0.0.7492-ddcb75d.tar.gz"}]},"app-ai/llama-cpp-0.0.7556":{"atom":"app-ai/llama-cpp-0.0.7556","category":"app-ai","package":"llama-cpp","revision":"0","catpkg":"app-ai/llama-cpp","eclasses":[["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["xdg-utils","14d00d009167652b1fa363e55effe213"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake","x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"relations_by_kind":{"BDEPEND":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake"],"DEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"RDEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"]},"metadata":{"BDEPEND":"vulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install prepare test","DEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","DESCRIPTION":"LLM inference in C/C++","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://github.com/ggml-org/llama.cpp","INHERITED":"toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake","IUSE":"static blas cuda vulkan","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","REQUIRED_USE":"","RESTRICT":"","SLOT":"0","SRC_URI":"https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7556 -\u003e llama-cpp-0.0.7556-4fd59e8.tar.gz"},"metadata_out":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\ncuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\n0\nhttps://api.github.com/repos/ggml-org/llama.cpp/tarball/b7556 -\u003e llama-cpp-0.0.7556-4fd59e8.tar.gz\n\nhttps://github.com/ggml-org/llama.cpp\nMIT\nLLM inference in C/C++\n*\ntoolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake\nstatic blas cuda vulkan\n\n\nvulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake\n7\n\ncompile configure install prepare test\n\n\n\n\n\n","manifest_md5":"130d79604020aa3f4860c95d060d6252","md5":"9aa95d3b5c3a4da6e2e2c2343862df7f","files":[{"src_uri":["https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7556"],"size":"28607013","hashes":{"blake2b":"996020e88bc84a1badaef80f3ab66c2fe95102f1c06fc68fdf2b2d6653530e5a203e4296b6e29e8a675b28728754008627094de8071952b94ece7fd6817e4b83","sha512":"8e6ecc5653f4a8fb202e8e053487695b5ea0ccf124740ea7dada45cbfb3b5a0ce9a10af75c6239898bbace081c5b030962445c48de3b9e06315b8279dea7f5b9"},"name":"llama-cpp-0.0.7556-4fd59e8.tar.gz"}]},"app-ai/llama-cpp-0.0.7622":{"atom":"app-ai/llama-cpp-0.0.7622","category":"app-ai","package":"llama-cpp","revision":"0","catpkg":"app-ai/llama-cpp","eclasses":[["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["xdg-utils","14d00d009167652b1fa363e55effe213"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake","x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"relations_by_kind":{"BDEPEND":["dev-util/vulkan-headers","dev-util/ninja","dev-util/cmake"],"DEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"],"RDEPEND":["x11-drivers/nvidia-drivers","dev-util/vulkan-tools","media-libs/vulkan-layers","media-libs/vulkan-loader","media-libs/shaderc","sci-libs/gsl","virtual/blas","virtual/lapack"]},"metadata":{"BDEPEND":"vulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install prepare test","DEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","DESCRIPTION":"LLM inference in C/C++","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://github.com/ggml-org/llama.cpp","INHERITED":"toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake","IUSE":"static blas cuda vulkan","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )","REQUIRED_USE":"","RESTRICT":"","SLOT":"0","SRC_URI":"https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7622 -\u003e llama-cpp-0.0.7622-c69c7eb.tar.gz"},"metadata_out":"cuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\ncuda? ( x11-drivers/nvidia-drivers ) vulkan? ( dev-util/vulkan-tools media-libs/vulkan-layers media-libs/vulkan-loader media-libs/shaderc sci-libs/gsl ) blas? ( virtual/blas virtual/lapack )\n0\nhttps://api.github.com/repos/ggml-org/llama.cpp/tarball/b7622 -\u003e llama-cpp-0.0.7622-c69c7eb.tar.gz\n\nhttps://github.com/ggml-org/llama.cpp\nMIT\nLLM inference in C/C++\n*\ntoolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic xdg-utils cmake\nstatic blas cuda vulkan\n\n\nvulkan? ( dev-util/vulkan-headers ) dev-util/ninja dev-util/cmake\n7\n\ncompile configure install prepare test\n\n\n\n\n\n","manifest_md5":"130d79604020aa3f4860c95d060d6252","md5":"0108932827f4a4c5c1b1a46656927eea","files":[{"src_uri":["https://api.github.com/repos/ggml-org/llama.cpp/tarball/b7622"],"size":"28630954","hashes":{"blake2b":"9b4a7b3df2d826abea81400ad71bc714d0863518b9d35f58908915360e88f3837993606c0240bc0d35b27f60af74f4ba7a5d2c2134f02a5f5b4de5a7e2eb46d7","sha512":"5662e8c3f63d1cf0134188bf3391234ddb5fcd4308e24135c313cf0d53fc3dd9e67621a9983a44dd8f04e24db49ff15f59bdf7b84bf218db9feb379826c72cc5"},"name":"llama-cpp-0.0.7622-c69c7eb.tar.gz"}]},"app-ai/ollama-0.13.1":{"atom":"app-ai/ollama-0.13.1","category":"app-ai","package":"ollama","revision":"0","catpkg":"app-ai/ollama","eclasses":[["xdg-utils","14d00d009167652b1fa363e55effe213"],["go-module","4668f84c6028afa6bc1e08d6824ff9a3"],["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"],["user","875055ddd565382c728563992ae3abf7"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake","sys-apps/shadow"],"relations_by_kind":{"BDEPEND":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake"],"DEPEND":["sys-apps/shadow"]},"metadata":{"BDEPEND":"\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install postinst prepare setup test unpack","DEPEND":"sys-apps/shadow","DESCRIPTION":"Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://ollama.com","INHERITED":"xdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user","IUSE":"systemd","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"","REQUIRED_USE":"","RESTRICT":"strip","SLOT":"0","SRC_URI":"https://api.github.com/repos/ollama/ollama/tarball/v0.13.1 -\u003e ollama-0.13.1-5317202.tar.gz mirror://macaroni/ollama-0.13.1-mark-go-bundle-5317202.tar.xz -\u003e ollama-0.13.1-mark-go-bundle-5317202.tar.xz"},"metadata_out":"sys-apps/shadow\n\n0\nhttps://api.github.com/repos/ollama/ollama/tarball/v0.13.1 -\u003e ollama-0.13.1-5317202.tar.gz mirror://macaroni/ollama-0.13.1-mark-go-bundle-5317202.tar.xz -\u003e ollama-0.13.1-mark-go-bundle-5317202.tar.xz\nstrip\nhttps://ollama.com\nMIT\nGet up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.\n*\nxdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user\nsystemd\n\n\n\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake\n7\n\ncompile configure install postinst prepare setup test unpack\n\n\n\n\n\n","manifest_md5":"97acf0f897bc25413ab146815fcbc76d","md5":"02b4eafdd2f775ec4ae2e7c3df3fbd4a","files":[{"src_uri":["https://api.github.com/repos/ollama/ollama/tarball/v0.13.1"],"size":"20813074","hashes":{"blake2b":"9ac0bdcc95fbf566ec0cfb78c0e01e7f364c929822c7cb5879a62adf4b607f35f97a19c8abc83247602636c868d0f8343a673902635aad1e63929de3a52c33ae","sha512":"4c5d0263e84a50acc829c159ea136cbb29ef2bb014afc4755d53a952d1727008ff003bdb1ee5d49c520c15f44f869cef49c8278c5e54dc8f8497b085c1242eda"},"name":"ollama-0.13.1-5317202.tar.gz"},{"src_uri":["mirror://macaroni/ollama-0.13.1-mark-go-bundle-5317202.tar.xz"],"size":"124632188","hashes":{"blake2b":"a76478030c6aee51b5581b6aa67822c9467968b438a8271e8c95c3d491e25eff8e69b4e331199ea4c66e64922181333bdd5ebaba14eaba503a5b5f08ea371f58","sha512":"c104681ac235838e3bbd17951ffcf8bf80ff482d6034bf2e4410d1feb6bb13a86cdfb3a03b84dffd38b3c3f84847d3f8b604beee9550b9451590f8fa7ad3bfac"},"name":"ollama-0.13.1-mark-go-bundle-5317202.tar.xz"}]},"app-ai/ollama-0.13.3":{"atom":"app-ai/ollama-0.13.3","category":"app-ai","package":"ollama","revision":"0","catpkg":"app-ai/ollama","eclasses":[["xdg-utils","14d00d009167652b1fa363e55effe213"],["go-module","4668f84c6028afa6bc1e08d6824ff9a3"],["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"],["user","875055ddd565382c728563992ae3abf7"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake","sys-apps/shadow"],"relations_by_kind":{"BDEPEND":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake"],"DEPEND":["sys-apps/shadow"]},"metadata":{"BDEPEND":"\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install postinst prepare setup test unpack","DEPEND":"sys-apps/shadow","DESCRIPTION":"Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://ollama.com","INHERITED":"xdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user","IUSE":"systemd","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"","REQUIRED_USE":"","RESTRICT":"strip","SLOT":"0","SRC_URI":"https://api.github.com/repos/ollama/ollama/tarball/v0.13.3 -\u003e ollama-0.13.3-709f842.tar.gz mirror://macaroni/ollama-0.13.3-mark-go-bundle-709f842.tar.xz -\u003e ollama-0.13.3-mark-go-bundle-709f842.tar.xz"},"metadata_out":"sys-apps/shadow\n\n0\nhttps://api.github.com/repos/ollama/ollama/tarball/v0.13.3 -\u003e ollama-0.13.3-709f842.tar.gz mirror://macaroni/ollama-0.13.3-mark-go-bundle-709f842.tar.xz -\u003e ollama-0.13.3-mark-go-bundle-709f842.tar.xz\nstrip\nhttps://ollama.com\nMIT\nGet up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.\n*\nxdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user\nsystemd\n\n\n\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake\n7\n\ncompile configure install postinst prepare setup test unpack\n\n\n\n\n\n","manifest_md5":"97acf0f897bc25413ab146815fcbc76d","md5":"fc4826f4d1c9a246534f20c68bc4f224","files":[{"src_uri":["https://api.github.com/repos/ollama/ollama/tarball/v0.13.3"],"size":"20920881","hashes":{"blake2b":"9b53f8a42556da41a8f87685ba6307428dbe8728d190c5d83df48299039b1ce59a6af86dc500d414d2dbe3a0c9888024211fd4c7e01f2f664332933e10cfe43a","sha512":"cc391798c553db3da58800fd830b106503fa3f4008489dab9fa58825869cba0d722310171f03897173f5f4afe056a156a15158b71bb81744df8f4682e8e4faa7"},"name":"ollama-0.13.3-709f842.tar.gz"},{"src_uri":["mirror://macaroni/ollama-0.13.3-mark-go-bundle-709f842.tar.xz"],"size":"124632492","hashes":{"blake2b":"d3d8e8d4fc6e2155f633aeadc5753ccb0607faac81fa54022a4e76ef7f9982237bde358da4c2c50f7436750546b971a90e314c496afcb5e98a8e2a6a2934d09a","sha512":"fb95d51d2c28cbc3448e8c5068f60c01ff4f5e65df89ebda961a3c083114c271c683a5e23f4a7a0cc71be38c87244b83c7d6cc88bdefe726f3f3e7c98d65347c"},"name":"ollama-0.13.3-mark-go-bundle-709f842.tar.xz"}]},"app-ai/ollama-0.13.5":{"atom":"app-ai/ollama-0.13.5","category":"app-ai","package":"ollama","revision":"0","catpkg":"app-ai/ollama","eclasses":[["xdg-utils","14d00d009167652b1fa363e55effe213"],["go-module","4668f84c6028afa6bc1e08d6824ff9a3"],["toolchain-funcs","24921b57d6561d87cbef4916a296ada4"],["multilib","d410501a125f99ffb560b0c523cd3d1e"],["multiprocessing","cac3169468f893670dac3e7cb940e045"],["ninja-utils","e7575bc4a90349d76e72777013b2bbc2"],["eutils","6e6c2737b59a4b982de6fb3ecefd87f8"],["flag-o-matic","d0939f99dd528dd0c5ec25284877bf5c"],["cmake","ac7cb516f6a288b6a82bc0649ce49878"],["user","875055ddd565382c728563992ae3abf7"]],"kit":"ai-kit","branch":"mark-xl","relations":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake","sys-apps/shadow"],"relations_by_kind":{"BDEPEND":["dev-lang/go","app-arch/unzip","dev-util/ninja","dev-util/cmake"],"DEPEND":["sys-apps/shadow"]},"metadata":{"BDEPEND":"\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake","DEFINED_PHASES":"compile configure install postinst prepare setup test unpack","DEPEND":"sys-apps/shadow","DESCRIPTION":"Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.","EAPI":"7","HDEPEND":"","HOMEPAGE":"https://ollama.com","INHERITED":"xdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user","IUSE":"systemd","KEYWORDS":"*","LICENSE":"MIT","PDEPEND":"","PROPERTIES":"","PYTHON_COMPAT":"","RDEPEND":"","REQUIRED_USE":"","RESTRICT":"strip","SLOT":"0","SRC_URI":"https://api.github.com/repos/ollama/ollama/tarball/v0.13.5 -\u003e ollama-0.13.5-7325791.tar.gz mirror://macaroni/ollama-0.13.5-mark-go-bundle-7325791.tar.xz -\u003e ollama-0.13.5-mark-go-bundle-7325791.tar.xz"},"metadata_out":"sys-apps/shadow\n\n0\nhttps://api.github.com/repos/ollama/ollama/tarball/v0.13.5 -\u003e ollama-0.13.5-7325791.tar.gz mirror://macaroni/ollama-0.13.5-mark-go-bundle-7325791.tar.xz -\u003e ollama-0.13.5-mark-go-bundle-7325791.tar.xz\nstrip\nhttps://ollama.com\nMIT\nGet up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.\n*\nxdg-utils go-module toolchain-funcs multilib multiprocessing ninja-utils eutils flag-o-matic cmake user\nsystemd\n\n\n\u003e=dev-lang/go-1.24 \u003e=dev-lang/go-1.12 app-arch/unzip dev-util/ninja dev-util/cmake\n7\n\ncompile configure install postinst prepare setup test unpack\n\n\n\n\n\n","manifest_md5":"97acf0f897bc25413ab146815fcbc76d","md5":"6abad3a684ff0ed0a6a851b269eb3edd","files":[{"src_uri":["https://api.github.com/repos/ollama/ollama/tarball/v0.13.5"],"size":"20965208","hashes":{"blake2b":"d5a20b6d6e3f0b600cd9f04a464b4530b9c389fd36bba7a5d5f6c2f74b769e9ce235ffec1bf3445dad695503ee98c6b33b137943df7a51b65a38009e49b1d2aa","sha512":"90538ad4729abc23e2ff8da187895f6382231b08a026a6e27f7ee159fbc3498f6c80e155cc72051c563ef680b9dd932876136e7281836c4e7f16635b01d22687"},"name":"ollama-0.13.5-7325791.tar.gz"},{"src_uri":["mirror://macaroni/ollama-0.13.5-mark-go-bundle-7325791.tar.xz"],"size":"124632372","hashes":{"blake2b":"a2e47b94d69384fb20c267fe1e4bef76a13a12403a32464d10a3749b073c46acef0cbe8f7cc3f60118addb61dfa0a8dcea8afac3cbb46a71da8ef60e134d6cf9","sha512":"9f92c4fb9bed3bb46304afb0aea5bd9c89bfe5c7c11223eaf9040cf9f0f9d68aa2b55974066564143741a6c258f978e5319cc3b6f8dc5feecae858f2cbf6b3e3"},"name":"ollama-0.13.5-mark-go-bundle-7325791.tar.xz"}]}}}