{"id":247234,"date":"2025-07-08T05:54:09","date_gmt":"2025-07-08T05:54:09","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/247234\/"},"modified":"2025-07-08T05:54:09","modified_gmt":"2025-07-08T05:54:09","slug":"weka-debuts-neuralmesh-axon-for-exascale-ai-deployments","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/247234\/","title":{"rendered":"WEKA Debuts NeuralMesh Axon For Exascale AI Deployments"},"content":{"rendered":"<p class=\"prntac\">New Offering Delivers a Unique Fusion Architecture That&#8217;s Being Leveraged by Industry-Leading AI Pioneers Like Cohere, CoreWeave, and NVIDIA to Deliver Breakthrough Performance Gains and Reduce Infrastructure Requirements For Massive AI Training and Inference Workloads<\/p>\n<p>PARIS and CAMPBELL, Calif., July 8, 2025 \/PRNewswire\/ &#8212; From RAISE SUMMIT 2025: <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=3637918604&amp;u=https%3A%2F%2Fwww.weka.io%2F&amp;a=WEKA\" rel=\"nofollow noopener\" target=\"_blank\">WEKA<\/a>\u00a0unveiled NeuralMesh Axon, a breakthrough storage system that leverages an innovative fusion architecture designed to address the fundamental challenges of running exascale AI applications and workloads. NeuralMesh Axon seamlessly fuses with GPU servers and AI factories to streamline deployments, reduce costs, and significantly enhance AI workload responsiveness and performance, transforming underutilized GPU resources into a unified, high-performance infrastructure layer.<\/p>\n<p>    &#13;<br \/>\n      &#13;<br \/>\n&#13;<br \/>\n      &#13;<\/p>\n<p>          <a href=\"#\" class=\"tabfocus\" role=\"button\"><img title=\"WEKA's NeuralMesh Axon delivers an innovative fusion architecture designed to address the fundamental challenges of running exascale AI applications and workloads.\" data-getimg=\"https:\/\/mma.prnewswire.com\/media\/2726225\/WEKA_og_pr_wire_neuralmesh_axon.jpg?w=600\" id=\"imageid_2\" alt=\"WEKA's NeuralMesh Axon delivers an innovative fusion architecture designed to address the fundamental challenges of running exascale AI applications and workloads.\" class=\"gallery-thumb img-responsive\" rel=\"newsImage\" itemprop=\"contentUrl\" loading=\"lazy\"\/><\/a><\/p>\n<p>        &#13;<br \/>\n      &#13;<br \/>\n&#13;<br \/>\n      &#13;<br \/>\n&#13;<br \/>\n      &#13;<br \/>\n        WEKA&#8217;s NeuralMesh Axon delivers an innovative fusion architecture designed to address the fundamental challenges of running exascale AI applications and workloads.&#13;<br \/>\n      &#13;<\/p>\n<p>Building on the company&#8217;s <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=1792786488&amp;u=https%3A%2F%2Fwww.weka.io%2Fcompany%2Fweka-newsroom%2Fpress-releases%2Fweka-introduces-neuralmesh%2F&amp;a=recently+announced+NeuralMesh+storage+system\" rel=\"nofollow noopener\" target=\"_blank\">recently announced NeuralMesh storage system<\/a>, the new offering enhances its containerized microservices architecture with powerful embedded functionality, enabling AI pioneers, AI cloud and\u00a0neocloud service providers to accelerate AI model development at extreme scale, particularly when combined with <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=3950171428&amp;u=https%3A%2F%2Fwww.nvidia.com%2Fen-us%2Fdata-center%2Fproducts%2Fai-enterprise%2F&amp;a=NVIDIA+AI+Enterprise\" rel=\"nofollow noopener\" target=\"_blank\">NVIDIA AI Enterprise<\/a>\u00a0software stacks for advanced model training and inference optimization. NeuralMesh Axon also supports real-time reasoning, with significantly improved time-to-first-token and overall token throughput, enabling customers to bring innovations to market faster.<\/p>\n<p><b>AI Infrastructure Obstacles Compound at Exascale<\/b><br class=\"dnr\"\/>Performance is make-or-break for large language model (LLM) training and inference workloads, especially when running at extreme scale. Organizations that run massive AI workloads on traditional storage architectures, which rely on replication-heavy approaches, waste NVMe capacity, face significant inefficiencies, and struggle with unpredictable performance and resource allocation.<\/p>\n<p>The reason? Traditional architectures weren&#8217;t designed to process and store massive volumes of data in real-time. They create latency and bottlenecks in data pipelines and AI workflows that can cripple exascale AI deployments. Underutilized GPU servers and outdated data architectures turn premium hardware into idle capital, resulting in\u00a0costly downtime for training workloads. Inference workloads struggle with memory-bound barriers, including key-value (KV) caches and hot data, resulting in reduced throughput and increased infrastructure strain. Limited KV cache offload capacity creates data access bottlenecks and complicates resource allocation for incoming prompts, directly impacting operational expenses and time-to-insight. Many organizations are transitioning to NVIDIA accelerated compute servers, paired with NVIDIA AI Enterprise software, to address these challenges. However, without modern storage integration, they still encounter significant limitations in pipeline efficiency and overall GPU utilization.<\/p>\n<p><b>Built For The World&#8217;s Largest and Most Demanding Accelerated Compute Environments<\/b><br class=\"dnr\"\/>To address these challenges,\u00a0NeuralMesh Axon&#8217;s high-performance, resilient storage fabric fuses directly into accelerated compute servers by leveraging local NVMe, spare CPU cores, and its existing network infrastructure. This unified, software-defined compute and storage layer delivers consistent microsecond latency for both local and remote workloads\u2014outpacing traditional local protocols like NFS.<\/p>\n<p>Additionally, when leveraging WEKA&#8217;s Augmented Memory Grid capability, it can provide near-memory speeds for KV cache loads at massive scale. Unlike replication-heavy approaches that squander aggregate capacity and collapse under failures, NeuralMesh Axon&#8217;s\u00a0unique erasure coding design tolerates up to four simultaneous node losses, sustains full throughput during rebuilds, and enables predefined resource allocation across the existing NVMe, CPU cores, and networking resources\u2014transforming isolated disks into a memory-like storage pool at exascale and beyond while providing consistent low latency access to all addressable data.<\/p>\n<p>Cloud service providers and AI innovators operating at exascale require infrastructure solutions that can match the exponential growth in model complexity and dataset sizes. NeuralMesh Axon is specifically designed for organizations operating at the forefront of AI innovation that require immediate, extreme-scale performance rather than gradual scaling over time. This includes AI cloud providers and neoclouds building AI services, regional AI factories, major cloud providers developing AI solutions for enterprise customers, and large enterprise organizations deploying the most demanding AI inference and training solutions that must agilely scale and optimize their AI infrastructure investments to support rapid innovation cycles.<\/p>\n<p><b>Delivering Game-Changing Performance for Accelerated AI Innovation<\/b><br class=\"dnr\"\/>Early adopters, including <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=2704359166&amp;u=https%3A%2F%2Fcohere.com%2F&amp;a=Cohere\" rel=\"nofollow noopener\" target=\"_blank\">Cohere<\/a>, the industry&#8217;s leading security-first enterprise AI company, are already seeing transformational results.<\/p>\n<p>Cohere is among WEKA&#8217;s first customers to deploy NeuralMesh Axon to power its AI model training and inference workloads. Faced with high innovation costs, data transfer bottlenecks, and underutilized GPUs, Cohere first deployed NeuralMesh Axon in the public cloud to unify its AI stack and streamline operations.<\/p>\n<p>&#8220;For AI model builders, speed,\u00a0GPU optimization, and cost-efficiency are mission-critical. That means using less hardware, generating more tokens, and running more models\u2014without waiting on capacity or migrating data,&#8221; said Autumn Moulder, vice president of engineering at Cohere. &#8220;Embedding WEKA&#8217;s NeuralMesh Axon into our GPU servers enabled us to maximize utilization and accelerate every step of our AI pipelines. The performance gains have been game-changing: Inference deployments that used to take five minutes can occur in 15 seconds, with 10 times faster checkpointing. Our team can now iterate on and bring revolutionary new AI models, like North, to market with unprecedented speed.&#8221;<\/p>\n<p>To improve training and help develop\u00a0<a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=1147650451&amp;u=https%3A%2F%2Fcohere.com%2Fnorth&amp;a=North\" rel=\"nofollow noopener\" target=\"_blank\">North<\/a>, Cohere&#8217;s secure AI agents platform, the company is deploying WEKA&#8217;s NeuralMesh Axon on <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=2162600138&amp;u=https%3A%2F%2Fwww.coreweave.com%2F&amp;a=CoreWeave+Cloud\" rel=\"nofollow noopener\" target=\"_blank\">CoreWeave Cloud<\/a>, creating a robust foundation to support real-time reasoning and deliver exceptional experiences for Cohere&#8217;s end customers.<\/p>\n<p>&#8220;We&#8217;re entering an era where AI advancement transcends raw compute alone\u2014it&#8217;s unleashed by intelligent infrastructure design. CoreWeave is redefining what&#8217;s possible for AI pioneers by eliminating the complexities that constrain AI at scale,&#8221; said Peter Salanki, CTO and co-founder at CoreWeave. &#8220;With WEKA&#8217;s NeuralMesh Axon seamlessly integrated into CoreWeave&#8217;s AI cloud infrastructure, we&#8217;re bringing processing power directly to data, achieving microsecond latencies that reduce I\/O wait time and deliver more than 30 GB\/s read, 12 GB\/s write, and 1 million IOPS to an individual GPU server. This breakthrough approach increases GPU utilization and empowers Cohere with the performance foundation they need to shatter inference speed barriers and deliver advanced AI solutions to their customers.&#8221;<\/p>\n<p>&#8220;AI factories are defining the future of AI infrastructure built on NVIDIA accelerated compute and our ecosystem of NVIDIA Cloud Partners,&#8221; said Marc Hamilton, vice president of solutions architecture and engineering at NVIDIA. &#8220;By optimizing inference at scale and embedding ultra-low latency NVMe storage close to the GPUs, organizations can unlock more bandwidth and extend the available on-GPU memory for any capacity. Partner solutions like WEKA&#8217;s NeuralMesh Axon deployed with CoreWeave provide a critical foundation for accelerated inferencing while enabling next-generation AI services with exceptional performance and cost efficiency.&#8221;<\/p>\n<p><b>The Benefits of Fusing Storage and Compute For AI Innovation<\/b><br class=\"dnr\"\/>NeuralMesh Axon delivers immediate, measurable improvements for AI builders and cloud service providers operating at exascale, including:\u00a0<\/p>\n<ul type=\"disc\">\n<li><b>Expanded Memory With Accelerated Token Throughput:\u00a0<\/b>Provides tight integration with WEKA&#8217;s Augmented Memory Grid technology, extending GPU memory by leveraging it as a token warehouse. This has delivered a 20x improvement in time to first token performance across multiple customer deployments, enabling larger context windows and significantly improved token processing efficiency for inference-intensive workloads. Furthermore, NeuralMesh Axon enables customers to dynamically adjust compute and storage resources and seamlessly supports just-in-time training and just-in-time inference.<\/li>\n<li><b>Huge\u00a0GPU Acceleration and Efficiency Gains:<\/b>\u00a0Customers are achieving dramatic performance and GPU utilization improvements with NeuralMesh Axon, with AI model training workloads exceeding 90%\u2014a three-fold improvement over the industry average. NeuralMesh Axon also reduces the required rack space, power, and cooling requirements in on-premises data centers, helping to lower infrastructure costs and complexity by leveraging existing server resources.<\/li>\n<li><b>Immediate Scale for Massive AI Workflows:<\/b>\u00a0Designed for AI innovators who need immediate extreme scale, rather than to grow over time. NeuralMesh Axon&#8217;s containerized microservices architecture and cloud-native design enable organizations to scale storage performance and capacity independently while maintaining consistent performance characteristics across hybrid and multicloud environments.<\/li>\n<li><b>Enables Teams to Focus on Building AI, Not Infrastructure<\/b>: Runs seamlessly across hybrid and cloud environments, integrating with existing Kubernetes and container environments to eliminate the need for external storage infrastructure and reduce complexity.<\/li>\n<\/ul>\n<p>&#8220;The infrastructure challenges of\u00a0exascale AI are unlike anything the industry has faced before. At WEKA, we&#8217;re seeing organizations struggle with low GPU utilization during training and GPU overload during inference, while AI costs spiral into millions per model and agent,&#8221; said Ajay Singh, chief product officer at WEKA. &#8220;That&#8217;s why we engineered NeuralMesh Axon, born from our deep focus on optimizing every layer of AI infrastructure from the GPU up. Now, AI-first organizations can achieve the performance and cost efficiency required for competitive AI innovation when running at exascale and beyond.&#8221;<\/p>\n<p><b>Availability<\/b><br class=\"dnr\"\/>NeuralMesh Axon is currently available in limited release for large-scale enterprise AI and neocloud customers, with general availability scheduled for fall 2025. For more information, visit:<\/p>\n<p><b>About\u00a0WEKA<\/b><br class=\"dnr\"\/>WEKA is transforming how organizations build, run, and scale AI workflows through NeuralMesh\u2122, its intelligent, adaptive mesh storage system. Unlike traditional data infrastructure, which becomes more fragile as AI environments expand, NeuralMesh becomes faster, stronger, and more efficient as it scales, growing with your AI environment to provide a flexible foundation for enterprise and agentic AI innovation. Trusted by 30% of the Fortune 50 and the world&#8217;s leading neoclouds and AI innovators, NeuralMesh maximizes GPU utilization, accelerates time to first token, and lowers the cost of AI innovation. Learn more at <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=2134788139&amp;u=http%3A%2F%2Fwww.weka.io%2F&amp;a=www.weka.io\" rel=\"nofollow noopener\" target=\"_blank\">www.weka.io<\/a>, or connect with us on <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=483385444&amp;u=https%3A%2F%2Fwww.linkedin.com%2Fcompany%2Fweka-io&amp;a=LinkedIn\" rel=\"nofollow noopener\" target=\"_blank\">LinkedIn<\/a>\u00a0and <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=2520009930&amp;u=https%3A%2F%2Fx.com%2FWekaIO%2F&amp;a=X\" rel=\"nofollow noopener\" target=\"_blank\">X<\/a>.<\/p>\n<p class=\"prntac\">WEKA and the W logo are registered trademarks of WekaIO, Inc. Other trade names herein may be trademarks of their respective owners.<\/p>\n<p>Photo &#8211; <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=3475262379&amp;u=https%3A%2F%2Fmma.prnewswire.com%2Fmedia%2F2726225%2FWEKA_og_pr_wire_neuralmesh_axon.jpg&amp;a=https%3A%2F%2Fmma.prnewswire.com%2Fmedia%2F2726225%2FWEKA_og_pr_wire_neuralmesh_axon.jpg\" rel=\"nofollow noopener\" target=\"_blank\">https:\/\/mma.prnewswire.com\/media\/2726225\/WEKA_og_pr_wire_neuralmesh_axon.jpg<\/a><br class=\"dnr\"\/>Logo &#8211; <a href=\"https:\/\/c212.net\/c\/link\/?t=0&amp;l=en&amp;o=4463805-1&amp;h=2021096678&amp;u=https%3A%2F%2Fmma.prnewswire.com%2Fmedia%2F1796062%2FWEKA_v1_Logo_new.jpg&amp;a=https%3A%2F%2Fmma.prnewswire.com%2Fmedia%2F1796062%2FWEKA_v1_Logo_new.jpg\" rel=\"nofollow noopener\" target=\"_blank\">https:\/\/mma.prnewswire.com\/media\/1796062\/WEKA_v1_Logo_new.jpg<\/a><\/p>\n<p><img decoding=\"async\" alt=\"\" src=\"https:\/\/rt.prnewswire.com\/rt.gif?NewsItemId=EN25632&amp;Transmission_Id=202507080100PR_NEWS_EURO_ND__EN25632&amp;DateId=20250708\" style=\"border:0px; width:1px; height:1px;\"\/><\/p>\n","protected":false},"excerpt":{"rendered":"New Offering Delivers a Unique Fusion Architecture That&#8217;s Being Leveraged by Industry-Leading AI Pioneers Like Cohere, CoreWeave, and&hellip;\n","protected":false},"author":2,"featured_media":247235,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3164],"tags":[1942,3730,3728,22101,3731,3284,95775,3732,53,16,15,95774],"class_list":{"0":"post-247234","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-artificial-intelligence","9":"tag-cloud-computing-internet-of-things","10":"tag-computer-electronics","11":"tag-computer-hardware","12":"tag-computer-software","13":"tag-computing","14":"tag-general-sports","15":"tag-new-products-services","16":"tag-technology","17":"tag-uk","18":"tag-united-kingdom","19":"tag-weka"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/114816069181784927","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/247234","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=247234"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/247234\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/247235"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=247234"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=247234"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=247234"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}