{"id":2496,"date":"2025-07-13T06:38:58","date_gmt":"2025-07-13T06:38:58","guid":{"rendered":"https:\/\/violethoward.com\/new\/cracking-ais-storage-bottleneck-and-supercharging-inference-at-the-edge\/"},"modified":"2025-07-13T06:38:58","modified_gmt":"2025-07-13T06:38:58","slug":"cracking-ais-storage-bottleneck-and-supercharging-inference-at-the-edge","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/cracking-ais-storage-bottleneck-and-supercharging-inference-at-the-edge\/","title":{"rendered":"Cracking AI\u2019s storage bottleneck and supercharging inference at the edge"},"content":{"rendered":" \r\n<br><div>\n\t\t\t\t<div id=\"boilerplate_2682874\" class=\"post-boilerplate boilerplate-before\">\n<p><em>Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders.<\/em> <em>Subscribe Now<\/em><\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity is-style-wide\"\/>\n<\/div><p>As AI applications increasingly permeate enterprise operations, from enhancing patient care through advanced medical imaging to powering complex fraud detection models and even aiding wildlife conservation, a critical bottleneck often emerges: data storage.<\/p>\n\n\n\n<p>During VentureBeat\u2019s Transform 2025, Greg Matson, head of products and marketing, Solidigm and Roger Cummings, CEO of PEAK:AIO spoke with Michael Stewart, managing partner at M12 about how innovations in storage technology enables enterprise AI use cases in healthcare.<\/p>\n\n\n\n<figure class=\"wp-block-embed aligncenter is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><p>\n<iframe loading=\"lazy\" title=\"Cracking AI\u2019s Storage Bottleneck | VB Transform 2025\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube.com\/embed\/oxjDHmcsi3A?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe>\n<\/p><\/figure>\n\n\n\n<p>The MONAI framework is a breakthrough in medical imaging, building it faster, more safely, and more securely. Advances in storage technology is what enables researchers to build on top of this framework, iterate and innovate quickly. PEAK:AIO partnered with Solidgm to integrate power-efficient, performant, and high-capacity storage which enabled MONAI to store more than two million full-body CT scans on a single node within their IT environment.<\/p>\n\n\n\n<p>\u201cAs enterprise AI infrastructure evolves rapidly, storage hardware increasingly needs to be tailored to specific use cases, depending on where they are in the AI data pipeline,\u201d Matson said. \u201cThe type of use case we talked about with MONAI, an edge-use case, as well as the feeding of a training cluster, are well served by very high-capacity solid-state storage solutions, but the actual inference and model training need something different. That\u2019s a very high-performance, very high I\/O-per-second requirement from the SSD. For us, RAG is bifurcating the types of products that we make and the types of integrations we have to make with the software.\u201d<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-improving-ai-inference-at-the-edge\"><strong>Improving AI inference at the edge<\/strong><\/h2>\n\n\n\n<p>For peak performance at the edge, it\u2019s critical to scale storage down to a single node, in order to bring inference closer to the data. And what\u2019s key is removing memory bottlenecks. That can be done by making memory a part of the AI infrastructure, in order to scale it along with data and metadata. The proximity of data to compute dramatically increases the time to insight.<\/p>\n\n\n\n<p>\u201cYou see all the huge deployments, the big green field data centers for AI, using very specific hardware designs to be able to bring the data as close as possible to the GPUs,\u201d Matson said. \u201cThey\u2019ve been building out their data centers with very high-capacity solid-state storage, to bring petabyte-level storage, very accessible at very high speeds, to the GPUs. Now, that same technology is happening in a microcosm at the edge and in the enterprise.\u201d<\/p>\n\n\n\n<p>It\u2019s becoming critical to purchasers of AI systems to ensure you\u2019re getting the most performance out of your system by running it on all solid state. That allows you to bring huge amounts of data, and enables incredible processing power in a small system at the edge.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-the-future-of-ai-hardware\"><strong>The future of AI hardware<\/strong><\/h2>\n\n\n\n<p>\u201cIt\u2019s imperative that we provide solutions that are open, scalable, and at memory speed, using some of the latest and greatest technology out there to do that,\u201d Cummings said. \u201cThat\u2019s our goal as a company, to provide that openness, that speed, and the scale that organizations need. I think you\u2019re going to see the economies match that as well.\u201d<\/p>\n\n\n\n<p>For the overall training and inference data pipeline, and within inference itself, hardware needs will keep increasing, whether it\u2019s a very high-speed SSD or a very high-capacity solution that\u2019s power efficient.<\/p>\n\n\n\n<p>\u201cI would say it\u2019s going to move even further toward very high-capacity, whether it\u2019s a one-petabyte SSD out a couple of years from now that runs at very low power and that can basically replace four times as many hard drives, or a very high-performance product that\u2019s almost near memory speeds,\u201d Matson said. \u201cYou\u2019ll see that the big GPU vendors are looking at how to define the next storage architecture, so that it can help augment, very closely, the HBM in the system. What was a general-purpose SSD in cloud computing is now bifurcating into capacity and performance. We\u2019ll keep doing that further out in both directions over the next five or 10 years.\u201d<\/p>\n\n\n\n\n<div id=\"boilerplate_2660155\" class=\"post-boilerplate boilerplate-after\"><div class=\"Boilerplate__newsletter-container vb\">\n<div class=\"Boilerplate__newsletter-main\">\n<p><strong>Daily insights on business use cases with VB Daily<\/strong><\/p>\n<p class=\"copy\">If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.<\/p>\n<p class=\"Form__newsletter-legal\">Read our Privacy Policy<\/p>\n<p class=\"Form__success\" id=\"boilerplateNewsletterConfirmation\">\n\t\t\t\t\tThanks for subscribing. Check out more VB newsletters here.\n\t\t\t\t<\/p>\n<p class=\"Form__error\">An error occured.<\/p>\n<\/p><\/div>\n<div class=\"image-container\">\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/venturebeat.com\/wp-content\/themes\/vb-news\/brand\/img\/vb-daily-phone.png\" alt=\"\"\/>\n\t\t\t\t<\/div>\n<\/p><\/div>\n<\/div>\t\t\t<\/div>\r\n<br>\r\n<br><a href=\"https:\/\/venturebeat.com\/ai\/cracking-ais-storage-bottleneck-and-supercharging-inference-at-the-edge\/\">Source link <\/a>","protected":false},"excerpt":{"rendered":"<p>Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now As AI applications increasingly permeate enterprise operations, from enhancing patient care through advanced medical imaging to powering complex fraud detection models and even aiding wildlife conservation, a critical bottleneck [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":2497,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[33],"tags":[],"class_list":["post-2496","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-automation"],"aioseo_notices":[],"jetpack_featured_media_url":"https:\/\/violethoward.com\/new\/wp-content\/uploads\/2025\/07\/VBTRANSFORM25-0204-X2.jpg","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/2496","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/comments?post=2496"}],"version-history":[{"count":0,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/2496\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media\/2497"}],"wp:attachment":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media?parent=2496"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/categories?post=2496"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/tags?post=2496"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69e302c146fa5c92dc28ac12. Config Timestamp: 2026-04-18 04:04:16 UTC, Cached Timestamp: 2026-04-29 13:41:02 UTC -->