• Menu

The Storage Inquirer

January 22, 2021
Skip to content
  • News U Can Use
  • Old School Storage
  • Hall of Fame
  • Data Storage Jobs
Home › News U can Use › The Most Interesting Thing in IBM’s Big Storage Announcement

The Most Interesting Thing in IBM’s Big Storage Announcement

 Author on Source    11/02/2018    News U can Use

This week IBM made a massive storage announcement covering a variety of impressive offerings. This announcement included huge upgrades to its flash solution (including a noteworthy entry-level offering), Non-Volatile Memory Express over Fabrics (NVMe-oF) updates, a very interesting storage gridlock tool, a bunch of tape updates and a series of new and updated SAP HANA solutions.

But the most interesting to me was IBM Spectrum Discover. It is the first use of AI at scale that I’ve seen (I don’t see everything) using AI to speed critical data research.

Pivot to Importance

Over the last decade I’ve watched a number of powerful IT types argue that that the industry’s focus on big data was wrong headed. The most compelling talk I saw was by President Obama’s former election IT manager, who waxed eloquent on why this focus on collecting massive amounts of data, before really understanding what to do with it, was a colossally bad idea.

It led to the creation of huge repositories, like the one the NSA built in the middle of the country that doesn’t appear to be doing much more than running up huge electrical bills at the moment. The lesson was that the focus should have always been on what was going to be done with the data rather than the collection of it.

https://o1.qnsr.com/log/p.gif?;n=203;c=204657336;s=9478;x=7936;f=201808231619130;u=j;z=TIMESTAMP;a=20403940;e=i

My own experience, both in market research and internal audit, is that it is better to have small samples of data but strong analysis, than it is to have overwhelming amounts of data you can’t parse and effectively analyze.

The goal is to have an actionable result — not to create an enormous stinking pile of unstructured data that is equally expensive and can’t be analyzed. But thanks to the effort of “big data,” I’m aware that there are a lot of enterprises and government organizations like the NSA that are now in this mess.

IBM Spectrum Discover

IBM Spectrum Discover is a brand-new AI driven product that appears to be specifically targeting this problem. It has the ability to go into one of these colossal and largely unmanageable data repositories and fix them so they can be properly analyzed. This offering, which came out of IBM Research (indicating it is cutting edge and likely unique), enhances and then leverages the metadata surrounding each element at scale. It does this by rapidly ingesting, consolidating and indexing this metadata for the billions of files in these immense repositories so that the data can then either be sampled and analyzed or potentially analyzed as a whole. (This last tends to be costly and unnecessary if a sample will accomplish much the same result.)

What is also fascinating about this offering is that it initially doesn’t just support IBM Cloud Object Storage and IBM Spectrum Scale, but it will also support Dell EMC Isilon in 2019. IBM, of late, has become far more aggressively hardware platform-agnostic, which means even those that haven’t invested in IBM’s own storage solutions will increasingly benefit from existing and future IBM tools like IBM Spectrum Discover.

Big Data and Big, Bad Decisions

The intense focus on big data around a decade ago led to some really bad decisions with regard to priorities. The biggest was the excessive focus on collecting data, which should have been preceded by a focus on what the heck you were going to do with it. This resulted in some huge unstructured data repositories that are wicked expensive and aren’t providing much value now. IBM’s Spectrum Discover appears to be uniquely targeted at that problem, making it the most important part of the massive IBM announcement that came out this week.

I think this also highlights a common bad practice of not fully thinking through a technology solution when you buy it. Vendors often want to focus you on buying more hardware and software, but your focus needs to be on what will result. Having an enormously expensive data repository you can’t effectively analyze is really not a ton better than not having one — and given the cost, I’d argue it could be a ton worse.

Information is always more important than raw data and should always remain in your sights when making a decision in this area.

Photo courtesy of Shutterstock.

Source: The Most Interesting Thing in IBM’s Big Storage Announcement

 AI, analysis, analytics, Big Data, Dell EMC Isilon, Flash, IBM Research, IBM Spectrum Discover, metadata, NSA, NVMe-oF, SAP HANA, storage, Tape, via bookmarklet
Post navigation
 A quarter of data in the cloud can be classed as sensitive | Telecom Asia | Qumulo Welcomes Robyn Singh as Vice President of People 

Archives

  • May 2019
  • April 2019
  • March 2019
  • February 2019
  • January 2019
  • December 2018
  • November 2018
  • October 2018
  • September 2018
  • August 2018
  • July 2018
  • June 2018
  • May 2018
  • April 2018
  • March 2018
  • February 2018
  • January 2018
  • December 2017
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • July 2017
  • June 2017
  • May 2017
  • April 2017
  • March 2017
  • February 2017
  • January 2017
  • December 2016
  • May 2016
  • April 2016
  • March 2016
  • February 2016
  • January 2016
  • December 2015
  • November 2015
  • October 2015
  • September 2015
  • August 2015
  • July 2015
  • June 2015
  • Home
  • Data Storage Jobs
  • News U Can Use
  • Hall of Fame
  • Old School Storage
© 2021 - The Storage Inquirer
Newsframe Theme by Edward R. Jenkins