A Trip to the Cyber Range: An Immersive Response and Strategic Planning Exercise

By Joe Clabby, Clabby Analytics

Does your organization have a comprehensive plan in place to deal with major data breaches? Does your organization have the right tools and the right policies and procedures in place to isolate the cause of a breach and shut it down? Does that plan include ways to analyze breaches; ways to assess the damage; and ways to interact with the onslaught of press, customers and regulators who will demand answers and assurances afterward? Do your employees know their exact responsibilities should a breach occur?   These are the types of questions that business and information technology (IT) executives will find themselves asking should they visit IBM’s new X-Force Command Center Cyber Range in Cambridge, Massachusetts.

Last week when I had the opportunity to visit IBM’s Cyber Range I was guided to a large conference room that featured over 30 displays– in an imaginary immersive security operations center. IBM security faculty started my lesson in cybersecurity breach response with an overview of the Cyber Range, followed by a discussion of goals and objectives.  The individual who guided me through the exercise explained a bit about typical security challenges, and then detailed how the security drill would work. Essentially, I would be in charge of a fictitious company and guide its response to a security breach in which customer data was published on the web for all to see.

And then, all heck broke loose. The exercise started with a blindside – a crisis situation I was not prepared to handle. (I’ll not describe the blindside – it would ruin the fun for future attendees – but trust me, it catches you off-guard).

As I attempted to deal with the crisis, a litany of questions jumped into my mind:

  • What data has been published?
  • Was the leak plugged?
  • Was it internal or external?
  • At what point do I address the media, and what do I tell them?

As the crisis unfolded, I was asked in rapid fashion to handle the customer outcry, the press inquiries, and to deal with regulators. Again, this is another activity that I’ll be scarce on details with in order to maintain the element of surprise – but let’s just say this: as you go through this exercise you realize that a plan should have been put in place to handle a crisis like the one I was facing BEFORE the situation occurred. After around forty-five stressful minutes, the exercise drew to a close with a final press interview, which IBM confirmed that I handled well.

Looking back at the experience I realized that there has got to be a process in place, supported by the right security tools, the right project management/process flow tools — and supported by specific individuals assigned to handle a wide variety of tasks that must be addressed should a breach occur.

The tools

As I watched the exercise unfold, I was shown a lot of tools that were used to identify the source of the breach; to analyze internal behaviors; to manage the process flow; to assess the potential damage in terms of regulatory fines; and more. I have seen most of IBM’s security product offerings in demonstrations at various conferences – but I had never seen them all in action, working together in a seamless, integrated fashion. Impressive.

I have written about several of IBM’s security tools over the past few years (see this report, this report and this report) – so I had a working knowledge of the features and functions of the respective products. I was most impressed by IBM’s Resilient environment, an incident response platform that provides a means to work collaboratively through the breach process in an organized, project management-like fashion.

IBM Resilient offers a series of working “playbooks” that help guide executives through all of the steps of the breach process. Activities can be set up in advance, with contact names, a description of the task that must be performed, and the name and contact information for individuals assigned to execute tasks. This software can even extrapolate a company’s financial exposure in terms of regulatory fines (and remember that if certain tasks aren’t executed in certain time frames, the fines go up exponentially).

IBM’s i2 visual investigation and analysis tools played a major role in the breach analysis process.

As I claim in one of the above-mentioned reports, I contend that cognitive computing is on a path to become pervasive in the security community. During the course of the breach exercise, IBM’s cognitively-enabled QRadar was used to formulate a “threat query” which it forwarded to IBM Watson for Cyber Security for additional analysis.

Watson searched its corpus of structured and unstructured data – and it searches for other threat entities – and let me know about threats that may have been related to the breach that I was attempting to manage. The information gleaned by Watson for Cyber Security was refined by QRadar – using insights identified by QRadar, by Watson for Cyber Security and by the security administrator. In a relatively short amount of time these tools enabled me to identify the source of the breach.

IBM’s BigFix Detect was also used in the analysis process. This program is an endpoint detection and response (EDR) environment designed to identify malicious behavior at the point where external threats begin. When used with Watson for Cyber Security, the combo can deliver targeted remediation to endpoints that have been compromised quickly – within minutes – helping cut off cyber attacks.

IBM did not overtly try and sell me on any of these products – the company merely showed what needed to be done during a cyber attack, then used the tools at its disposal to do so. Interestingly, IBM did not mention any of the services that it could use to help me implement a breach response plan or formulate a comprehensive security plan. This was likely due to the abbreviated nature of my visit – the exercise can run anywhere from 4-8 hours, whereas I could only devote 60 minutes to the exercise.

Service offerings

Though IBM did not describe its services, I think those offerings are worth mentioning here. I know from previous research that IBM’s security services fall into seven categories: 1) data and application security services; 2) offensive security testing; 3) incident response and intelligence services; 4) identity and access management; 5) infrastructure and endpoint security; 6) security strategy, risk and compliance; and, 7) security intelligence and optimization)

As background, IBM’s data and application security services help secure data and applications. The company’s data security capabilities are based on its Guardium offering, which provides data and file activity monitoring, threat detection analytics, vulnerability and more. The company also offers application security on cloud consulting services. IBM’s offensive security testing services include a programmatic approach to security testing of all types, including human, process, hardware, IoT, application and infrastructure.

The IBM X-Force Red portal helps provide visibility into asset vulnerabilities and offensive security reporting facilities. IBM X-Force Incident Response and Intelligence Services are backed by an organization that helps prepare clients to instantly respond to security incidents. The services offered include a proactive retainer program known as IBM X-Force IRIS Vision Retainer and cybersecurity consulting services to deal with active threats.

IBM’s identity and access management services protect against breaches, and include identity and access strategy and assessment, cloud identity and insider threat protection services. IBM’s infrastructure and endpoint security aim at transforming existing email, Web, network, server and endpoint environments into modern, well secured environments through managed security services, cloud security services, network protection services (such as a managed firewall service), and through managed detection and response services.

The company’s security strategy, risk and compliance service aims at meeting regulatory requirements. This service makes recommendations for the better management of risks, compliance and governance – and includes IBM’s GDPR privacy service.

IBM’s security intelligence and optimization service aims at proactively detecting and prioritizing threats. This service is designed to assist clients throughout the Threat Management lifecycle, including strategic security operations center planning and deployment, 24×7 threat monitoring and analysis, rules and use case management, and closed-loop intelligence processing. This offering also includes services from the IBM X-Force Command Centers such as the IBM X-Force Hosted Threat Analysis Service. Managed SIEM and strategic SOC consulting are two important services in this area.

Summary observations

I thoroughly enjoyed my visit to IBM’s Cyber Range. It was a fun, immersive and extremely informative exercise—a great way to spend an afternoon in Cambridge other than strolling Havad Yahd.

When I started the exercise, I had no idea what to expect. It could’ve been like other briefings – typically discussions of a vendor’s strategies, products and services. But IBM has taken this process to the next level, making it an immersive learning and teaching experience. I “got into” the exercise big time, marveling at the technologies I had at my disposal. IBM’s suite of integrated tools made it easy for me to ascertain the source of the problem and respond rapidly, in an organized fashion, to a data breach.

Due to time constraints, I could not finish the full experience. I was told that it can last for hours, and that teams of security personnel are provided learning materials and planning materials that they fill out in real time to help them figure out how to build their own security strategies. The Cyber Range is really an educational opportunity made fun through the use of immersion technologies.

I loved the parting words of the lead faculty member as he concluded our time together. “The Cyber Range is where security best practices meet. It’s where the game of CLUE converges with security process flow on a Disney roller coaster ride.” From my perspective, he’s 100% right – it’s fun, it’s exciting and it is a fantastic learning experience for those looking to IBM to help them build a comprehensive, fast-response security breach plan.

Posted in Uncategorized | Leave a comment

PowerPlex 2017 – Makers Making an Impact

By Jane Clabby, Clabby Analytics

This year’s PowerPlex conference was held in Atlanta, Georgia on May 8-11. The event built upon the “making things” theme introduced last year—highlighting the manufacturing focus of Plex, as well as the company’s customers, prospects and partners. Each attendee’s badge told us, for example, “we make automobile braking systems” or “we make candy” or “we make things happen”.

Plex Systems, based in Troy Michigan, makes cloud-based manufacturing ERP software linking the “shop floor to the top floor” with an end-to-end solution for financials, HR, manufacturing operations, customer and sales management, and supply chain planning—capturing information across the business for analytics and reporting.

Plex plays a large role in helping manufacturing businesses make things—with close to 600 customers in 1800 facilities in 20+ countries, representing $35B in total annual revenue. The Plex Manufacturing Cloud processes over 5 billion transactions every day—and 50 percent of those transactions come from connected devices and machines on the shop floor. With an annual renewal rate of over 95 percent, these customers rely on Plex so that they can focus on their core business. In fact, over the past 12 months Plex has delivered 99.995 percent uptime to its customers—just 26 minutes of unplanned downtime.

Fourth Industrial Revolution

The Fourth Industrial Revolution (4IR) was a central theme at this year’s conference. Three market forces—cloud computing, the evolution of manufacturing and the rise of the Industrial Internet of Things (IIoT), are driving 4IR. By taking advantage of these forces, the Plex community has pioneered the concept of connected manufacturing—connected systems, processes, smart products, and people across the manufacturing floor—the foundation of 4IR. Plex customers are ahead of the curve, leveraging these trends to “make things” and to make things better and more efficiently. Here are several examples.

  • GenZe, a maker of electric bikes and scooters, is part of the rising sharing economy dedicated to providing affordable transportation across the US. The scooters are internet-enabled and connected to the Plex Cloud, allowing GenZe to collect information that can be analyzed to understand customer usage patterns and to provide monitoring of the scooter’s parts to support proactive maintenance.
  • Polamer Precision, winner of the 2017 Plex Impact Award for innovation, is using Microsoft HoloLens with Plex on the shop floor in a pilot program. The 3D technology is being used to illustrate the layout of the manufacturing floor—to see visually how forklifts move across the manufacturing floor—with the goal of automating that process.
  • Firstronic, a provider of electronics manufacturing services (EMS) and a 2016 Industry Week Best Plants winner leverages Plex’s cloud architecture to shorten the process of bringing on new plants from 12-18 months to fewer than 90 days. Plex provides the transparency to track material movement, and traceablity at the component level. The company will expand Plex usage to do more predictive planning, using Plex to identify which parts are running low and to order additional stock.


Skills Shortage

In my conversations with attendees at this year’s Plex conference, the skills shortage was a common frustration. Many businesses are using Plex to automate processes that were previously done manually, enabling some shop-floor activities to be accomplished with fewer people and allowing reallocation of resources throughout the company. Others are doing more on-the-job training with mobile devices and looking ahead to using augmented reality to instruct workers on the shop floor. While these are great solutions, they don’t really get to the heart of the problem. Our young people need to be educated in skills that will prepare them for the jobs that are available.

Mike Rowe, well-known TV host and podcaster, delivered the opening keynote on Day Two of the conference and provided his insight into the issue. Inspired by a trip he made to the San Francisco sewers to do an interview for Evening Magazine (a program on the local CBS affiliate), Mike traveled across 50 states and apprenticed at 300 “dirty” jobs, ranging from road kill cleaner to worm dung farmer.

After this experience, Mike had a “peripeteia” — reaching an unexpected reversal in his thinking about “dirty” jobs. With 5.6 million available jobs and 1.3 trillion in student loans funding education for jobs that don’t exist, the world needs to change their thinking.

As a result, Mike founded mikeroweWORKS Foundation, a public charity that is dedicated to providing the technical and vocational training required to fill the 3 million available skilled trade jobs currently available in the US—jobs that pay well but don’t necessarily require a college degree. During the conference, Plex presented Mike Rowe with a $25,000 donation which will be used to provide scholarships to people getting trained for high-demand skilled jobs.

Collaboration and Community

Finally, collaboration was an overarching theme at the conference. With over 13,000 active users in the Plex community worldwide, many rely on other users to work through issues or provide suggestions as to how Plex can be used in different and innovative ways to meet business goals. One third of the sessions at Power Plex were hosted by customers, sharing ideas and insight with the rest of the community.

Along with customers, the conference includes prospects and partners, and Plex prospects are encouraged to solicit product feedback from existing customers. Finally, PowerPlex 2017 marked the third year that the conference included a PowerPlex session devoted to women. This year’s luncheon included a cross-section of strong, articulate and successful women in executive positions who had great advice for women pursuing careers in technology and manufacturing.

Final Thoughts

Makers are truly making an impact, and based my many conversations at PowerPlex 2017, Plex customers are already defining the fourth industrial revolution and what it means to the future of manufacturing.

Posted in Uncategorized | Leave a comment

IBM Interconnect 2017: Cloud Solutions; Watson-enabled Applications and Blockchain

By Joe Clabby, Clabby Analytics

Last week approximately 20,000 people attended IBM’s Interconnect Cloud Computing event at the Mandalay Bay complex in Las Vegas, and their reasons for attendance varied widely. Some wanted to learn more about IBM’s cloud and cognitive computing products and strategies, others were interested in traditional enterprise computing products, still others in application development or asset management or security. Interconnect, with its broad spectrum of keynotes, product demonstrations, customer testimonials and hands-on labs, was able to address the broad range of requirements of its attendees.

I went to Interconnect 2017 without a specific research agenda. Interconnect is the industry’s largest cloud computing showcase, so I went to learn more about IBM’s cloud strategy, its cloud products, its new cloud services and to learn what vendors and customers are saying about the cloud marketplace. After listening to the presentations of several IBM executives, after talking with numerous product managers at exhibition booths, after listening to customer testimonial after customer testimonial, and after watching product demonstrations – I left Interconnect with several new perspectives.

My biggest finding is that IBM has done a masterful job building an enterprise-class hybrid cloud environment that features analytics, machine learning and cognitive computing. I stopped publishing reports on IBM’s cloud efforts back into 2014 because I perceived that IBM was having a difficult time getting its cloud act together (more on this later). What I found at this conference is that IBM has a clear cloud strategy with offerings that are distinctly different from the cloud offerings of Microsoft, Amazon and Google. And due to these differences, I believe that IBM will see stronger growth in enterprise cloud computing as compared with its leading competitors over the next several years.

I also noted that IBM has become more aggressive in building-out its Watson/third party software ecosystem.  For the past several years IBM has focused strongly on adding cognitive and analytics overlays onto its traditional enterprise software environments. But at Interconnect 2017 I found dozens of Watson-enabled third party software vendors – and learned that IBM’s Watson organization is looking to recruit thousands of third party software makers to its Watson/cloud environment.

Finally, I observed that IBM is “all in” when it comes to Blockchain technology – and it has the perfect platform on which to implement Blockchain: LinuxONE. Further, IBM offers a secure, IBM-managed Blockchain service environment that should appeal to customers who want to offload their transaction processing to a company with a deep transaction processing heritage.

IBM’s Hybrid Cloud

After several years of watching IBM try to architect its own cloud environment (Smart Cloud), and several years of watching IBM acquire a whole bunch of cloud-related companies, including Cast Iron, Coremetrics, Sterling Commerce, Unica, Emptoris, Varicent and Kenexa – with no apparent cohesive rhyme or reason for these acquisitions – I lost confidence that IBM would be able to recover against new generation cloud competitors such as Akamai, Rackspace, Amazon, Google, Microsoft and the telcos. I determined that I would come back to covering IBM cloud computing when IBM had a clear strategy, a coherent product line and rising marketshare potential.

I’m back.

At Interconnect 2017, Arvind Krishna, senior vice president, Hybrid Cloud and Director, IBM Research, presented a slide that brought me back into the fold (see Figure 1). This slide shows an integrated, secure, cohesive Watson/analytics cloud environment on which a variety of IBM software solutions can run. The design of this environment is distinctly different from the design of competing clouds that lack deep cognitive and analytics elements. IBM has moved its rich portfolio of home grown analytics, transaction processing and integrated software solutions to this unique, enterprise-strength cloud – contributing to and helping drive the company’s $24.5 billion in software revenue. In the shaded area (upper left), IBM has also developed several ancillary businesses that run on its hybrid cloud – new businesses with unique offerings that exploit cognitive computing and analytics, and that represent a huge growth opportunity for the company. Healthcare, financial services and the Internet of Things are examples of these high growth opportunities.  In short, the company now has a clear strategy; it has a coherent product line – and, given the uniqueness of its position in the cloud marketplace, the company’s cloud environment has strong potential to significantly grow market share.

IBM’s Chairman, President and CEO Ginni Rometty spoke about the solid revenue growth that she is seeing in both Watson and the IBM Cloud. She also described how the IBM cloud can be differentiated from the clouds of other vendors such as Microsoft, Google and Amazon. She emphasized that IBM is in the hybrid cloud business, helping customers build clouds that suit their enterprise needs without having to give away their data. According to Ginni, the value of data is that it generates unique insights. Competitors, she argued, are often in favor of democratizing and sharing data – and this gives away the advantage of developing unique competitive insights. IBM advocates helping enterprises gather data from a variety of sources, both internal and external, to help generate those unique insights. Ginni’s speech made it perfectly clear that IBM knows its position in the cloud marketplace; that it knows what its customers want; and that the company can now deliver solutions to its customers that its cloud competitors cannot.

Expanding the Watson/Cloud Ecosystem

When I attended the World of Watson conference last year I noted that a growing number of vendors were starting to overlay Watson on top of their traditional software offerings. The big news at this year’s Interconnect conference was that cloud leader Salesforce.com has now integrated Watson with its own Einstein program in order to simplify the use of its software and help users make better decisions.

At last year’s World of Watson conference, however, I left without a clear understanding of how IBM was going to make it possible for the broader ecosystem of third-party vendors to overlay Watson on top of their own infrastructure, management and application solutions. At Interconnect I learned that IBM is wide open to helping third-party software vendors, including competitors; integrate their solutions with Watson on the IBM cloud.

In short, third-party software makers who are not integrating analytics, cognitive services and cloud architecture into their product offerings are going to have a hard time surviving. IBM offers an integrated, secure cognitive/cloud environment on which third-party software makers can deploy their solutions. Vendors looking to the next wave in computing would be well served to look into deploying their solutions on IBM’s Watson Cloud.

Software makers who wish to know more about this topic should contact me at joeclabby@AOL.com.

Blockchain, LinuxONE and Blockchain for Hyperledger Fabric v1.0

In September, 2016, I started writing about Blockchain, a new way of processing transactions based upon sharing a common ledger (see this blog). I’ve since come to the conclusion that IBM’s LinuxONE platform, and its new Blockchain services are the best approaches to take when building Blockchain networks.

To build a Blockchain transaction processing environment, Blockchain users need secure platforms, networks and a ledger that accounts for transactions. For those not familiar with IBM’s LinuxONE, it is a platform that offers high-speed architecture, fast elliptic curve processing, and dedicated hardware for acceleration. It is 2.3 times faster than competitors – with integrated security levels such as EAL Level 5+ that no other architecture in the industry can match. Hardware security also includes FIPS 140-2 compliance. As part of the security features of this architecture, security keys are placed into protected memory and placed into hardware secure modules in order to prevent access and tampering. IBM does not make it possible for code to take control of secured modules, and focuses on encryption protection to prevent administrative abuse.   Try finding performance this strong and security this deep on any other platform…

At Interconnect 2017, IBM introduced a new Blockchain service, IBM’s Blockchain for Hyperledger Fabric v1.0 – a service that offers both security and scale, as well as related governance tools. This service, which uses LinuxONE servers on the back-end, helps protect from external as well as internal attacks, offers the levels of security previously described, uses the secure service containers and hardware security models previously described – all running on a highly auditable operating environment. I travel to parts of the world that are not familiar with LinuxOne architecture – where IT executives are reluctant to use it, preferring x86-based solutions – even though LinuxOne uses an operating environment that is quite familiar in the industry: Linux. For those geographies that want the best Blockchain implementation available, but don’t want to run their own Blockchain environment, I would strongly recommend evaluating the above mentioned Blockchain service.

One last thing on Blockchain: regulation. As transactions take place across various domains, they run into dozens or perhaps hundreds of regulations that must be addressed. Watson now offers a regulatory service that can help streamline the complicated problem of dealing with regulations. The Blockchain story – LinuxONE combined with automated Watson regulation services – will be difficult for other vendors to mimic and overcome.

Summary Observations

Over the past few years, I’ve watched the giant behemoth called IBM change its course. It has changed its cloud strategy – it no longer tries to compete with the public cloud vendors or the traditional enterprise cloud makers per se – it now offers a solution that is completely different. IBM calls its offering a “hybrid cloud” – a blend of integrated public and private cloud facilities driven by analytics and cognitive computing. The company has taken its traditional enterprise management tools and cloud-enabled them – weaving them into the fabric of the cloud. In fact, wherever it makes sense, it has cloud- and Watson-enabled its entire product line, from management tools and infrastructure through applications, reporting tools and databases. And it has modified its pricing and go to market models – actually promoting many of its offerings first and foremost as “software-as-a-service” solutions.

These changes at IBM involved huge internal cultural changes, including migrating the “enterprise computing old guard” to “new disciples of analytics and cognitive computing.” These changes involved “new think” – like how do we most efficiently develop and support applications under the new, analytics/machine-driven cloud model? And how do we offer traditional, on-premise solutions as well as Software-as-a-service solutions? They involved retooling, integration efforts, the introduction of new APIs to join the old world to the new mobile world – and the new cloud environment to the existing premise environment. IBM has worked to integrate existing traditional management applications into the new model of cloud computing – and to find or create new solutions more suited for the new hybrid cloud environments.

The IBM Cloud has come a long way since I stopped covering it in 2014. This week, at Interconnect 2017, I saw firm evidence that IBM has made it over the hump – and has transitioned to a modern, extremely competitive analytics/cognitive cloud company that can address enterprise needs for a blended, rationalized public/private/hybrid computing environment. Enterprises planning for a future of cognitively-enabled applications and secure Blockchain must take a closer look at the IBM Cloud.

Posted in Uncategorized | Leave a comment

IBM Connect 2017: Collaboration Under the Radar

By Joe Clabby, Clabby Analytics

IBM runs six large technology events each year: World of Watson, InterConnect, Amplify, Edge, Vision and Connect. I usually attend the two largest, World of Watson and InterConnect, along with 15,000 to 20,000 other interested parties. Both of these events are huge and cover a wide range of technologies ranging from systems to infrastructure, to management, to cloud, to cognitive computing, to analytics and more. Although I suspect that IBM’s deep portfolio of collaboration products can be found at some of these events, I just plain don’t notice them.

My earliest memories of IBM date back to the 1970s when I was selling competitive products against their word processing and computer systems. In the 1980s, IBM collaborative computing “office products” started making the scene. Does anyone remember PROFS and DISOSS? In the 1990s IBM’s stunned the marketplace by purchasing an office product company by the name of Lotus for a whopping $3 billion. As a research analyst at the time, I could not fathom what IBM saw in Lotus and its email and collaboration products. By the early 2000s, I could clearly see the value of the purchase as IBM brought to market a myriad of new office and document management offerings, creating a multibillion-dollar collaboration software environment that dwarfed their now “miniscule” investment in Lotus.

This year I chose to attend IBM’s Connect 2017 event in San Francisco – leaving my 75° abode in in sunny Charleston, South Carolina, to travel to cold and rainy San Francisco. And as surprising as this may sound, I’m glad I did. Why so?

  • IBM Connect 2017 was an industry tradeshow built around IBM’s and its business partners’ collaborative solutions. It offers roadmaps and commitments to older IBM products, such as Sametime, Domino and Notes. It also showcases new IBM products such as Verse (a versatile modern day mail and messaging environment with a collaborative overlay), and Connection (a social business network platform), and  Watson Workspace and Work Services (a conversational collaboration environment and a set of APIs that allow developers to build applications that understand conversations and intent, and allow for integration into existing work applications).
  • Business partner participation at the event was strong, featuring new collaborative product offerings between IBM and Cisco and between IBM and Box . Connect also highlighted many blended solutions by vendors whose products overlay IBM offering, such as project management environments blended with Notes, and mobile interfaces blended with Domino environments.   Also featured were a slew of new products that integrate Watson cognitive technologies with existing business applications, such as Watson Workplace into traditional software offerings, thus delivering new functionalities to market.

What I Heard and Saw: Business Partners

I always seem to gravitate toward the EXPO floors at these events. I think the main reason is that I just plain love to play with technology – I like to see the way it’s used; I like to see what new and innovative directions that developers have taken with their hardware and software solutions; and the EXPO floors are a great place to talk with vendors about what’s really happening in the marketplace, as well as to meet IT buyers.

At Connect, I had a long conversation with Alex Homsi, the CEO of Trilog Group, an IBM business partner, who I asked to help put IBM Connect 2017 into context. The way Mr. Homsi described it, the IBM Connect events are all about getting things done. They are about increasing productivity and efficiency, but mostly about collaboration in the processing of complex workflows. As I stood in the Trilog booth, Mr. Homsi gestured around the floor “look over there – you see companies that offer telepresence, that create virtual project rooms, that offer sales collaboration tools and much, much more.” When I prodded him for more information about why people come to the event, Mr. Homsi told me that “They come to solve complex, mega-problems” and then he proceeded to talk about how his own project solutions help customers save millions (or in once case, hundreds of millions) of dollars by digitally capturing content, effectively communicating it and then coordinating the efforts of large groups. Incidentally, Mr. Homsi is also CEO of a company called Darwino, a company that helps customers mobilize IBM Notes Domino applications and migrate to the cloud.

I also talked at length with the company by the name of Oblong, a maker of a digital content management environment that I wish I owned. For sci-fi fans who may have seen the movie “Minority Report”, you might recall Tom Cruise pulling files and data streams from a wide variety of sources, which he examined in the holographic 3D air space that surrounded him. He could expand files, shrink files, push files to the side, look at multiple displays of realtime and recorded data in real time – easily moving between static and dynamic filing environments at the touch of his hand. Oblong makes a highly scalable environment that can scale across hundreds of displays (sorry, no holographs yet) where information that it collects can be collaboratively shared amongst large teams. I pictured disaster response use cases where a room of people look to coordinate a response to an event – and I have to admit, I thought back to the NASA launch room in the movie “Apollo 13” where a group of scientists collaborated on a way to help astronauts return to earth after a failed lunar landing (I guess I’ve been watching too many old movies lately…). Anyway, you get the point, the world is now digital, in these digital sources can be easily harvested and displayed – enabling people to more easily collaborate and make better decisions.

A company by the name of Imaging Systems, Inc. out of Slovenia also caught my attention with a product called IMiS MOBILE that can be used to easily mobile-enable legacy applications – thus broadening the platforms that can be used to conduct collaborative activities using hand-held devices. I liked this product because of its programming simplicity.

I went to at least a dozen other booths, including the IBM Verse and Connection booths (covered in the next section).

Last but not least, business partners were truly excited about using Watson to make their applications “smarter”. I wrote about this trend in my Pund-IT trip summary after attending World of Watson 2016 – and this trend is becoming omnipresent (I’m seeing it everywhere across the traditional software applications markets) as ISVs are recognizing that using Watson can simplify the use their products while expanding the types of and accuracy of the solutions they create. Watching the industry move to “Watson Enablement” is truly one of the most fascinating trends I’ve ever seen in the computing industry – it seems that the sky’s the limit in terms of what machine intelligence blended with traditional applications can now do.

What I Heard and Saw From IBM

The lead speaker at Connect was Inhi Cho Suh, IBM’s General Manager of Collaborative Solutions. She took the stage to tell the audience of about 2,000 Connect attendees what was going on in the collaboration marketplace and what IBM was doing to address the needs of the market.   is– I’ve known Ms. Suh for years – I first met her in the early days of IBM’s foray into the analytics marketplace – and she’s a real straight shooter. Bringing a person with her background to the collaboration space is a very smart move on IBM’s behalf – she knows analytics technology extremely well, and she knows how to Watson-enable IBM’s collaborative offerings as well as how to help business partners to do so.

Ms. Suh took the stage to tell the audience about the trends that she is seeing in the collaborative computing market space. She contended that the way we engage with others in the work world is changing thanks to new innovations, especially the use of Watson cognitive services that are being used to simplify products and extend their capabilities. She also talked about open collaboration where companies in the collaboration marketplace are working more closely together to build jointly integrated solutions. I saw a clear example of this with the joint IBM/Cisco announcement that integrates Cisco collaboration products with IBM collaboration products – in the past these two companies would’ve been strong competitors who likely would not have worked together, but now both are pleased to show how the best of each company’s solutions can be blended to create a more powerful and integrated collaborative environment.

Ms. Suh also talked about how collaboration products are getting better at streamlining process flows.

Other speakers talked about how cognitive computing is being blended with analytics. In short, cognitive computing is being used to help sort and prioritize “what is important to me”; it is streamlining the flow of work; it’s using bots and virtual assistants to help aid humans in their decision-making processes; it’s involved in using the Internet of Things sensory devices to aid decision-making; and it’s helping individuals focus better.

There was also an interesting discussion about whether today’s tools “create more noise” than help. From what I saw on the demo floor, today’s tools can take a lot of the uncertainty and human guesswork out decision-making, while at the same time making processes flow more easily. The “more noise” argument does not hold with me.

As for IBM products, I got a close look at IBM Connections and IBM Verse. Verse was pretty cool, a modern mail-and-messaging environment with the collaborative overlay that made it simple to access and sort the work of fellow team members as well as handle external inputs. I especially liked that the product could be used in a mode where you don’t have to delete your emails and related documents, you just leave them in your inbox when you’re done with them – and if you need to refer back to them you perform a search on the few keywords that you may remember and your document appears. I’ve deleted emails and messages for decades – what a silly concept… As for IBM Connections, I rarely have a need to collaborate with a group of people to conduct project work, but if I did, I’d consider using this social, collaborative environment.

As an Aside

I had the opportunity to attend a customer-presented session on deploying Blockchain on IBM’s Bluemix. For those not aware of Blockchain, it’s a new way of processing transactions based upon Bitcoin technology. It deals with distributed databases, distributed servers, encrypted data, the mining of that data – and establishes consensus between nodes. This technology is being used to create trusted, synchronized transactions. If a transaction is tampered with in any way, all stakeholders know about it and the transaction is thus broken. The ability to conduct secure transactions is one of the major points of this technology – but probably the biggest selling point is that it takes intermediaries who add processing, time and cost overhead to transactions out of the picture. It’s pretty exciting stuff and represents a whole new, more efficient and secure way of processing transactions – and I will have the pleasure of speaking about it at a government conference in Dubai in May. It was fun to see how another practitioner handled the topic.


I truly enjoy going to technology shows. Throughout my lifetime, I have been a technology sales representative, a project manager and a technology researcher. I remember how I used to do things in the old days, and I like seeing how technological advances have simplified tasks, and have made people more productive and efficient.

As I looked at the wide array of technological solutions presented at Connect 2017, I kept asking myself “How can I use these tools as a research analyst to my advantage?” I left Connect with some new ideas to explore, and with the self-realization that I go to events such as this to learn, to look for new innovations, to talk with people who have used or developed these technologies – and most of all to surround myself with people with like interests who enjoy technology and innovation as much as I do.

Will I go again to IBM’s Connect next year? I hope so…




Posted in Uncategorized | Leave a comment

IBM Enhances Spectrum Storage Management, Integration and Flexibility

IBM’s February 7, 2017 software-defined storage announcement was so chock full of new capabilities for the company’s Spectrum Storage and Cloud Object Storage, it’s tough to sum it up in one sentence or even a paragraph. But a look at the history of IBM Spectrum Storage will provide some context and illustrate IBM’s prime objectives: consistency, integration and flexibility.

The Spectrum Family of software- defined storage was first announced in February 2015 — a rebranding of existing IBM storage solutions with names more indicative of their functions. In early 2016, IBM announced the IBM Spectrum Storage Suite, a single capacity-based license that includes all the IBM Spectrum Storage offerings. Over time, the suite has become more of a “family”, with a consistent user experience across products by using IBM Storage Design Language (based on IBM design language) and improved integration between members of the product family.

For example, Spectrum Scale and Spectrum Accelerate are managed through Spectrum Control, and Spectrum Scale can be used as a target with Spectrum Protect – an example of integration between these two products. Another great example is the integration of IBM Cloud Object Storage (formerly Cleversafe –acquired in 2015) with IBM Spectrum Storage offerings (we’ll discuss this in greater detail later). IBM Cloud Object Storage is now a full-fledged member of the Spectrum Storage Suite with no change in the license charge.

In most cases, IBM Spectrum Storage solutions are available in flexible deployment models as software-only and as-a-service offerings delivered via a hybrid cloud model—including on-premises solutions, private and public cloud implementations. Data can be easily shared across on-premises and cloud deployment models—providing maximum flexibility for customers.

IBM Spectrum Storage Suite Announcements

IBM announced enhancements to IBM Spectrum Virtualize, IBM Spectrum Control/Storage Insights, IBM Spectrum Accelerate and IBM Cloud Object Storage. Here are the details.

IBM Spectrum Virtualize

The big news with IBM Spectrum Virtualize is that it is now available as software-only and can be deployed on x86 servers , providing more options for service providers who may also choose to offer Spectrum Control for management and Spectrum Protect for back-up. Bare metal support on Supermicro SuperServer 2028U-TRTP+ is being introduced (in addition to existing support for Lenovo System x3650 M5). As a result, service providers can offer new services to their customers (for example disaster recovery as a service for clients with IBM’s SVC and/or Storwize) while also adding new capabilities to their own infrastructure.

Other enhancements include I/O throttling for host clusters, GUI support for host groups, increased background FlashCopy transfer rates (4x improvement), automatic resizing of remote mirror volumes, and automatic consistency protection for metro/global mirror relationships.

IBM Spectrum Control and Storage Insights (SaaS version)

Perhaps one the most exciting enhancements to Storage Insights is support for Dell EMC storage (VNX, VNXe and VMAX) enabling cloud-based storage analytics for Dell EMC customers, providing better insight, flexibility and efficiency for customers with heterogeneous storage environments  For example, administrators can see at-a-glance which storage volumes need to re-tiered for cost-optimized storage, examine future capacity requirements and reallocate unused storage across both IBM and Dell EMC storage infrastructure. A new consolidated chargeback reporting tool provides better accounting for data owners and administrators and adds support for Dell EMC VNXe.

IBM Spectrum Accelerate

The latest revision of IBM Spectrum Accelerate (V11.5.4) provides the following:

  • Non –disruptive data-at-rest encryption – data security is improved by using SED drive-based encryption and can be done on –premise or in the cloud, Standard key management tools including IBM Security Key Lifecycle Manager and SafeNet KeySecure are supported and the functionality is available at no additional charge.
  • Hyper-Scale Manager 5.1 enables centralized management across up to 144 systems on or off-premise running IBM Spectrum Accelerate (FlashSystem A900/A900R, XIV Gen 3, Spectrum Accelerate software-only).
  • VMware 6 support enables enhanced IBM Spectrum Accelerate solutions in hyperconverged configurations including VMware ESXi 6.0 and vCenter6.0. vSphere Web Client certification provides IBM native storage visibility. These features are delivered in IBM Spectrum Control Base Edition for centralized management across the entire IBM storage portfolio.

IBM Cloud Object Storage

As mentioned earlier, IBM Cloud Object Storage is now included in the IBM Spectrum Storage portfolio, having been integrated with:

  • IBM Spectrum Scale (for an additional storage tier behind IBM Spectrum Scale);
  • IBM Spectrum Protect ( to provide an electronic vault, initial target or storage in the cloud);
  • IBM Spectrum Control (to enable monitoring of IBM Cloud Object Storage); and
  • IBM Spectrum CDM (for data archiving in IBM Cloud Object Storage).

Supported environments include on-premise single and multi-site, hybrid cloud dedicated multi-site, and cloud-only running on dedicated supported cloud infrastructure. Existing clients get the new capabilities and can use them immediately.

Other new features include:

  • Unified NFS/Object access adds the ability to use the object store via an NFS file system interface. IBM will use its own NFS native file access capability for object storage to enable support for applications that communicate to storage using file system protocols in addition to those that speak natively to object storage – with all the scalability, security, reliability and capacity characteristics (hundreds of PBs and billions of files) of IBM Cloud Object Storage. Included is the ability to interoperate between the file system interface and the object storage RESTful API, a common storage pool for NFS and object data, and file to object migration. Customers can, for example, use the file system data to ingest the data set, but for processing that data for an analytics application they can access data via the object interface. IBM Object Cloud Storage NFS is ideal for large-scale file archives, long-term backup retention, and file accessible content repository applications.
  • Enhanced storage-as-a-service (STaaS) increases the maximum number of user accounts that can be supported on a single system. The system can scale to millions of users with their own accounts and isolated storage containers all on a single system. This system is designed to be used both for a commercial STaaS product that can be sold to other businesses but also for employees within an organization with use cases including medical archive-as-a-service, archive/back-up as-a-service, IT storage as-a-service and cloud storage as-a-service.
  • IPv6 is now supported for management of all devices and all nodes— nodes can be configured to support IPv4 or IPv6, so IBM Cloud Object storage can easily fit into a customer’s existing network architecture if they have adopted IPv6.
  • Preconfigured bundles of IBM Object Cloud Storage are now available in Entry, Workhorse and Enterprise classes. These bundles are turnkey, simple to order complete systems with a balanced, pre-tuned set of components for various levels of capacity and performance— designed to simplify a customer’s entry into object storage.

Summary Observations

Having followed IBM Storage for several years, the direction the company is taking with its storage portfolio is definitely making my job easier. Back in the old days, there were so many disparate hardware and software offerings handled by different siloed groups at IBM Storage, it was difficult to keep up with them all. And it was even tougher to figure out how all the products worked together. In some cases, I learned that it was tough to figure out because they didn’t.

Since the rebranding in early 2015, IBM has been making steady progress in integrating its storage portfolio at both the marketing and technical level. A consistent user experience makes the products feel like a family and makes storage administration easier and more efficient. Synergies between the products enable enterprise customers and service providers to offer new services. IBM’s software-defined Spectrum Storage strategy gives customers more flexibility, enabling Spectrum products to be purchased as software-only, as an integrated solution or as-a-service. In addition, IBM supports a wide range of underlying hardware platforms and a hybrid cloud storage model.

These latest enhancements to the portfolio continue to build on the themes of integration, flexibility, and consistent management— which ultimately improve efficiency and drive down storage costs, while at the same time enabling new use cases. I am particularly impressed by how quickly Cleversafe, now IBM Cloud Object storage, has been integrated with IBM Spectrum Storage and enhanced to support unified file and object access. IBM Spectrum Storage has something for everyone— and now it’s easier than ever to find the integrated software-defined storage solution that your organization needs.

Posted in Uncategorized | Leave a comment

IBM DS8880F All-Flash Data Systems – Integral to IBM’s Cognitive Computing Strategy

By Jane Clabby, Clabby Analytics

When we think of data storage today, chances are that the first things that come to mind are “object storage”, “cloud storage”, “commodity storage” and “software-defined storage”. Nobody really talks much about SANs and block storage anymore–except for high-end database and transaction oriented applications. The same has been true for IBM’s storage strategy.

For the past couple of years, IBM has made many announcements and published many marketing documents that position its Storwize and Spectrum portfolios, including “all-flash” versions of those products. However, the DS8000 family is typically relegated to the background, and enhancements have focused primarily on “speeds and feeds” – x times faster, x more capacity, x times lower latency etc. – rather than new use cases and support for different types of workloads.

Imagine my surprise when IBM’s first storage briefing of 2017 highlighted not only the new all-flash DS8880F arrays, but also the strategic nature of the DS8880F family and how it underpins IBM’s cognitive computing and analytics offerings. In addition, this announcement provides a lower-cost entry point (starting at 95,000 USD) for DS8880-class storage, providing high-performance low-latency storage for mid-range customers. According to IBM, providing a more cost-effective offering enables businesses in emerging geographies such as Latin America and Eastern Europe to afford DS8880F benefits in a smaller package.

New use cases/workloads

Here are a couple examples of workloads that are ideal for the DS8880 family:

·         Cognitive computing

IDC forecasts global spending on cognitive systems will reach nearly $31.3 billion in 2019 with a five year compound annual growth rate (CAGR) of 55%. With IBM Watson, IBM has been a pioneer in cognitive computing –using Watson-based artificial intelligence technology in applications across a wide range of industries, enabling humans to solve complex problems more efficiently. Cognitive applications ingest, process and correlate huge volumes of data which requires a robust computing infrastructure and high performance, high capacity storage.

·         Real-time analytics

Another area where IBM is seeing interest from customers is in real-time analytics, where information can be collected, processed, and analyzed in real-time for applications such as credit card fraud detection and Internet of Things (IoT). The DS8880F provides the robust resiliency and data security for the source OLTP data set as well as the throughput and performance to enable real-time analytics. As a result, financial institutions can match credit card use patterns with real-time data collection, detecting potential fraud before it happens and saving millions of dollars. In transportation, IoT data collected from cars, street lights, cell phones and other sources combined with real-time analytics will enable driverless cars.   In health care, patients can be monitored–collecting and analyzing several metrics simultaneously in real-time– to alert medical staff to potential problems, improving patient outcomes.

New generation analytics and cognitive applications simply can’t be handled by traditional shared nothing Hadoop clusters, HDFS, and commodity storage. Those systems are designed for scale and low-cost, but don’t have the bandwidth or performance to support real-time analytics and cognitive systems that businesses collect and need to analyze data instantaneously. The new DS8880F data systems are designed specifically to handle these types of workloads that require lower latency and sub-millisecond response time.

Workload consolidation

The new DS8880F models also enable the consolidation of all mission-critical workloads, both new and traditional, for IBM z Systems and IBM Power Systems under a single all-flash storage family.

  • Cognitive – Including Watson Explorer, Watson Content analytics and Watson APIs that allow customers and ISV’s to create their own Watson- based applications.
  • Analytics- Including IBM Cognos Analytics, IBM SPSS, IBM Infosphere Big Insights, SAS business intelligence, Elasticsearch, Apache Solr and others.
  • Traditional/Database – Including IBM DB2, Oracle, SAP, PostgreSQL, MongoDB, Cassandra and others.

DS8880F all-flash data systems –A closer look

IBM’s new line-up of DS8880F data systems includes three all-flash arrays designed to support applications from entry level business class to enterprise class to the “analytics” class required for real-time analytics and cognitive workloads. IBM uses a different approach in the design of its all-flash arrays. Rather than just replacing HDD’s with Flash cards IBM has completely rearchitected the system, optimizing it for Flash for better performance. The systems use the Flash Enclosure Gen2 with 2.5” drives in a 4U enclosure and are built on the Power CEC’s (Central Electronic Complex) used in Power servers. The models vary based on number of cores, capacity, cache and number of FibreChannel /FICON ports, but the functionality is the same across the product line.

Here are the details:

  • DS8884F Business Class
    • Built with IBM Power Systems S822
    • 6-core POWER8 processor per S822
    • 256 GB Cache (DRAM)
    • 32 Fibre channel/FICON ports
    • 6.4TB to 154 TB of flash capacity
  • DS8886F Enterprise Class
    • Built with IBM Power Systems S824
    • 24-core POWER8 processor per S824
    • 2 TB Cache (DRAM)
    • 128 Fibre channel/FICON ports
    • 6.4TB to 614.4 TB of flash capacity
  • DS8888F Analytics Class
    • Built with IBM Power Systems E850
    • 48-core POWER8 processor per E850
    • 2 TB Cache (DRAM)
    • 128 Fibre channel/FICON ports
    • 6.4TB to 1.22 PB of flash capacity

The Flash Enclosure Gen 2 provides significant improvements over the previous generation for both IOPs (read: 500,000+47%, write: 300,000+50%) and throughput (read: 14 GB/s + 268%, write: 10.5GB/s +288%). Aside from the performance gains seen from using all-flash, the unique architectural design attaches the enclosure to the PCI 3 bus rather than through the device adapter, eliminating a portion of the data path which lowers latency, and improves response time. The drawer itself has also been redesigned to provide higher bandwidth into the I/O drawer and to use ASIC rather than FPGA in the drawer itself (also improving bandwidth).

Other features of the DS8880F series include greater than “six-nines” availability; point-in-time copy functions with IBM FlashCopy; and Remote Mirror and Copy functions with Metro Mirror, Global Copy, Global Mirror, Metro/Global Mirror, IBM z/OS/Global Mirror, and z/OS Metro/Global Mirror providing disaster recovery/replication for up to 4 sites and 1000 miles with 3-5 second RPO (recovery point objective) for businesses where continuous operation is critical.

Summary observations

With this announcement, IBM expands its offerings of high-end enterprise-class storage by adding a new lower cost entry point that makes the systems affordable for a broader range of customers, while also enabling new uses cases that require the performance of all-flash. The optimized architecture further reduces latency and response time which will differentiate IBM in markets that require real-time analysis and results. Not only will these systems improve performance in their traditional database and OLTP realm, but the DS8880F family will become a key building block in infrastructure built to support cognitive and analytics applications.

In order to attract new customers, IBM also plans to expand marketing outreach by promoting the DS8880 systems as part of its broader cognitive computing strategy. The recent analyst briefing was a good start. Like IBM, I have been guilty of shunting aside the DS8000 family in favor of covering IBM Storwize, Cleversafe object storage and the IBM Spectrum software-defined family. In the future, I will pay closer attention. And IBM customers and prospects – you should, too.

Posted in Uncategorized | Leave a comment

Year End Review: IBM Systems Group

By Joe Clabby, Clabby Analytics

Tom Rosamilia, Senior Vice President of IBM Systems group, recently presented his annual year-end recap. In his opening statement he started with a review of IBM’s mission in the IT marketplace; to “continue to transform and innovate in technology, business models, and skills for the cognitive era.” He then pointed out several of IBM acquisitions that illustrate the company’s focus in these areas, including: Truven Health Analytics, explorsys, Merge, clearleap, USTREAM, The Weather Company and more.

What jumped out at me was that most of the new solutions that IBM has brought to market are focused on specific industries, like healthcare, or are focused on helping other vendors bring analytics solutions to the market, such as The Weather Company. Although IBM continues to deliver the infrastructure, database and transaction products it always has, it must be noted that the company truly sees itself as a leader in analytics and cognitive solutions.

IBM Systems’ accelerated 2016

Rosamilia underscored this point with a review of the key products in his portfolio that support the company’s emphasis on analytics. These included the activities that have taken place around IBM’s POWER architecture, its z Systems/LinuxONE offering, and in storage. Coincidently, in 2016, Clabby Analytics wrote in-depth reports on each of these topics, including this POWER 9 report, this LinuxONE report and this storage/software defined report.

Rosamilia’s POWER commentary highlighted innovations around the company’s OpenPOWER initiative (which we cover in-depth in this blog), as well as IBM’s emphasis on extraordinarily fast Power-based servers that leverage other processor architectures, such as NVIDIA graphical processing units (GPU) to serve the analytics marketplace.

He focused on Google and Rackspace’s efforts to develop an open server specification using IBM’s forthcoming POWER9 architecture, then described 2016’s arrival of POWER LC servers that combine POWER8 processors with NVIDIA’s GPUs and NVLink interconnect technologies. Rosamillia also spent time discussing IBM open sourcing its CAPI (Coherent Accelerator Processor Interface) technology and the progress being made within the OpenCAPI Consortium; and the company’s continuing efforts with NVIDIA to deliver Power AI solutions.

My key take-aways from this discussion was that IBM is continuing to aggressively build “accelerated systems” that use multiple types of processors to accelerate analytics workloads, and that IBM is successfully engaging open communities to help build solutions on its POWER architecture and complementary technologies.

Z Systems, LinuxONE and Blockchain

The key points that Rosamilia chose to focus on regarding the company’s z Systems/LinuxONE mainframe architecture centered on positioning LinuxONE for hybrid cloud environments; the use of z13s for encrypted hybrid clouds; the relationship of the z/OS operating environment and the Apache Spark movement (a better way of processing large volumes of analytics data than Hadoop); the EZSource acquisition (for code analysis) ; and the availability of secure cloud services for Blockchain on LinuxONE.

The new “news” in Rosamilia’s review of z System/LinuxONE was his emphasis on Blockchain and HSBN (the company’s “high security business network”). Blockchain serves as the basis for creating a new way to perform transaction processing, one that features a secure “open ledger” that is shared amongst all concerned parties during the transaction. This new approach streamlines transaction and business processes and enables significantly greater security that traditional approaches.

I had not been aware that IBM had created a service offering featuring IBM LinuxONE servers overlaid with Hyperledger that enables customers to form smart contracts, to create shared ledger’s, to gather consensus along the route of completing a transaction – all taking place in a secure and private environment. IBM claims that it is making solid headway with this offering in the securities, trade, finance, syndicated loans, supply chain, retail banking, public records and digital property management industries. Rosamilia shared examples of success stories using this LinuxONE/Bluemix offering, including the activities taking place at Wells Fargo, Walmart and Everledger.

IBM Storage in a Flash

In storage, Rosamilia focused on2016 activities that resulted in IBM Flash and software defined solutions. He described efforts to round out the company’s Flash array offerings from the low-end all the way through the high-end, and also described how the company is providing storage solutions driven by software and appliance designs, along with IBM’s storage as a service cloud offerings.

Rosamillia also provided examples of how IBM’s software-defined storage products are being used, including a discussion of DESY (the Deutsches Elektronen-Synchrotron research facility in Germany which is using IBM’s Spectrum Scale to analyze petabytes of data in minutes, not days), the Arizona State Land Department (super-efficient land administration using IBM Flash Systems), and bitly (using IBM Cloud Object Storage to accelerate consumer research with faster, easier access to data derived from capturing over 10 billion online clicks per month).

Summary Observations

Rosamilia’s year-end review of IBM Systems’ highlights was a good 50,000 foot overview of the most important activities that have taken place in 2016. But there is far more going on within this group than meets the eye.

Two years ago, IBM’s POWER organization was struggling: its former UNIX market stronghold was weakening as customers shifted to Linux on x86 architecture and revenues were in strong decline. To right the ship, IBM decided to open source its POWER architecture to the industry. And, as a result, the company has revived its revenue stream while fostering advanced and innovative systems from the OpenPOWER community. What IBM’s POWER organization has done is truly remarkable, they rescued this architecture from the declines suffered by competitors, including Oracle (Sun) and HPE, opened it up for collaborative systems integration, and built incredibly powerful new system designs using POWER processors, GPUs and FPGAs (field programmable gate arrays).

For over 20 years, ever since industry pundits in the mid-1990s forecast the demise of the IBM mainframe, Clabby Analytics has taken the position that there is no other architecture better suited for processing secure transactions (and now in-transaction analytics workloads) than IBM’s z System. Given this position, we see IBM’s new LinuxONE mainframe servers as ideally positioned to support a projected major market move toward Hyperledger and Blockchain transaction processing over the coming years. This movement should greatly escalate the sale of mainframe servers. Long live the mainframe!

As for storage, the markets that IBM and every other enterprise vendor focus on have changed tremendously over the past few years as customers shifted from traditional workloads to include more compute- and data-intensive workloads (genomics, simulations, what/if analysis, cognitive), and next generation Big Data and born-in-the cloud applications, like Hadoop, NoSQL, Spark and Docker. Accordingly, IT executives are now looking for storage and software defined infrastructure options that provide better IT performance, scalability, and agility at significantly lower cost. In addition, these same executives are grappling with rationalizing which workloads belong on-premise – and which workloads can be shifted to low-cost public cloud storage.

To address traditional storage requirements, as well as the new generation of compute- and/or data-intensive applications, IBM has revamped its storage line to include a complete range of solutions (including software-based offerings, services and storage hardware options such as appliances and all-Flash arrays). To many, IBM’s myriad storage offerings may seem confusing but if you look from the perspective of IT managers and executives, storage needs to make full use of varying technologies , needs to accommodate private and public clouds; and needs to support both traditional applications and new workloads, including analytics. IBM storage accomplishes all of these objectives.

IBM’s year-end review was excellent at a high level. But for more details on each initiative, take a look at our free reports on IBM LinuxONE, IBM storage and software defined storage, and POWER architecture at www.ClabbyAnalytics.com.

Posted in Uncategorized | 1 Comment