Get the lowest-cost and the best server colocation service in the business. Learn more.
Information Technology News.

IDC: high performance data analysis looking better for 2014

Share on Twitter.

Install your server in Sun Hosting's modern colocation center in Montreal. Get all the details by clicking here.

Do it right this time. Click here and we will take good care of you!

Click here to order our special clearance dedicated servers.

Get the most reliable SMTP service for your business. You wished you got it sooner!

June 19, 2014

IDC recently said that 2014's first quarter was a downer for high-end storage sales, but now it also predicts that things are looking up in storage used for what it now calls high performance data analysis (HPDA).

IDC's vice president for HPC and data analysis Steve Conway places HDPA in perspective with a quote explaining that it's the kind of hardware one would buy for a high performance server, but applied to analytics.

“To be sure, leading commercial companies in a variety of industries are turning to HPC technologies for challenging big data analytics workloads that enterprise IT technology alone cannot tackle effectively,” he writes.

“HPC systems can handle more complex queries, more variables, and faster turnaround requirements,” he added.

Throw in the fact that the likes of Hadoop emerged from the HPC-inspired rigs at hyperscale web operators and just what HDPA involves become even clearer.

IDC says that the field is also going to be a big business. Compounded annual growth rates of 23.5 percent will result in US $4.3 billion of HDPA technology each year, IDC added.

About $2.7 billion of that will be for servers and a further $1.7 billion will be storage. A companion study, Worldwide High-Performance Data Analysis Storage 2014–2018 Forecast, says storage sales will increase at an even faster pace, with 2013's HDPA storage sales of $514 million to grow at 26.5 percent to reach 2018's predicted total.

The analyst firm says this HDPA initiative is its first formal effort in the field, but that it plans to do it more often in near future.

In other IT news

NetApp said earlier this morning that it has completed the overhauling of its unified storage FAS arrays with FAS-2500s at the low-end and a bigger FAS-8080 EX on the high end of the range.

We got the basic details, except for the FAS 8080 which has 36 TB of virtual storage tier flash and not the 18 TB we initially expected.

The new solution can scale up to 69 PB through clustering and handle 600+ IO connections.

Just a quick point about the scale-out-- the FAS-8080 EX can scale to 24 nodes (12 high-availability pairs) with file access but only 8 nodes (4 HA pairs) with SAN access.

The separate V-Series goes away as a hardware implementation, with its software now incorporated in Data OnTap.

At the low-end, the FAS-2220 gets replaced with the FAS-2520, the FASA-2240 with the FAS-2552 and the FAS-3220 with the FAS-2554, completing the removal of the mid-range FAS-3200s, the FAS-3250 having been replaced by the FAS-8020 a few weeks ago.

NetApp says the FAS-2500s, with the 4 TB of flash, can accelerate workloads by up to 46 percent and increase usable capacity by 48 percent.

The entire FAS hardware range has been refreshed in just five months and now characterized as hybrid flash+disk or, with the FAS-8000s, all-flash, if users configure them that way.

The FAS-8080 EX can perform at around 4 million IOPS and more than 4.6 PB of capacity if configured with only flash drives.

As a hybrid array, its flash cache amounts to almost 500 PB. There is a NetApp All-Flash FAS Reference Architecture for VDI.

The company says all-flash FAS arrays are suited to workloads needing both high-performance and low data access latency.

Its OnCommand Insight management software enables storage cost management, rationalisation, and optimization of storage services and there is a tailored offering, NetApp Services for OnCommand Insight, for enterprise customers deploying high-end FAS-8000 systems across heterogeneous datacentres.

To encourage reluctant customers, NetApp is offering a 90 day payment holiday to save while transitioning to these new FAS arrays.

The company is now embracing the cloud storage world with Cloud ONTAP and also beginning to strengthen its object offering.

As soon as an attractive new data storage silo comes along, a 3rd-party array virtualization function, and object storage NetApp either adds functionality to FAS or makes a targeted acquisition.

To be sure, let's just say that FlashRay is a departure from this as it's a complete in-house development of an all-flash array and not, as with Dell, HDS and HP all-flash arrays a facility inside a current array and inheriting all of its data management features.

In that respect, FlashRay will be unique, as NetApp's other all-flash array competitors either bought technology in Cisco, EMC and NetApp or are startups (Kaminario, Nimbus Data, Pure Storage, Skyera, SolidFire and Violin Memory).

In other IT news

Flash storage solutions provider SanDisk said this morning that it is acquiring Fusion-io for $1.1 billion in an all cash transaction.

Fusion-io’s PCIe flash card and ioControl shared flash array businesses will now be integrated with SanDisk’s other enterprise flash acquisitions and business.

This acquisition will provide SanDisk an immediate and high-profile PCIe flash card hardware and software product line, almost overnight.

Market analyst firm Gartner recently said that Fusion-io was fifth overall in the enterprise SSD market (SSDs + PCIe flash), behind fourth-placed SanDisk, which has virtually no PCIe presence, so the acquisition can be viewed as stategic.

SanDisk has a $6 billion annual revenue run rate and made a billion dollar profit in 2013. It certainly is in a healthy state.

Over the past couple of years SanDisk has also acquired:

  • SMART Storage for its SSDs and flash DIMMs for $307 million in July 2013
  • FlashSoft for PCIe flash aching software in February 2012
  • Pliant for SSD controllers in May 2011
  • Overall, Fusion-io is its largest flash business acquisition so far. The company will benefit from the addition of Fusion-io’s leading PCIe solutions to SanDisk’s vertically integrated business model.

    SanDisk CEO Sanjay Mehrotra was quoted as saying-- “Fusion-io will accelerate our efforts to enable the flash-transformed data center, helping companies better manage increasingly heavy data workloads at a lower total cost of ownership.”

    SanDisk is hell-bent on penetrating the enterprise flash market, pushing the concept of an all-flash data centre.

    Fusion-io has just announced its third generation “Atomic” ioMemory PCIe card products. SanDisk will inherit all of its server OEM and channel partner relationships and be able to guarantee a steady supply of flash chips from its own flash foundry partnership with Toshiba.

    The acquisition is expected to close in the third quarter of SanDisk’s fiscal 2014, unless somebody else jumps in with a bid and prolongs the affair.

    The flash industry is consolidating at a dizzying rate, with Seagate acquiring LSI’s PCIe card flash business from Avago for $450 million last month.

    It followed on from multiple acquisitions by Western Digital, and Toshiba buying Violin's Velocity PCIe card business.

    Click here to order the best dedicated server and at a great price.

    This $1.1 billion SanDisk acquisition price is slightly more than double that, so there is a premium for Fusion-io’s market success over LSI, but still, two times its annual revenue? Sounds like a good bargain.

    Fusion-io CEO Shane Robinson was pleased and he said-- “This transaction represents a compelling opportunity for Fusion-io’s employees, customers and shareholders. Fusion-io’s innovative hardware and software solutions will be improved by SanDisk’s worldwide scale and vertical integration, enabling a combined company that can offer an even more compelling value proposition for customers and partners.”

    In other IT news

    Microsoft said earlier this morning that it has found a new method to greatly increase the computing capabilities of its data centers, despite the fact that Moore's Law is wheezing towards its inevitable demise sooner as opposed to later.

    In a paper to be presented this week at the International Symposium on Computer Architecture (ISCA) titled 'A Reconfigurable Fabric for Accelerating Large-Scale Datacenter Services' a group of top Microsoft Research engineers explain how the company has dealt with the slowdown in single-core clock-rate improvements that has occurred over the past eleven years.

    To get around that issue, Microsoft has designed a system it calls Catapult, which automatically offloads some of the advanced technology that powers its Bing search engine onto clusters of highly efficient, low-power FPGA chips attached to typical Intel Xeon server processors.

    Think of FPGAs (field-programmable gate arrays) as chips whose circuits can be customized and tweaked as required, allowing specific tasks to be offloaded from the Xeon processors and accelerated in FPGA hardware.

    This approach may save Microsoft from a rarely acknowledged problem that lurks in the technology industry-- that processors are not getting much faster these days. And that would be a reality.

    For the past fifty years or so, almost every aspect of our global economy has been affected by Moore's Law, which states that the number of transistors on a chip of the same size will double every 1 1/2 year or so, resulting in faster performance and much improved power efficiency.

    Source: IDC.

    Get the most dependable SMTP server for your company. You will congratulate yourself!

    Share on Twitter.

    IT News Archives | Site Search | Advertise on IT Direction | Contact | Home

    All logos, trade marks or service marks on this site are the property of their respective owners.

    Sponsored by Sure Mail™, Avantex and
    by Montreal Server Colocation.

           © IT Direction. All rights reserved.