Data Efficiency in the News


Microsoft Fortifies Commitment to Open Source, Becomes Linux Foundation Platinum Member

| Home | The Linux Foundation

The Linux Foundation, the nonprofit advancing professional open source management for mass collaboration, today announced that Microsoft has joined the organization at a Platinum member during Microsoft’s Connect(); developer event in New York.

From cloud computing and networking to gaming, Microsoft has steadily increased its engagement in open source projects and communities. The company is currently a leading open source contributor on GitHub and earlier this year announced several milestones that indicate the scope of its commitment to open source development. The company released the open source .NET Core 1.0; partnered with Canonical to bring Ubuntu to Windows 10; worked with FreeBSD to release an image for Azure; and after acquiring Xamarin, Microsoft open sourced its software development kit. In addition, Microsoft works with companies like Red Hat, SUSE and others to support their solutions in its products.

“As a cloud platform company we aim to help developers achieve more using the platforms and languages they know,” said Scott Guthrie, Executive Vice President, Microsoft Cloud and Enterprise Group. “The Linux Foundation is home not only to Linux, but many of the community’s most innovative open source projects. We are excited to join The Linux Foundation and partner with the community to help developers capitalize on the shift to intelligent cloud and mobile experiences.”

Microsoft already contributes to several Linux Foundation projects, including Node.js Foundation, OpenDaylight, Open Container Initiative, R Consortium and Open API Initiative.

John Gossman, Architect on the Microsoft Azure team, will join The Linux Foundation Board of Directors.

“Microsoft has grown and matured in its use of and contributions to open source technology,” said Jim Zemlin, Executive Director of The Linux Foundation. “The company has become an enthusiastic supporter of Linux and of open source and a very active member of many important projects. Membership is an important step for Microsoft, but also for the open source community at large, which stands to benefit from the company’s expanding range of contributions.”

To view a full roster of Linux Foundation members, please visit

About The Linux Foundation

The Linux Foundation is the organization of choice for the world’s top developers and companies to build ecosystems that accelerate open technology development and commercial adoption. Together with the worldwide open source community, it is solving the hardest technology problems by creating the largest shared technology investment in history. Founded in 2000, The Linux Foundation today provides tools, training and events to scale any open source project, which together deliver an economic impact not achievable by any one company. More information can be found at

Read more


On-Premises Can Be Just as Effective as On-Cloud


There are plenty of good reasons to push the enterprise infrastructure model onto the cloud. It is cheaper, more flexible, and is becoming increasingly more secure and reliable as both hardware and software architectures evolve.

But there are plenty of good reasons to maintain a certain amount of on-premises infrastructure as well, and as time goes by, the deployment and management of that infrastructure will become vastly less complicated and expensive than what you find in today’s bloated data center.

Even Amazon is starting to rethink the all-cloud strategy that has driven its system development over the years. The company undoubtedly still thinks it provides the optimum environment for the bulk of enterprise workloads, but with releases like the recent Amazon Linux AMI (Amazon Machine Image), it also recognizes that some infrastructure will remain in the enterprise, at least for the time being. The system, according to E-Commerce Times, allows users to deploy the Linux AMI on internal resources, essentially providing tighter integration between local infrastructure and Amazon data centers. In this way, even high-performance applications can maintain a stable and secure link across hybrid deployments which, according to Amazon, would be most useful when developing and testing new apps and workloads at home before deploying them in the cloud.

In addition, there is the ability to deploy data reduction in either or bot on-premises or the cloud with Linux. Simply employ Permabit VDO for Data Centers in both and you can reduce data costs, increase data density and minimize network bandwidth consumption.

Whether on-cloud or on-premises, however, the final call will be made by the enterprise, which will now have to take into consideration its workload and application requirements, cost factors, performance goals, and a host of other elements for each and every deployment.


Whether on-cloud or on-premises, however, the final call will be made by the enterprise, which will now have to take into consideration its workload and application requirements, cost factors, performance goals, and a host of other elements for each and every deployment.

Read more


Rackspace Announces Support for Red Hat Ceph Storage 2

| Marketwired

Rackspace today announced support for Red Hat Ceph Storage 2, expanding its Rackspace Private Cloud powered by Red Hat porftfolio of services. This new version from Red Hat is based on the 10.2 Jewel community version release of Ceph and addresses key stability requirements for enterprise customers.

Data storage growth is both a reality and a burden for today’s enterprises. With the move to digital and data intensive workloads, enterprises are storing an increasing amount of data.

With Rackspace Private Cloud powered by Red Hat support for the Red Hat Ceph Storage 2 release, customers can gain the performance benefits of software defined storage, without taking on the burden of having to deploy, operate, monitor and troubleshoot their OpenStack or OpenStack storage environment.

This new release is the most stable version yet of Ceph and provides numerous features and benefits to enterprise users including:

  • Exabyte-level scalability
  • Open APIs
  • Security
  • Reliability and availability
  • Multi-Datacenter Support
  • Performance
  • Cost effectiveness

In addition Permabit Data reduction can be deployed with Ceph to effectively reduction storage consumption and cost. The impact of data reduction can increase data density enabling more efficient use of Rackspace resources lowering costs for each customer deployment.


Read more


Permabit and Calsoft announce reseller agreement for storage and cloud solutions

| PR Newswire

Permabit Technology Corporation, the open source data reduction experts, and Calsoft Pvt. Ltd., a leading IT services company specializing in storage, networking, virtualization and cloud business verticals have entered into a reseller agreement for Permabit VDO.  The agreement will enable Calsoft to deliver industry-leading data reduction solutions via their consulting and systems integration services for storage and cloud solutions.

Calsoft will deploy Permabit VDO in its storage solutions for software-defined data centers and storage in addition to cloud solutions for public, private and hybrid clouds.

“We selected Permabit because it is a solid data reduction offering that spans OEM and ODM deployments in Linux. Permabit data reduction complements our existing efforts in storage, cloud development and integration services,” said Anupam Bhide, CEO of Calsoft Pvt. Ltd.  “With Permabit VDO we will be able to deploy across Red Hat, Ubuntu and CentOS based solutions enabling our clients to optimize data center efficiency, increasing data density while avoiding costly expansion.”

The leader in data reduction technology, Permabit Technology Corporation recently announced the latest release of its Virtual Data Optimizer (VDO) software, VDO 6 – the only modular data reduction solution available for the Linux block storage stack.  VDO delivers the company’s patented deduplication, HIOPS Compression™ and thin provisioning in a commercial software package for Linux for enterprise hybrid cloud data centers and cloud service providers.

“Calsoft excels in software development and the delivery of integration services for complex data center storage and cloud solutions,” said Tom Cook, Permabit CEO and President. “We are thrilled to welcome Calsoft to Permabit’s reseller network and are fully committed to their success.”

To learn more about Permabit VDO Data Reduction software visit:

About Calsoft

Calsoft is a leading software product engineering Services Company specializing in StorageNetworkingVirtualization and Cloud business verticals. Calsoft provides End-to-End Product Development, Quality Assurance Sustenance, Solution Engineering and Professional Services expertise to assist customers in achieving their product development and business goals. For more information ,visit

Follow Calsoft on Twitter:

And/or on LinkedIn:

About Permabit:

Permabit pioneers the development of data reduction software that provides data deduplication, compression, and thin provisioning. Our innovative products enable customers to get to market quickly with solutions that cut effective cost, accelerate performance, and gain a competitive advantage. Just as server virtualization revolutionized the economics of compute, Permabit software is transforming the economics of storage today.

Permabit is headquartered in Cambridge, Massachusetts with operations in California, Korea and Japan. For more information, visit

Follow Permabit on Twitter:

And/or on LinkedIn:

Read more


Cloud adoption keeps moving ahead

| Computerworld

Companies continue their migration of both applications and computing infrastructure to the cloud at a steady pace. They have moved 45% of their applications and computing infrastructure to the cloud already, and they expect well over half of their IT environment to be cloud-based by 2018, according to a recent IDG Enterprise survey of 925 IT decision makers.

On average, IT decision-makers (ITDMs) plan to allocate more than a quarter of their total IT budgets to cloud spending, but organizations with fewer than 1,000 employees are making significantly different choices than larger enterprises in how they spend that money.

The four main drivers moving IT decision makers to cloud computing are:

  1. Lower total cost of ownership
  2. Replacing on-premise legacy systems
  3. Enabling business continuity
  4. Speed of development

That’s not to say there aren’t concerns about the move to cloud; ITDMs name the following as their top three worries in each of the major cloud models:

Public cloud

  • 43% Where data is stored
  • 41% Security
  • 21% Vendor lock-in

Private cloud

  • 24% Vendor lock-in
  • 22% Lack of appropriate skills
  • 21% Security concerns

Hybrid cloud

  • 24% Security
  • 19% Where data is stored
  • 18% Lack of appropriate

In addition the need to optimize public cloud, hybrid cloud and  private cloud data and its costs cannot be underscored.  Applying data reduction to a Linux implementation in any of these deployment models will drive down TCO and improve operating efficiency.  Data reduction can now be deployed in almost any Linux OS!

Read more


Top Priority for IT Investments: Improve Service to Quickly Meet Business Needs

| Stock Market

The research, conducted across 125 IT decision makers in the US, revealed the number one priority for IT investments: Improve service to quickly meet business needs. Reducing risk was the second major priority, and the third, as stated by respondents, was realizing higher levels of performance to support mission-critical applications.

“The findings of this research clearly indicate that the number one priority for IT decision makers is ensuring that IT becomes an enabler of business and not a hindrance,” said Joshua Yulish, president and CEO of TmaxSoft, Inc. “To this end, IT must provide open systems that afford greater flexibility and speed at a lower cost. Businesses are looking to not only improve service to respond to business needs, but innovate faster and realize higher levels of performance to support key objectives.”

Dave Lasseter, VP Power Systems Sales at Mainline, an IBM and TmaxSoft partner, added: “These findings mirror what we are seeing in the market today. IT must take the initiative in delivering solutions and services that support innovation and enable the business to adapt to changes in strategy, market conditions, and regulatory requirements.”

Key findings include:

  • The top priority among 24% of respondents was improving service to dynamically respond to business needs.
  • The second #1 priority was ensuring uptime, cited by 21%, and third was the need to reduce the administrative cost and burden by consolidating systems, cited by 19% of respondents.
  • The top-rated second priority for IT decision makers was reducing risk (identified by 21% of the sample), followed by realizing higher levels of performance to support mission-critical applications (18% of respondents).

Not only are open systems a requisite. There is also a need for Linux based data reduction that can deliver enterprise wide operating and storage efficiency. This will result in lower data bandwidth needs, improved data density (fewer servers and storage devices) and reduce data center footprint which will improve operating efficiency.


Read more


OpenStack expands both its customer reach and deployments size

| ZDNet

In 451 Research’s recent report on OpenStack adoption among enterprise private cloud users, they found that 72 percent of OpenStack-based clouds are between 1,000 and 10,000 cores and three-fourths choose OpenStack to increase operational efficiency and app deployment speed.

They also found that OpenStack is not just for large enterprises. Two-thirds of respondents (65 percent) are in organizations of between 1,000 and 10,000 employees.

The survey also uncovered that OpenStack-powered clouds have moved beyond small-scale deployments. Approximately 72 percent of OpenStack enterprise deployments are between 1,000 to 10,000 cores in size. Additionally, 5 percent of OpenStack clouds among enterprises top the 100,000-core mark. So, while OpenStack may be expanding its reach into smaller companies, it’s being used for larger deployments.

Curiously, OpenStack users are adopting containers at a faster rate than the rest of the enterprise market with 55 percent of OpenStack users also using containers, compared to 17 percent with other cloud users. What’s odd about this, as Mark Shuttleworth, founder of Canonical and Ubuntu, pointed out to me at an OpenStack Summit meeting, is OpenStack is not especially well-suited for containers.

Well, not yet anyway. But, it will be with companies both moving the technology forward and customers demanding it.

OpenStack is also moving along to real enterprise workloads rather than just testing and development work. These include infrastructure services (66 percent), business applications and big data (60 percent and 59 percent, respectively), and web services and ecommerce (57 percent).

You’ll also find OpenStack clouds running in a wide variety of businesses. While 20 percent cited the technology industry, manufacturing (15 percent), retail/hospitality (11 percent), professional services (10 percent), healthcare (7 percent), insurance (6 percent), transportation (5 percent), communications/media (5 percent), wholesale trade (5 percent), energy and utilities (4 percent), education (3 percent), financial services (3 percent), and government (3 percent) were all represented.

Why are so many businesses across so many industries adopting OpenStack? Simple. Increasing operational efficiency and accelerating innovation/deployment speed are top business drivers for enterprise adoption of OpenStack, at 76 and 75 percent, respectively. Supporting DevOps is a close second, at 69 percent. Reducing cost and standardizing on OpenStack APIs were close behind, at 50 and 45 percent, respectively. In addition to operational efficiency, data efficiency is also becoming table stakes in today’s data center. Not just to reduce effective storage costs but also to increase data density which reduces /eliminates data center expansion. The bottom line is that openstack and its efficiency impact  help business bottom line and that’s why the adoption is increasing.

“Our research in aggregate indicates enterprises globally are moving beyond using OpenStack for science projects and basic test and development to workloads that impact the bottom line,” said Al Sadowski, 451 Research’s research vice president. “This is supported by our OpenStack Market Monitor which projects an overall market size of over $5 billion in 2020 with APAC, namely China, leading the way in terms of growth.”

Mark Collier, COO of the OpenStack Foundation, agreed, “The research [is] telling us that OpenStack is not merely an interesting technology, but it’s a cornerstone technology. Companies are using OpenStack to do work that matters to their businesses, and they’re using it to support their journey to a changing landscape in which rapid development and deployment of software is the primary means of competitive advantage.”

Read more


Open Software Defined Storage needs Data Reduction

| Storage Swiss

Software defined storage (SDS) promises to abstract storage services from storage hardware, freeing organizations from the “lock” of having to use specific storage hardware. But the price for this “freedom” is imprisonment to a single storage software vendor. While certainly an improvement, it is not the panacea that SDS vendors promote. Open SDS takes data center flexibility to the next step. With Open SDS a customer has not only the flexibility to select the storage hardware they want, they also have flexibility on the software side of the equation. Examples of these more flexible SDS solutions are Red Hat’s SDS solutions CEPH and Gluster.

Flexible storage software is not only good for customers, it is also good for the third party vendors looking to add value to an existing SDS feature set. One can imagine an app store type of concept where Open SDS vendors can present a variety of extensions to the current SDS capabilities. In this scenario everyone wins. The vendor that owns the Open SDS support can offer more reasons for the customer to do business with them, the third party vendors can respond faster to customer needs because they don’t have to recreate an entire storage software stack, and the customer wins because they should see greater variety and innovation in the options available.

Data Reduction for CEPH/Gluster

An excellent example of the innovation that an Open SDS solution can foster is found in Permabit’s recent update of its VDO product. VDO or Virtual Disk Optimizer is a deduplication and compression solution and the latest version of VDO is tested and certified with Red Hat CEPH and Gluster. VDO has been on the market for over five years, and we did an extensive test of the solution a couple of years ago. Thanks to its OEM roots, VDO is potentially the most heavily tested data reduction solution on the market today. To learn more about CEPH and Gluster see our article “Product Analysis: Open Software Defined Storage – Ceph or Gluster?”.

Permabit designed their data reduction solution to provide high performance, so high as to not interfere when used in all-flash configurations. Bringing data reduction to CEPH and Gluster opens up a world of possibilities for data centers. Imagine a CEPH solution with all-flash, or mostly flash, supporting a virtual or containerized environment. Or a Gluster solution running parallel analytics processing on IoT data hosted on flash storage. Both approaches with Permabit VDO will also substantially increase data center density, another win for data center managers! While both CEPH and Gluster have supported flash for a long time, Permabit data reduction technology makes their use much more competitively appealing from a price perspective without impacting performance.

Pricing That Makes Sense

One aspect of third party solutions that usually ruins their adoption is the price. Most of the time these vendors want so much for their solution that IT professionals decide to take a pass. Permabit is a clear exception here. It is pricing the solution at an almost unheard of $199 per 16TBs of data and $3000 for 256TB of data with annual maintenance at the same price points. Using Permabit’s very conservative 2.5:1 efficiency claim, the 256TB license/maintenance costs would then enable 256TB of storage to act like 640TBs of capacity representing 384TBs in savings. Last time I checked you can’t get 384TBs for $3000.

StorageSwiss Take

We are not sure who the biggest winner is here. Red Hat, especially as they mature and simplify CEPH/Gluster, have another advantage over many other SDS solutions on the market today. It is now Open and Optimized. Permabit’s VDO at these price points should be an absolute no brainer for CEPH/Gluster customers. And customers are getting more than double their storage, without impacting performance for a fraction of the cost of actually buying it.

Read more


Alternative storage options: Ceph object storage, Swift and more

| VMware information, news and tips

Alternative storage options: Ceph object storage, Swift and more by Object storage is rapidly replacing proprietary SAN filers as the go-to choice for storage in the modern data center. But is it right for your virtual environment?

Object storage is changing the data center. Commodity storage offerings provide a well-performing alternative for expensive proprietary SAN filers.

There are currently three different products for object storage dominating the market: the legacy Swift, Amazon Simple Storage Service (S3) and the more recent Ceph object storage offering. Swift is mostly used in OpenStack cloud environments and works with applications that address Swift object storage through direct API calls. That means it’s fairly limited in use: If you have a generic application or OS, there’s no easy way integrate with Swift.

S3 has been around for a long time and works in Amazon cloud environments. Its access methods are limited as well, which means it’s not the best candidate for a generic object storage product. S3 is best used to deploy images in an Amazon Web Services cloud environment. Unfortunately, this isn’t helpful if you’re using VMware vSphere.

Ceph is the most open of all the object storage offerings, not only because it’s open source, but also because it offers several different client interfaces:

API access. This is the most common access model in object storage, but it doesn’t work for VMware environments, as you would need to rewrite the vSphere code to access it.

  • The Ceph file system. This is a special-purpose file system that can be used on the object storage client. Since this object storage client would be an ESXi server, this option also isn’t very usable in VMware environments.
  • The RADOS Block Device. This adds a block device to the client OS by loading a kernel module and integrating it on ESXi; this is also difficult to use in a VMware environment.
  • The new iSCSI interface. This is a new and promising development in Ceph object storage. In the new iSCSI interface, the Ceph storage cluster includes an iSCSI target, which means the client will access it like any other iSCSI-based SAN offering.

Of these four access methods, the iSCSI interface is the only one that really works in a VMware environment. You may be wondering, doesn’t that just replace one SAN product with another? The answer is absolutely not. Even if the client only sees an iSCSI target, you’ll be dealing with a flexible, scalable and affordable SAN offering on the back end, which is much cheaper than traditional SAN environments.

The iSCSI target interface for Ceph object storage is relatively new, and you’ll notice it may not be available on all Ceph object storage products. It is included in Ceph’s SUSE-supported offering, SUSE Enterprise Storage 3, and it is likely that other Ceph vendors, such as Red Hat, will soon follow suit. The iSCSI interface code shows in SUSE first because SUSE is its main developer.

Since Ceph object storage is revolutionizing the world of enterprise storage, it might be a good idea to take the time to explore its possibilities, especially in VMware vSphere environments. Once configured, it will behave just like any other iSCSI data store.

Read more