Data Efficiency in the News

curata__kt4Wdywr67xX1JP.jpeg

On-Premises Can Be Just as Effective as On-Cloud

| itbusinessedge.com
curata__kt4Wdywr67xX1JP.jpeg

There are plenty of good reasons to push the enterprise infrastructure model onto the cloud. It is cheaper, more flexible, and is becoming increasingly more secure and reliable as both hardware and software architectures evolve.

But there are plenty of good reasons to maintain a certain amount of on-premises infrastructure as well, and as time goes by, the deployment and management of that infrastructure will become vastly less complicated and expensive than what you find in today’s bloated data center.

Even Amazon is starting to rethink the all-cloud strategy that has driven its system development over the years. The company undoubtedly still thinks it provides the optimum environment for the bulk of enterprise workloads, but with releases like the recent Amazon Linux AMI (Amazon Machine Image), it also recognizes that some infrastructure will remain in the enterprise, at least for the time being. The system, according to E-Commerce Times, allows users to deploy the Linux AMI on internal resources, essentially providing tighter integration between local infrastructure and Amazon data centers. In this way, even high-performance applications can maintain a stable and secure link across hybrid deployments which, according to Amazon, would be most useful when developing and testing new apps and workloads at home before deploying them in the cloud.

In addition, there is the ability to deploy data reduction in either or bot on-premises or the cloud with Linux. Simply employ Permabit VDO for Data Centers in both and you can reduce data costs, increase data density and minimize network bandwidth consumption.

Whether on-cloud or on-premises, however, the final call will be made by the enterprise, which will now have to take into consideration its workload and application requirements, cost factors, performance goals, and a host of other elements for each and every deployment.

 

Whether on-cloud or on-premises, however, the final call will be made by the enterprise, which will now have to take into consideration its workload and application requirements, cost factors, performance goals, and a host of other elements for each and every deployment.

Read more

curata__enVFBp0vTZYpfUg.jpeg

Rackspace Announces Support for Red Hat Ceph Storage 2

| Marketwired
curata__enVFBp0vTZYpfUg.jpeg

Rackspace today announced support for Red Hat Ceph Storage 2, expanding its Rackspace Private Cloud powered by Red Hat porftfolio of services. This new version from Red Hat is based on the 10.2 Jewel community version release of Ceph and addresses key stability requirements for enterprise customers.

Data storage growth is both a reality and a burden for today’s enterprises. With the move to digital and data intensive workloads, enterprises are storing an increasing amount of data.

With Rackspace Private Cloud powered by Red Hat support for the Red Hat Ceph Storage 2 release, customers can gain the performance benefits of software defined storage, without taking on the burden of having to deploy, operate, monitor and troubleshoot their OpenStack or OpenStack storage environment.

This new release is the most stable version yet of Ceph and provides numerous features and benefits to enterprise users including:

  • Exabyte-level scalability
  • Open APIs
  • Security
  • Reliability and availability
  • Multi-Datacenter Support
  • Performance
  • Cost effectiveness

In addition Permabit Data reduction can be deployed with Ceph to effectively reduction storage consumption and cost. The impact of data reduction can increase data density enabling more efficient use of Rackspace resources lowering costs for each customer deployment.

 

Read more

curata__s8dZYCNYdbPFJ5l.png

Permabit and Calsoft announce reseller agreement for storage and cloud solutions

| PR Newswire
curata__s8dZYCNYdbPFJ5l.png

Permabit Technology Corporation, the open source data reduction experts, and Calsoft Pvt. Ltd., a leading IT services company specializing in storage, networking, virtualization and cloud business verticals have entered into a reseller agreement for Permabit VDO.  The agreement will enable Calsoft to deliver industry-leading data reduction solutions via their consulting and systems integration services for storage and cloud solutions.

Calsoft will deploy Permabit VDO in its storage solutions for software-defined data centers and storage in addition to cloud solutions for public, private and hybrid clouds.

“We selected Permabit because it is a solid data reduction offering that spans OEM and ODM deployments in Linux. Permabit data reduction complements our existing efforts in storage, cloud development and integration services,” said Anupam Bhide, CEO of Calsoft Pvt. Ltd.  “With Permabit VDO we will be able to deploy across Red Hat, Ubuntu and CentOS based solutions enabling our clients to optimize data center efficiency, increasing data density while avoiding costly expansion.”

The leader in data reduction technology, Permabit Technology Corporation recently announced the latest release of its Virtual Data Optimizer (VDO) software, VDO 6 – the only modular data reduction solution available for the Linux block storage stack.  VDO delivers the company’s patented deduplication, HIOPS Compression™ and thin provisioning in a commercial software package for Linux for enterprise hybrid cloud data centers and cloud service providers.

“Calsoft excels in software development and the delivery of integration services for complex data center storage and cloud solutions,” said Tom Cook, Permabit CEO and President. “We are thrilled to welcome Calsoft to Permabit’s reseller network and are fully committed to their success.”

To learn more about Permabit VDO Data Reduction software visit:
http://permabit.com/products-overview/albireo-virtual-data-optimizer-vdo/

About Calsoft

Calsoft is a leading software product engineering Services Company specializing in StorageNetworkingVirtualization and Cloud business verticals. Calsoft provides End-to-End Product Development, Quality Assurance Sustenance, Solution Engineering and Professional Services expertise to assist customers in achieving their product development and business goals. For more information ,visit http://www.calsoftinc.com

Follow Calsoft on Twitter: https://twitter.com/CalsoftInc

And/or on LinkedIn: https://www.linkedin.com/company/calsoft

About Permabit:

Permabit pioneers the development of data reduction software that provides data deduplication, compression, and thin provisioning. Our innovative products enable customers to get to market quickly with solutions that cut effective cost, accelerate performance, and gain a competitive advantage. Just as server virtualization revolutionized the economics of compute, Permabit software is transforming the economics of storage today.

Permabit is headquartered in Cambridge, Massachusetts with operations in California, Korea and Japan. For more information, visit www.permabit.com.

Follow Permabit on Twitter: https://twitter.com/Permabit

And/or on LinkedIn: https://www.linkedin.com/company/permabit

Read more

curata__5cd092b74bfbc44b4220a3e182130728.PNG

Cloud adoption keeps moving ahead

| Computerworld
curata__5cd092b74bfbc44b4220a3e182130728.PNG

Companies continue their migration of both applications and computing infrastructure to the cloud at a steady pace. They have moved 45% of their applications and computing infrastructure to the cloud already, and they expect well over half of their IT environment to be cloud-based by 2018, according to a recent IDG Enterprise survey of 925 IT decision makers.

On average, IT decision-makers (ITDMs) plan to allocate more than a quarter of their total IT budgets to cloud spending, but organizations with fewer than 1,000 employees are making significantly different choices than larger enterprises in how they spend that money.

The four main drivers moving IT decision makers to cloud computing are:

  1. Lower total cost of ownership
  2. Replacing on-premise legacy systems
  3. Enabling business continuity
  4. Speed of development

That’s not to say there aren’t concerns about the move to cloud; ITDMs name the following as their top three worries in each of the major cloud models:

Public cloud

  • 43% Where data is stored
  • 41% Security
  • 21% Vendor lock-in

Private cloud

  • 24% Vendor lock-in
  • 22% Lack of appropriate skills
  • 21% Security concerns

Hybrid cloud

  • 24% Security
  • 19% Where data is stored
  • 18% Lack of appropriate

In addition the need to optimize public cloud, hybrid cloud and  private cloud data and its costs cannot be underscored.  Applying data reduction to a Linux implementation in any of these deployment models will drive down TCO and improve operating efficiency.  Data reduction can now be deployed in almost any Linux OS!

Read more

curata__f423c5616ffe8e5b903e1d5a3d306c5b.PNG

Top Priority for IT Investments: Improve Service to Quickly Meet Business Needs

| Stock Market
curata__f423c5616ffe8e5b903e1d5a3d306c5b.PNG

The research, conducted across 125 IT decision makers in the US, revealed the number one priority for IT investments: Improve service to quickly meet business needs. Reducing risk was the second major priority, and the third, as stated by respondents, was realizing higher levels of performance to support mission-critical applications.

“The findings of this research clearly indicate that the number one priority for IT decision makers is ensuring that IT becomes an enabler of business and not a hindrance,” said Joshua Yulish, president and CEO of TmaxSoft, Inc. “To this end, IT must provide open systems that afford greater flexibility and speed at a lower cost. Businesses are looking to not only improve service to respond to business needs, but innovate faster and realize higher levels of performance to support key objectives.”

Dave Lasseter, VP Power Systems Sales at Mainline, an IBM and TmaxSoft partner, added: “These findings mirror what we are seeing in the market today. IT must take the initiative in delivering solutions and services that support innovation and enable the business to adapt to changes in strategy, market conditions, and regulatory requirements.”

Key findings include:

  • The top priority among 24% of respondents was improving service to dynamically respond to business needs.
  • The second #1 priority was ensuring uptime, cited by 21%, and third was the need to reduce the administrative cost and burden by consolidating systems, cited by 19% of respondents.
  • The top-rated second priority for IT decision makers was reducing risk (identified by 21% of the sample), followed by realizing higher levels of performance to support mission-critical applications (18% of respondents).

Not only are open systems a requisite. There is also a need for Linux based data reduction that can deliver enterprise wide operating and storage efficiency. This will result in lower data bandwidth needs, improved data density (fewer servers and storage devices) and reduce data center footprint which will improve operating efficiency.

  

Read more

curata__MqHHqUqXkFiIa2T.png

OpenStack expands both its customer reach and deployments size

| ZDNet
curata__MqHHqUqXkFiIa2T.png

In 451 Research’s recent report on OpenStack adoption among enterprise private cloud users, they found that 72 percent of OpenStack-based clouds are between 1,000 and 10,000 cores and three-fourths choose OpenStack to increase operational efficiency and app deployment speed.

They also found that OpenStack is not just for large enterprises. Two-thirds of respondents (65 percent) are in organizations of between 1,000 and 10,000 employees.

The survey also uncovered that OpenStack-powered clouds have moved beyond small-scale deployments. Approximately 72 percent of OpenStack enterprise deployments are between 1,000 to 10,000 cores in size. Additionally, 5 percent of OpenStack clouds among enterprises top the 100,000-core mark. So, while OpenStack may be expanding its reach into smaller companies, it’s being used for larger deployments.

Curiously, OpenStack users are adopting containers at a faster rate than the rest of the enterprise market with 55 percent of OpenStack users also using containers, compared to 17 percent with other cloud users. What’s odd about this, as Mark Shuttleworth, founder of Canonical and Ubuntu, pointed out to me at an OpenStack Summit meeting, is OpenStack is not especially well-suited for containers.

Well, not yet anyway. But, it will be with companies both moving the technology forward and customers demanding it.

OpenStack is also moving along to real enterprise workloads rather than just testing and development work. These include infrastructure services (66 percent), business applications and big data (60 percent and 59 percent, respectively), and web services and ecommerce (57 percent).

You’ll also find OpenStack clouds running in a wide variety of businesses. While 20 percent cited the technology industry, manufacturing (15 percent), retail/hospitality (11 percent), professional services (10 percent), healthcare (7 percent), insurance (6 percent), transportation (5 percent), communications/media (5 percent), wholesale trade (5 percent), energy and utilities (4 percent), education (3 percent), financial services (3 percent), and government (3 percent) were all represented.

Why are so many businesses across so many industries adopting OpenStack? Simple. Increasing operational efficiency and accelerating innovation/deployment speed are top business drivers for enterprise adoption of OpenStack, at 76 and 75 percent, respectively. Supporting DevOps is a close second, at 69 percent. Reducing cost and standardizing on OpenStack APIs were close behind, at 50 and 45 percent, respectively. In addition to operational efficiency, data efficiency is also becoming table stakes in today’s data center. Not just to reduce effective storage costs but also to increase data density which reduces /eliminates data center expansion. The bottom line is that openstack and its efficiency impact  help business bottom line and that’s why the adoption is increasing.

“Our research in aggregate indicates enterprises globally are moving beyond using OpenStack for science projects and basic test and development to workloads that impact the bottom line,” said Al Sadowski, 451 Research’s research vice president. “This is supported by our OpenStack Market Monitor which projects an overall market size of over $5 billion in 2020 with APAC, namely China, leading the way in terms of growth.”

Mark Collier, COO of the OpenStack Foundation, agreed, “The research [is] telling us that OpenStack is not merely an interesting technology, but it’s a cornerstone technology. Companies are using OpenStack to do work that matters to their businesses, and they’re using it to support their journey to a changing landscape in which rapid development and deployment of software is the primary means of competitive advantage.”

Read more

curata__94268666bb6689de67da501a638be5e2.PNG

Open Software Defined Storage needs Data Reduction

| Storage Swiss
curata__94268666bb6689de67da501a638be5e2.PNG

Software defined storage (SDS) promises to abstract storage services from storage hardware, freeing organizations from the “lock” of having to use specific storage hardware. But the price for this “freedom” is imprisonment to a single storage software vendor. While certainly an improvement, it is not the panacea that SDS vendors promote. Open SDS takes data center flexibility to the next step. With Open SDS a customer has not only the flexibility to select the storage hardware they want, they also have flexibility on the software side of the equation. Examples of these more flexible SDS solutions are Red Hat’s SDS solutions CEPH and Gluster.

Flexible storage software is not only good for customers, it is also good for the third party vendors looking to add value to an existing SDS feature set. One can imagine an app store type of concept where Open SDS vendors can present a variety of extensions to the current SDS capabilities. In this scenario everyone wins. The vendor that owns the Open SDS support can offer more reasons for the customer to do business with them, the third party vendors can respond faster to customer needs because they don’t have to recreate an entire storage software stack, and the customer wins because they should see greater variety and innovation in the options available.

Data Reduction for CEPH/Gluster

An excellent example of the innovation that an Open SDS solution can foster is found in Permabit’s recent update of its VDO product. VDO or Virtual Disk Optimizer is a deduplication and compression solution and the latest version of VDO is tested and certified with Red Hat CEPH and Gluster. VDO has been on the market for over five years, and we did an extensive test of the solution a couple of years ago. Thanks to its OEM roots, VDO is potentially the most heavily tested data reduction solution on the market today. To learn more about CEPH and Gluster see our article “Product Analysis: Open Software Defined Storage – Ceph or Gluster?”.

Permabit designed their data reduction solution to provide high performance, so high as to not interfere when used in all-flash configurations. Bringing data reduction to CEPH and Gluster opens up a world of possibilities for data centers. Imagine a CEPH solution with all-flash, or mostly flash, supporting a virtual or containerized environment. Or a Gluster solution running parallel analytics processing on IoT data hosted on flash storage. Both approaches with Permabit VDO will also substantially increase data center density, another win for data center managers! While both CEPH and Gluster have supported flash for a long time, Permabit data reduction technology makes their use much more competitively appealing from a price perspective without impacting performance.

Pricing That Makes Sense

One aspect of third party solutions that usually ruins their adoption is the price. Most of the time these vendors want so much for their solution that IT professionals decide to take a pass. Permabit is a clear exception here. It is pricing the solution at an almost unheard of $199 per 16TBs of data and $3000 for 256TB of data with annual maintenance at the same price points. Using Permabit’s very conservative 2.5:1 efficiency claim, the 256TB license/maintenance costs would then enable 256TB of storage to act like 640TBs of capacity representing 384TBs in savings. Last time I checked you can’t get 384TBs for $3000.

StorageSwiss Take

We are not sure who the biggest winner is here. Red Hat, especially as they mature and simplify CEPH/Gluster, have another advantage over many other SDS solutions on the market today. It is now Open and Optimized. Permabit’s VDO at these price points should be an absolute no brainer for CEPH/Gluster customers. And customers are getting more than double their storage, without impacting performance for a fraction of the cost of actually buying it.

Read more

curata__fMSz69LZH1jqSRL.png

Alternative storage options: Ceph object storage, Swift and more

| VMware information, news and tips
curata__fMSz69LZH1jqSRL.png

Alternative storage options: Ceph object storage, Swift and more by Object storage is rapidly replacing proprietary SAN filers as the go-to choice for storage in the modern data center. But is it right for your virtual environment?

Object storage is changing the data center. Commodity storage offerings provide a well-performing alternative for expensive proprietary SAN filers.

There are currently three different products for object storage dominating the market: the legacy Swift, Amazon Simple Storage Service (S3) and the more recent Ceph object storage offering. Swift is mostly used in OpenStack cloud environments and works with applications that address Swift object storage through direct API calls. That means it’s fairly limited in use: If you have a generic application or OS, there’s no easy way integrate with Swift.

S3 has been around for a long time and works in Amazon cloud environments. Its access methods are limited as well, which means it’s not the best candidate for a generic object storage product. S3 is best used to deploy images in an Amazon Web Services cloud environment. Unfortunately, this isn’t helpful if you’re using VMware vSphere.

Ceph is the most open of all the object storage offerings, not only because it’s open source, but also because it offers several different client interfaces:

API access. This is the most common access model in object storage, but it doesn’t work for VMware environments, as you would need to rewrite the vSphere code to access it.

  • The Ceph file system. This is a special-purpose file system that can be used on the object storage client. Since this object storage client would be an ESXi server, this option also isn’t very usable in VMware environments.
  • The RADOS Block Device. This adds a block device to the client OS by loading a kernel module and integrating it on ESXi; this is also difficult to use in a VMware environment.
  • The new iSCSI interface. This is a new and promising development in Ceph object storage. In the new iSCSI interface, the Ceph storage cluster includes an iSCSI target, which means the client will access it like any other iSCSI-based SAN offering.

Of these four access methods, the iSCSI interface is the only one that really works in a VMware environment. You may be wondering, doesn’t that just replace one SAN product with another? The answer is absolutely not. Even if the client only sees an iSCSI target, you’ll be dealing with a flexible, scalable and affordable SAN offering on the back end, which is much cheaper than traditional SAN environments.

The iSCSI target interface for Ceph object storage is relatively new, and you’ll notice it may not be available on all Ceph object storage products. It is included in Ceph’s SUSE-supported offering, SUSE Enterprise Storage 3, and it is likely that other Ceph vendors, such as Red Hat, will soon follow suit. The iSCSI interface code shows in SUSE first because SUSE is its main developer.

Since Ceph object storage is revolutionizing the world of enterprise storage, it might be a good idea to take the time to explore its possibilities, especially in VMware vSphere environments. Once configured, it will behave just like any other iSCSI data store.

Read more

curata__w6RbiCiIrczejre.png

Open source no longer scares the enterprise

| ComputerWeekly.com
curata__w6RbiCiIrczejre.png

Open source breaks the rules on corporate procurement, but developers never play by the rules and now open source has sneaked in through the back door.

A study by Vanson Bourne for Rackspace reports that businesses are making big savings by using open source.

In the survey of 300 organisations, three out of five respondents cited cost savings as the top benefit, reducing average cost per project by £30,146.

With most IT projects at the lower end of the cost scale, such savings are significant.

About half of the organisations in the study reported greater innovation because of open source – and 46% said they used open source because of the competitive opportunities.

In fact, 30% said open source gave them the opportunity to respond more quickly to market trends as a driver. Almost half (45%) said it enabled them to get products and services to market sooner – with project lifecycles reduced by an average of six months.

Businesses are used to dealing with the major IT providers and some open source software companies have successfully mimicked commercial models to sell open source to the enterprise.

“Red Hat sells the perception of decreased risk, and it looks similar to proprietary software sales,” said Lindberg. “This was what most corporate procurement people are used to.”

But he added that, in his experience, open source tended to creep in from the bottom through unofficial routes. “It doesn’t require permission or payment. You can simply start using it to deliver business value. It is about people just trying to be more efficient at doing their jobs.”

According to Lindberg, it is a huge indictment of the traditional commercial software business model that no one wants to use the very expensive software stacks that corporate IT used to deploy. “People say they can do things better, faster, cheaper and more efficiently using community-oriented open source software,” he said.

 

Read more