Cloud service provider Nirvanix stunned its customers early this week telling them they have two weeks to find another home for their terabytes -- and in some cases petabytes -- of data because the company is shutting down.
First reported on Monday by the U.K.-based Web site Information Age, the sudden collapse of one of the largest cloud storage providers has left customers scrambling, especially those with huge amounts of data stored. Among them are National Geographic and Fox, which use Nirvanix to store their data, according to published reports. Many major providers, including Dell, HP, Riverbed, Symantec and TwinStrata, sourced their cloud storage from Nirvanix when offering their own services, according to StorageReview.com.
Since Monday's reports that Nirvanix has run out of money and informed customers and partners that they have two weeks to migrate their data to another cloud provider, the company has given them a slight reprieve: Now they have until Oct. 15, a customer support rep at the company told me Thursday.
Customers will no longer be permitted to upload any data as of next Monday, Sept. 23, but the customer service rep agreed with me that one might not want to upload another bit at this point.
More than 1,000 enterprise customers have data stored in the Nirvanix cloud service, according to Forrester Research analyst Henry Baltazar, who said in a blog post Wednesday that the provider's demise validates warnings that it's easier to upload to the cloud than to recover large quantities of data.
"The recent example with Nirvanix highlights why customers should also consider exit and migration strategies as they formulate their cloud storage deployments," Baltazar said. "One of the most significant challenges in cloud storage is related to how difficult it is to move large amounts of data from a cloud. While bandwidth has increased significantly over the years, even over large network links it could take days or even weeks to retrieve terabytes or petabytes of data from a cloud. For example, on a 1 Gbps link, it would take close to 13 days to retrieve 150 TB of data from a cloud storage service over a WAN link."
Gartner analyst Kyle Hilgendorf also emphasized in a blog post that failure to have an exit strategy when using a cloud service, especially for storage, can be a recipe for disaster. As for this week's Nirvanix news, Hilgendorf said: "What are clients do to? For most -- react...and react in panic. You may be facing the worst company fear -- losing actual data."
With everyone trying to migrate their data simultaneously and Nirvanix's limited network resources, many customers are finding themselves in a jam, said Andres Rodriguez, founder and CEO of Nasuni, which provides its own storage service that once used Nirvanix as its back-end target.
"What's happening now with Nirvanix is the equivalent of bank rush," Rodriguez said. "Everyone is trying to get their data out in a hurry and you know what that does to a network, and it's going to be very hard to get their data out."
Rodriguez said he has warned his customers for many months that he believed Nirvanix was at risk of going out of business. Rodriguez said when Nasuni used Nirvanix as its cloud storage provider two years ago, he became increasingly concerned it couldn't scale. Nasuni now uses Amazon Web Services Simple Storage Service (S3) for primary storage.
Nasuni runs annual tests against what Rodriguez believes are the largest cloud providers. The most recent test results released earlier this year concluded Amazon S3 and Windows Azure were the only two viable enterprise-grade storage services.
Nasuni just added a mirroring option that lets customers replicate their data stored in Amazon S3 to Windows Azure for added contingency. While Rodriguez believes Amazon S3 and Windows Azure are the most scalable and resilient, he warns it could be years before the majority of customers feel comfortable using the latter as their primary target.
Posted by Jeffrey Schwartz on 09/19/2013 at 3:49 PM2 comments
When Amazon Web Services (AWS) made EC2 available at reduced rates back in 2009 for those who sign long-term commitments, it helped kick start the lowering of cloud pricing. The idea behind its Reserved Instances was customers could lock in usage of capacity in one- and three-year terms.
Now Amazon is letting customers modify their Reserved Instances -- at least somewhat.
AWS this week said it's letting customers move their Reserved Instances between Availability Zones as long as the instances remain in the same region. Customers with accounts enabled for EC2-Classic can also move their Reserved Instances between EC2-Classic and EC2-VPC, explained AWS evangelist Jeff Barr on the AWS Blog.
"You can now make adjustments to your Reserved Instances as your needs and your architecture change," Barr noted. Customers can use the AWS Management Console or modify their Reserved Instances using the EC2 APIs or the AWS command line interface (CLI), noted Barr. The post explains how to make the change via the AWS console.
In addition to the pricing advantage Reserved Instances offer versus using capacity on demand is the assurance that capacity will be available when needed, Barr noted. But it's important to note that while you'll get the pricing advantage regardless of the network platform used, customers only get the capacity assurance within the network platform of the reserved instance.
Posted by Jeffrey Schwartz on 09/12/2013 at 2:57 PM0 comments
Workday, one of the fastest-growing Software as a Service (SaaS) companies, this week launched a big data analytics module to add to its portfolio of human resources and financial applications.
Delivered at the company's seventh annual Workday Rising conference in San Francisco, the new Workday Big Data Analytics is a key component of the company's latest update to its SaaS offerings called Workday 20. Workday Big Data Analytics gathers data from its portfolio as well as other data sources to help individuals develop specialized reports and benchmarks.
"For our customers, big data is about opening up the Workday cloud to import non-Workday data to more readily make business decisions," wrote Dan Beck, the company's vice president of product management.
Templates offered for those in HR include a market compensation comparison tool that gathers payroll data and industry salary averages from internal and external sources, headcount analysis to determine risks in workforce planning and one that helps determine performance of employees. Finance templates include one that helps a company determine its performance in comparison to competitors, sales performance reports and customer profitability analyses.
Partners including Deloitte Consulting and IBM Global Business Services have developed their own templates that, respectively, measure manager effectiveness and attrition reporting.
In addition to the big data analytics tool Workday 20 includes the release of its Notebooks on iPad, which lets managers gather worker profiles on the tablets and render reports on them.
Posted by Jeffrey Schwartz on 09/12/2013 at 2:48 PM0 comments
A startup led by some key VMware veterans and backed by the virtualization vendor's founder, Diane Greene, officially opened for business this week with the release of its cloud-based datacenter operations management service.
CloudPhysics, which describes itself as the Google of IT operations management, launched its namesake service aimed at simplifying the administration of virtual machines by using a vast real-time analytics engine that aggregates and analyzes billions of data points.
The Mountain View, Calif.-based company also said it has raised $10 million in a second round of venture capital financing from Kleiner Perkins. The company's first round came from Mayfield Fund.
CloudPhysics operates a cloud-based Software as a Service (SaaS) consisting of what it described as a sophisticated real-time data analytics engine. This knowledgebase, which constantly takes in new data feeds, diagnoses and troubleshoots thousands of issues that might affect the function of a VMware ESX virtual server cluster environment such as incorrectly configured scripts, network configuration errors, and memory and IO utilization issues.
"The administrator has multiple questions, literally thousands of questions that are very well-defined explorations or responses to very well-defined problems," explained Founder and CEO John Blumenthal, who is among the VMware veterans who helped launch CloudPhysics in 2011.
Blumenthal described the service as a big-data repository that collects more than 80 billion pieces of data each day from a variety of sources, ranging from technical blogs to configuration data from customers and other sources. The data is all "anonymized" and used to create patterns that are subsequently analyzed.
Data fed from customer datacenters and other sources are kept anonymous by using sophisticated cryptography to debunk concerns about the privacy and security of data, Blumenthal said. While I didn't dispute the wisdom of those measures, especially with heightened concerns about surveillance, I asked Blumenthal why an organization would be worried about their memory utilization getting into the wrong hands.
"It's more of a policy issue than anything else," Blumenthal said. "When you talk to users, they make extensive uses of SaaS services, including Salesforce.com, where actually the most sensitive data in a corporation is now off-prem in the form of the customer contact list. Usually, in most of our discussions with our users who raise these concerns, they back down from it very quickly when they stop and think it through."
More than 500 enterprises globally tested the service, which is hosted on the Amazon Web Services EC2 service, though Blumenthal said it can easily be moved to another Infrastructure as a Service (IaaS).
"It's not tied to Amazon in any way," Blumenthal said. "Amazon's back-end provides the running infrastructure for compliance and security."
Customers install a virtual appliance on their VMware ESX clusters, which function as an agent. Administrators can discover and troubleshoot hundreds of operational problems using specific analytic components that CloudPhysics calls Cards, available from an app store-type environment also launched this week. In addition to accessing cards that offer pre-configured reports, a customer can create their own with a tool called Card Builder.
The analytics engine is designed to help administrators optimize storage, compute, network and other components using various modeling methods that can address performance and cost benchmarks. A planning component lets administrators simulate the effects of adding new hardware, software and other components.
Given CloudPhysics' roots and dominant installed base, it's not surprising that the inaugural edition is designed for VMware environments. But the company also plans to support other virtual machines, including Microsoft's Hyper-V, Citrix Xen and Linux-based KVM.
CloudPhysics offers a free community edition. For a standard edition with more features and e-mail support, pricing starts at $49 for customers signing a one-year contract or $89 for those who opt to go month by month. An enterprise edition is available for $149/$189 per month and offers telephone support and the full menu of features.
As for Greene's roll, while she's an investor, she also advises CloudPhysics on technical direction, though she doesn't serve in an operational capacity, Blumenthal said.
"She occasionally sits down with us to talk over strategy and helps with team culture development," he said. "She's both an inspiration and an investor to this company."
Posted by Jeffrey Schwartz on 08/16/2013 at 2:12 PM0 comments
Well before Edward Snowden leaked classified information that disclosed, among other things, the PRISM surveillance operation led by the U.S. government's National Security Agency (NSA), the Cloud Security Alliance (CSA) had established mechanisms for service providers to disclose their data-protection practices.
A key initiative was the Security, Trust & Assurance Registry (STAR) Registry, launched by the CSA two years ago, which is where cloud providers like Amazon and Microsoft have provided audited security controls.
Now that Snowden has unleashed a flood of classified information that points to PRISM and the NSA's widespread use of surveillance to thwart terrorism, the CSA has sprung into action, calling attention to its efforts and leading the discussion on the effect of surveillance on cloud security.
The Snowden leaks come just as IT organizations have started to become more comfortable with the notion that data can be securely stored in the public cloud. Concerned the Snowden revelations might have a chilling backlash on cloud deployments, the CSA conducted a survey in late June into early July after the leaks became public. The findings showed 56 percent of respondents outside the United States are less likely to use a domestic cloud provider, while 10 percent have actually canceled a cloud deployment here.
Less than a third of all participants, including domestic participants, believe there is adequate transparency on how often the government accesses their information. That lack of transparency was a recurring topic in the CSA's first-ever town hall panel held Monday.
"Today, there's no mechanism in place for cloud customers, any user organizations that rely on these cloud providers, to know when their data was exposed," said moderator Elad Yoran, VP of finance with the New York City chapter of the CSA and the CEO of Vaultive, an up-and-coming provider of a cloud encryption service. This is an issue Yoran has studied quite intensely for obvious reasons.
Not only is there a lack of transparency by the NSA and other U.S. law enforcement agencies, but many key cloud providers have complained that their hands are tied in that they're restricted in what they're permitted to disclose.
"This is definitely a hot topic for me," said panelist Peter McGoff, general counsel of Box, the popular cloud storage provider. "One thing we look at as a cloud provider, and what we're asking for, is more transparency in the process. We want to be able to communicate to customers at a minimum the numbers of such requests that we get in and what our process is. Right now, it's not quite super clear that we have that flexibility."
McGoff did offer that Box hasn't received an overwhelming number of warrants for enterprise data.
Back in June, after Snowden alleged that Microsoft was giving the NSA a direct line to Outlook.com (formerly Hotmail), SkyDrive and Skype, Microsoft general counsel Brad Smith immediately denied the claim in an extensive blog post.
"Microsoft does not provide any government with direct and unfettered access to our customer's data," Smith stated. "Microsoft only pulls and then provides the specific data mandated by the relevant legal demand."
Microsoft only responds to requests for specific accounts and identities, and governments must serve court orders or subpoenas for account information, Smith added. Microsoft has filed a petition with the court to allow it to disclose more information. "We hope the Attorney General can step in to change this situation," Smith said.
The Obama administration has resisted supporting changes in the disclosure policies, but last week the president proposed that the government should step up its efforts to be transparent. The proposal was vague and opposition from both parties indicated nothing will change in the near term. However, panelists during the hour-long CSA town hall webcast said Obama's proposal was a positive move.
"It's a good first step," Box's McGoff said. "I felt much better with president Obama coming out and putting a bright light on this."
Robert Brammer, a senior advisor to the Internet2 Consortium and CEO of Brammer Technology, agreed. "The review the president has talked about with the intelligence process with one of the objectives to create more transparency in the process will improve the level of dialogue on this subject," he said.
While calling for more transparency, Brammer argued there's a lot of misinformation, if not hysteria, about government surveillance activities. "Some of the emotional and superficial and narrowly based commentary that's come out in the media -- either in the newspapers or Sunday morning talk shows -- frankly makes this problem worse," he said. "We need a substantive dialogue on the issues and not a bunch of emotional sound bites."
One substantive point, Brammer noted, was a whitepaper (PDF) released last week by the Obama administration that lays out how telecommunications providers access and analyze metadata gathered from calling information.
"This information is limited to telephony metadata, which includes information about what telephone numbers were used to make and receive the calls, when the calls took place, and how long the calls lasted," according to the whitepaper's executive summary. "Importantly, this information does not include any information about the content of those calls -- the Government cannot, through this program, listen to or record any telephone conversations."
While Snowden revealed surveillance efforts that were previously not public, much of the concern that has surfaced is old news, added Francoise Gilbert, founder and managing director of IT Law Group, a law firm focused on domestic and international information privacy and security. The U.S. government has had surveillance initiatives in place dating back to the late 1960s, and the Foreign Intelligence Surveillance Act (FISA) was initiated in 1978, Gilbert pointed out during the CSA panel discussion.
"The topic of government access to data is not something new," she said. "There have been many iterations and many amendments to these laws to keep up with technology, technology progress, and there has been a movement for the past two years to amend one of these laws -- the Electronic Communications Privacy Act -- to also bring it to the 21st century."
Gilbert also pointed to due-process requirements such as the Wiretap Act. While critics of the Foreign Intelligence Surveillance Court (FISC), created under FISA, believe the judges rubber-stamp most law enforcement warrants, Gilbert argued U.S. citizens have more protections than those in many foreign countries such as the United Kingdom.
"There is no FISA court -- they just come in and have access to your information," she said of many foreign counties. "In general, the laws I would say are definitely more favorable to the governments in foreign countries, especially in the U.K.," than in the United States.
Perhaps, but there's a growing chorus of critics in the United States who don't view the current laws along with the Patriot Act as very favorable to their privacy. While the government argues its surveillance efforts have thwarted potentially deadly attacks, even the panelists on this week's CSA webcast concurred that the feds are going to have to look at becoming more transparent.
What effect have the disclosures of programs like PRISM had on your plans to use public cloud services? Our sister publication Redmond magazine has fielded a survey to gauge your concerns. I invite you to take the survey, which can be accessed here.
Posted by Jeffrey Schwartz on 08/15/2013 at 10:49 AM0 comments
In a major boost for the VMware-launched Cloud Foundry initiative, IBM this week said it is backing the open source Platform as a Service (PaaS) project. IBM said it will collaborate with Pivotal, the company spun out of VMware, the sponsor of Cloud Foundry.
IBM's decision to join the Cloud Foundry bandwagon gives a major boost to the open source project, and the two said they will work toward establishing a governance model aimed at making Cloud Foundry independent. IBM said Cloud Foundry will provide an open cloud platform for building agile applications that are independent of application development, cloud programming and infrastructures models.
The support for multiple cloud Infrastructure as a Service (IaaS) environments means Cloud Foundry can run on various IaaS clouds, including Amazon Web Services EC2, VMware's vCloud Director and those based on OpenStack. That suits IBM well because it earlier this year committed to OpenStack as the IaaS that will host all of its public, private and hybrid cloud offerings.
"Basically, we see these as very complementary sets of technologies," said Christopher Ferris, an IBM distinguished engineer and the company's CTO for cloud interoperability. "And these communities can potentially collaborate with one another." OpenStack and Cloud Foundry are complementary in that one is IaaS and the other is PaaS, he added in a blog post.
Ferris explained that IBM started working with Pivotal after EMC and VMware spun it off earlier this year, when GE said it was investing in the project. "We have been internally installing it, developing with it and have gotten to a point where we felt it was right to engage the community more openly and let people know what we're up to," Ferris said.
IBM approached Pivotal with the prospect of writing a build pack for its WebSphere Application Server Liberty Core offering as an alternative to the Java build pack that comes by default with Cloud Foundry and includes the OpenJDK bundled in Tomcat. The two organizations collaborated to extend the WebSphere Application Liberty Core, a lightweight version of IBM's pure WebSphere Application Server, as a substitute, Ferris explained.
"We did this because we'd like to be able to have the WebSphere platform be a first-class citizen in Cloud Foundry," Ferris said. "But it was the willingness and the openness of the Pivotal engineering team and leadership to collaborate with us on that. They didn't have to -- they've got their own Spring platform that really competes with WebSphere, but they recognize, too, that an open cloud really needs to be truly open to all. This is one of the reasons that we've been partnering with Pivotal and that we're joining the community and hoping to drive and scale the community itself toward a more open form of self-governance."
Posted by Jeffrey Schwartz on 07/25/2013 at 12:49 PM0 comments
Last week marked the third anniversary of the OpenStack project, an effort led by Rackspace and NASA to create an open source cloud operating environment. OpenStack quickly gained momentum and has evolved as a huge force in cloud computing, with 235 member companies that include AT&T, IBM, Red Hat, Cisco, Hewlett-Packard, Rightscale, Internap and Mirantis.
Attendance at the semi-annual OpenStack Summit continues to increase exponentially, and the OpenStack Foundation claims enterprise adoption is growing, citing examples such as PayPal, Cisco WebEx, Best Buy, Bloomberg, the Gap and HubSpot, as well as the recently reported deployment by Fidelity Investments.
However, some prominent critics have questioned whether OpenStack is gaining meaningful adoption compared with the growth of Amazon Web Services, Google Compute Engine and Microsoft's Windows Azure, among others. Analyst David Linthicum noted in a GigaOM blog post that despite a strong ecosystem and buzz for OpenStack, overall adoption pales in comparison to the growth of AWS.
"While OpenStack, including Rackspace, HP, IBM, and many startups, is clearly the darling of the cloud tech community, the number of installations within traditional IT shops has been lackluster when you consider the expectations that were set," Linthicum wrote.
Perhaps the most noteworthy critic to take the wind out of OpenStack's sails was Randy Bias, the outspoken CEO of Cloudscaling, itself a founding OpenStack member. Bias posted an extensive critique of the existing technical agenda outlining why he believes the OpenStack Foundation's self-described native APIs lack true compatibility with Amazon Web Services APIs, which was the project's original mission. This is especially important now considering the widespread use and dominance of Amazon's cloud services. The OpenStack Nova compute APIs are largely identical with the Rackspace Cloud Servers public cloud service API, not Amazon's, Bias said.
"There is nothing 'native' about the Nova API," Bias wrote. "In fact, calling the Rackspace Cloud Servers API the 'native API' promulgates the notion that there is an OpenStack Nova API that is separate from Amazon's. It's now obvious that the original native API for OpenStack was in fact its AWS EC2 API."
Now that Rackspace has ceded control of the project to the OpenStack Foundation, the new governing board needs to revisit the API stack, according to Bias. "In short, the community controls the direction of the project, and it's time we advocate a public cloud compatibility strategy that is in all our best interests, not just those of a single, albeit substantial, contributor," he wrote. "Failing to make this change in strategy could ultimately lead to the project's irrelevance and death."
The reason, he contends, is that Amazon is far more dominant than any other public cloud Infrastructure as a Service (IaaS). "Embracing Amazon serves the interests of all community members by positioning OpenStack as the best choice for enterprises and SaaS providers that want an ecosystem approach to public cloud, one in which their applications can move to the infrastructure best suited to the job at that time," he said.
Also, despite the lack publicly disclosed information, Bias believes the recently released Google Compute engine IaaS is also growing rapidly. "If others arise, we should debate and evaluate embracing them only when their market position is established," he argued.
Specifically, Bias proposed the OpenStack Foundation do the following:
"1. Embrace major public cloud APIs. GCE, AWS, Azure, and possibly vCloud
"2. Rename the Nova API to the Rackspace Cloud Servers API
"3. Create a new low level API(4) and move to the bridged API model
"4. Expand testing and the work around refstack. Refstack should focus on public cloud interoperability & hybrid cloud
"5. Embrace existing AWS interoperability testing frameworks."
I reached out to officials at the OpenStack Foundation and Rackspace, and while I didn't hear back at press time, I did speak with IBM distinguished engineer and CTO for cloud interoperability Chris Ferris today, primarily to discuss Big Blue's decision to commit to the Cloud Foundry Platform as a Service (PaaS) stack, originated by VMware and spun off to the new Pivotal. Regarding Bias' post, Ferris said he disagreed.
"I've known Randy for a while and he's a very bright guy," Ferris said. "I respect his opinion but I disagree with his conclusion. It's not at all clear to me Amazon has won anything, but more importantly, IBM really believes firmly that open is the right way. Adopting a proprietary API that is in the exclusive control of one vendor is not open. If Amazon wants to contribute that and make it part of OpenStack under the Apache 2 license, maybe we would think about that. But the notion that we should cede the whole thing to Amazon is not my idea of a good idea."
Posted by Jeffrey Schwartz on 07/25/2013 at 12:49 PM0 comments
While Amazon Web Services (AWS) routinely reduces the pricing of its cloud services portfolio, last week's 80 percent slashing of its EC2 Dedicated Instances raised the ire of Rackspace CTO John Engates, who all but said, "You get what you pay for."
Amazon reduced the hourly price of EC2 Dedicated Instances from $10 to $2 per region. In a blog post Tuesday, Engates said not to underestimate the fact that the reductions apply to each region.
In a blog post Tuesday, Engates acknowledged Rackspace isn't looking to beat Amazon on price, but he also warned customers that they should read between the lines.
"A lower unit price doesn't always mean lower costs overall," Engates argued. "Nor does it always deliver value when one considers an apples-to-apples comparison of performance and support."
The way Amazon defines dedicated computing is "at odds" with how everyone else, including Rackspace, defines dedicated instances, according to Engates' missive. Amazon's EC2 Dedicated Instances run in single-tenant hardware dedicated to a single customer account that offer compliance advantages over multitenant instances, specifically for customers who don't want to share those instances with anyone else, Engates noted.
"But they do not provide the true isolation that customers get on dedicated, bare metal servers," he said. "True dedicated servers offer superior performance and customization. And, despite the recent price cuts on EC2 dedicated instances, they still cost more on a total-cost-for-performance basis than do true dedicated servers."
Taking that further, Engates said while Amazon's Dedicated EC2 Instances give customers their own hardware, they're still running in a "dedicated slice" of EC2 and not isolated from the public cloud. Plus, that doesn't apply to requirements for additional block storage. Engates also argued Amazon offers limited customization and its EC2 Dedicated Instances don't offer improved reliability because they're on the same class of servers as the multitenant offering.
The fact that Rackspace also doesn't charge a separate per-region fee can have a significant impact on cost and performance, Engates argued. For example, with Amazon's new pricing, it would cost $1,460 per month for a dedicated per-region fee to continuously run an instance, not including bandwidth and support. By comparison, a dedicated Rackspace server with eight CPU cores, 16 GB of RAM and 146 GB of capacity on two drives costs $538 per month for a managed server, with the option to scale up to 1.5 terabytes for 32 CPU cores.
"A Rackspace customer could get seven dedicated servers (managed with month-to-month pricing) for about $500 a month less than an AWS customer would pay for seven dedicated instances in one region," Engates explained. "In that scenario, Rackspace is $3,656 while AWS is $4,158. What's more, the performance on the dedicated servers would eclipse that of the AWS dedicated instances."
Engates' position notwithstanding, Amazon is still the cloud provider to beat, and its price reductions will undoubtedly give it a boost, making EC2 Dedicated Instances attractive to its existing customers and those looking for commodity services.
Posted by Jeffrey Schwartz on 07/18/2013 at 12:49 PM0 comments
Cloud bottlenecks can have numerous consequences but the most concerning one is their impact on user experience, according to a global survey of 468 IT decision makers published this week.
Nearly two-thirds, or 64 percent, of those surveyed said the impact on the end user experience was the top management concern, compared with 44 percent, who were worried about the effect poor performance can have on revenue. Fifty-one percent pointed to brand reputation as the biggest concern. The study was conducted by Research in Action and commissioned by application performance management vendor Compuware.
Certainly, it's no surprise that any vendor would publish a report saying that performance management and the hidden costs associated with it are concerns (in the latter case, it's a worry for 79 percent of respondents). Nor is it a revelation that anyone in IT is concerned about the impact of bottlenecks from any form of technology. So drilling deeper into the findings, it stood out to me that management is more concerned about the impact on user experience over the risk on revenues.
Does this mean management equates user experience to revenues, or simply that companies are not at the point where they're running apps in the cloud that could have consequences on revenues? I find it hard to believe the an IT decision maker these days would be more worried about a bottleneck's effect on user experience over revenues unless he or she sees the two intertwined.
Among some other findings in the study:
- Eighty-one percent already use or plan to deploy cloud-based e-commerce platforms within the next year.
- Seventy-three percent use outdated methods to track and manage app performance.
- Cloud infrastructure and services will be the No. 1 investment by CIOs in 2013 (12. 5 percent) followed by renegotiating outsourcing contracts (9.6 percent) and big data analytics (9.2 percent).
- Test and backup is the leading area IT will invest in cloud computing over the next year (24.1 percent), followed by building private clouds (17.1 percent), and public and hybrid cloud deployments (15 percent).
Posted by Jeffrey Schwartz on 07/18/2013 at 12:49 PM1 comments
IBM earlier this week said it has closed its acquisition of cloud provider SoftLayer. While IBM hasn't officially disclosed terms, numerous reports have pegged the deal at around $2 billion.
When Big Blue announced the agreement to acquire SoftLayer last month, the company said it would combine the large public cloud provider with its IBM SmartCloud global network, all of which would become part of the company's new cloud services division. IBM has tapped James Comfort to lead the new unit.
In addition to announcing the closing of the SoftLayer acquisition, IBM said it has partnered with a company called Flow Inc. to stream its real-time data analytics into the SmartCloud and SoftLayer cloud platforms. New York-based Flow helps large customers process and use real-time data. The service is designed to let users view information from their mobile devices.
Under the partnership announced with Flow, IBM will provide customers various solutions using the SoftLayer and IBM SmartCloud services. The solutions will enable information to be automatically routed to and from various enterprise apps and analytics services, IBM said. IBM said organizations will be able to create real-time dashboards and mobile apps without requiring IT support.
According to the companies, Flow switched from another undisclosed public cloud provider because it felt the SoftLayer and IBM SmartCloud infrastructure would offer performance improvements and higher levels of flexibility for streaming data in real-time.
"SoftLayer immediately delivered us dramatic performance improvements," said Flow CEO Eric Alterman in a prepared statement. "In addition, with IBM's SmartCloud, we are able to apply improved analytics to the data we stream."
Posted by Jeffrey Schwartz on 07/11/2013 at 12:49 PM0 comments
In its latest round of price cuts, Amazon Web Services this week has reduced the cost of its EC2 Dedicated Instances by up to 80 percent.
Amazon introduced EC2 Dedicated Instances over two years ago. As the name implies, they run on hardware dedicated to a specific customer. The service is designed to let organizations create their own virtual private clouds.
The cloud provider added Dedication Instances at the time to address those customers who were unwilling or unable, due to regulatory or compliance issues, to run their data on multitenant shared instances.
The price cuts apply to dedicated per-region fees and per-instance On-Demand and Reserved Instance fees across all regions, explained AWS evangelist Jeff Barr in a blog post announcing the price cuts.
Under the new pricing plan, Amazon has cut the dedicated per-region fee by 80 percent from $10 to $2 in any "Region" where at least one dedicated instance type is running, according to Barr.
The company cut the hourly rate for its Dedicated On-Demand Instances by up to 37 percent. Barr said for an m1.xlarge Dedicated Instance in the U.S. East (Northern Virginia) Region, the price drops from $0.840 per hour to $0.528 per hour. And Amazon cut the price of Dedicated Reserved Instances by up to 57 percent. Barr said Dedicated Reserved Instances cost 65 percent less than Dedicated On-Demand Services.
The price reductions took effect July 1. Amazon customers can launch Dedicated Instances with the AWS Management Console by selecting a target virtual private cloud and Dedicated Tenancy Option when configuring an instance, according to Barr.
Posted by Jeffrey Schwartz on 07/11/2013 at 12:49 PM0 comments
One of the highlights of latest version of Hyper-V, which arrived with the release of Windows Server 2012 late last year, is its virtual machine live migration capability. Microsoft claims that Hyper-V 3.0 offers faster migrations at speeds of up to 10 Gigabits per second, while allowing IT pros to conduct simultaneous live migrations. IT pros can also now perform live migrations outside a clustered environment.
So how is Microsoft upping the ante on live migration in Windows Server 2012 R2? Following up on a demo at TechEd last month, Microsoft Principal Program Manager Jeff Woolsey showed attendees at the company's Worldwide Partner Conference in Houston Monday just how much faster IT pros can perform live migrations with the new release. In the demo, Woolsey showed an 8 GB virtual machine running SQL Server, which he described as a worst-case scenario for live migration.
In the demo scenario, migrating Windows Server 2012 to a like system takes just under 1 minute 26 seconds, while the Windows Server 2012 R2 Preview performed the same migration in just over 32 seconds. Then using remote direct memory access (RDMA) during the live migration process combined with SMB Direct, it took just under 11 seconds, without utilizing added CPU resources.
"With compression we're taking advantage of the fact that we know the servers ship with an abundance of compute resources, and we're taking advantage of the fact that we know that most Hyper-V servers are never compute bound," Woolsey said during the WPC demo. "So we're using a little bit of that compute resource to actually compress the virtual machine inline during the live migration. This allows us to compress it and it's actually done a lot faster and much more efficiently. All of this is built into Windows Server 2012 R2."
For those testing Windows Server 2012 R2, are you impressed with the improvements to Live Migration in Hyper-V as well as other new capabilities Microsoft is bringing to its hypervisor? Feel free to comment here or write me at [email protected]
Posted by Jeffrey Schwartz on 07/10/2013 at 12:49 PM1 comments