It's not every day that a company plunks down $10 billion and the news becomes a sidebar. But that's what happened last week when Hewlett-Packard Co. said it will acquire Autonomy for that amount.
The deal was overshadowed by the news that HP is weighing the sale or spinoff of its PC business, that it was shutting down its webOS hardware business and that it was lowering its revenue outlook for the year. All of that, and questions about the merits of its Autonomy acquisition, led to analyst downgrades and a 20 percent drop in HP's share price Friday (it recovered a bit Monday, gaining 3.6 percent).
The fact that HP is using the bulk of its $13 billion in cash reserves to acquire Autonomy, which will only account for 1 percent of HP's revenues, has not been well received. HP has signaled that it is transforming itself from a consumer and enterprise company to just an enterprise company, following in the footsteps of IBM. Seemingly, the plan is: Get out of low-margin businesses and focus on more profitable niches like software, services and the cloud.
But will Autonomy get HP there? And is the company squandering its cash reserves at that price tag, which is nearly 11 times revenues? CEO Leo Apotheker said on HP's earnings call Thursday that it is worth it. "Autonomy represents an opportunity for HP, for us to accelerate our vision to decisively and profitably lead a large and growing space, which is the enterprise information management space," Apotheker said.
"It also brings HP higher value business solutions that will help customers manage the explosion of information. But as we execute this deal, this will position HP as a leader in the large and growing space. It will complement our existing technology portfolio and enterprise strategy. It will provide differentiated IP for services and extensive vertical capability in key industries. We will provide IPG [Imaging and Printing Group] a base for content management platform. It will, over time, significantly enhance HP's financial profile. And the board believes that the transaction is accretive to HP's non-GAAP earnings in its first full year after completion."
Autonomy's flagship Intelligent Data Operating Layer (IDOL) software, which provides cross-enterprise search and content management, brings together silos of unstructured data including e-mail, text, Web pages, voice and video. It's used for e-discovery, records management, archiving and Web content management, among other things.
HP noted that Autonomy's revenue growth was running at a compounded annual growth rate of 55 percent. Its 25,000 customers include major corporations, law firms and federal agencies such as AOL, BAE Systems, BBC, Bloomberg, Boeing, Citigroup, Coca-Cola, Daimler AG, Deutsche Bank, DLA Piper, Ericsson, FedEx, Ford, GlaxoSmithKline, Lloyds TSB, NASA, Nestlé, the New York Stock Exchange, Reuters, Shell, Tesco, T-Mobile, the U.S. Department of Energy, the U.S. Department of Homeland Security and the U.S. Securities and Exchange Commission.
It also has 400 OEM partners including Symantec, Citrix, HP, Novell, Oracle, Sybase and TIBCO.
Sounds good, but is all that worth $10.3 billion for a company that generated $870 million in revenues last year? Some analysts are questioning that. "The IDOL IP is stagnant. There hasn't been a major release of IDOL in over 5 years," wrote Forrester Research analyst Leslie Owens in a blog post.
"Buyers of its technology though have been less enthusiastic, regularly citing a firm that is arrogant in its dealings with customers, confused roadmap messages, and technology (particularly the core IDOL platform) that is overly complex and expensive to use," wrote Alan Pelz-Sharpe, a principal with The Real Story Group. "On the other side of the equation is HP, a hardware and services firm that has had very little success with software."
Others are more bullish on the deal. "The Autonomy acquisition brings a broad portfolio of information management technology, along with a growing number of search-based applications," IDC said in a research note. "Autonomy is a highly profitable software business, with a solid customer base worldwide and a significant cash reserve -- that is, it should add significantly to HP's margins."
Pointing to Autonomy's most recent earnings call, IDC noted that Autonomy's cloud business represents 62 percent of revenues today and is growing at a clip if 17 percent. "HP's converged infrastructure solutions are another linchpin of its risk management-aware cloud computing strategy," the IDC note said.
Apotheker said one-third of Autonomy's revenues are from its Software as a Service business. Autonomy also recently acquired Iron Mountain Digital for $380 million.
"Autonomy is successfully selling enterprise software into a market in which enterprise spending on technology is growing, even while IT budgets are flat or in decline," said Constellation Research analyst Michael Dortch in a blog post.
Nevertheless, the deal has a lot of people scratching their heads. One doesn't have to look too far to see HP doesn't have a problem making big investments and later changing strategies. While HP claims it will salvage its webOS software business, it seems all but inevitable that the company will be writing off its $1.2 billion outlay in Palm.
Time will tell if HP does better with Autonomy, but it sure seems like it spent a fortune for a company that only has some of the pieces of the puzzle it is trying to stitch together. And it remains to be seen how well those pieces fit.
Posted by Jeffrey Schwartz on August 22, 2011 at 11:58 AM0 comments
Cloud-based business intelligence (BI) startup GoodData has landed $15 million in venture funding, bringing the total it has raised up to $28.5 million.
The funding was led by Andreessen Horowitz, with existing investors General Catalyst Partners, Fidelity Growth Partners and Windcrest Partners also contributing to the latest round.
San Francisco-based GoodData offers BI in a Software as a Service model, providing dashboards, analytics and data warehousing capabilities to businesses as an alternative to traditional on-premises BI platforms.
Since its launch in 2009, GoodData has landed 2,500 customers, including Capgemini, Groupon, Nike and Time Warner Cable. GoodData said that its platform use has grown 500 percent this year. In July the company managed 6,500 unique data marts and produced 2 million reports.
SaaS cloud providers such as Zendesk, Twilio, Pardot and Get Satisfaction offer their own analytics apps that run on the GoodData BI service. GoodData can run analytics against Facebook, Microsoft Dynamics CRM, Salesforce.com and SugarCRM data, among others. The company said it intends to use the additional funding to build new analytics apps and expand its partner network.
GoodData also said it has added two new members to its board of directors: John O'Farrell, general partner at Andreessen Horowitz, and Dave Girouard, president of Google Enterprise.
Posted by Jeffrey Schwartz on August 18, 2011 at 11:58 AM0 comments
To say cloud providers are less than forthcoming on their approach to cloud security would be an understatement. Call it paranoia or prudence, customers are demanding more transparency about security practices before making the leap to the cloud.
The Cloud Security Alliance (CSA) this fall will launch a searchable registry that existing or prospective customers can access free of charge to query how cloud providers are approaching security. Customers will be able to look up cloud providers and review their security practices.
The CSA Security, Trust & Assurance Registry (STAR) aims to document the security controls in place by cloud computing providers, letting users determine how their existing or potential providers are addressing security. STAR allows providers to file reports that document security practices.
"The purpose of the registry is to prod the industry a bit to really be more transparent in their security practices," said Jim Reavis, executive director of the CSA. "We need to have security by transparency. It's really going to create a big mindset shift that while there are definitely a lot of the details about security practices that must be closely held, that in fact to have cloud actually function as a compute utility, we have to have a lot more knowledge about how it works and operates."
The CSA is looking to strike the right balance between transparency and secrecy, but Reavis believes right now it lies too far on the side of secrecy. As a result, it inhibits the adoption of cloud computing and holds back knowledge of what the security practices are of cloud providers.
"It's a high-level type of shift in the mentality and mindset in how we protect our systems, how we disclose things, how we respond to audits, how we do due diligence," Reavis said. "But I think that will have far-reaching impact on the whole of security and compliance and it could even forestall the need for some pretty heavy-handed government regulation of cloud computing, if we are actually are able to show that the industry can self-regulate to a degree and really expose a prudent amount of information about what they're doing. That's the big effect we're trying to get."
STAR is open to all types of cloud providers, which have the option of submitting two different reports that would indicate their compliance:
- The Consensus Assessments Initiative (CAI) is a questionnaire that lets providers document what security controls exist in their IaaS, PaaS and SaaS offerings, based on industry-accepted methods. It consists of 140 questions a customer or auditor might ask of a cloud provider.
- The Cloud Controls Matrix (CCM) is a spreadsheet-based tool of the CSA's recommended security controls across 13 domains.
STAR should be welcome by customers considering cloud providers. But so far, it remains to be seen how many will contribute to the registry. Reavis is confident there will be broad industry participation.
"Under NDA I have seen this documentation that we're asking for from virtually every cloud provider," he said, noting they've had to provide it for their bigger customers. "I think based on the fact that they've already done this work -- and we've had really positive conversations -- we expect most major cloud providers to have this documentation posted very close to our go-live date."
Posted by Jeffrey Schwartz on August 18, 2011 at 11:58 AM0 comments
VMware Inc.'s Cloud Foundry Platform as a Service (PaaS) has signed on several new supporters of its open-cloud effort.
Cloud Foundry, which runs atop VMware's vSphere and vCloud platforms, went into beta in April. It is striving to offer a platform by which customers can develop and deploy applications to multiple clouds.
Now Cloud Foundry has a number of new partners joining in its cause: Canonical, Dell, enStratus and Opscode, all of which are aiming to make it easier to deploy Cloud Foundry-based infrastructures.
"We are announcing collaborations with key industry partners, dramatically expanding the integrated tooling for running Cloud Foundry almost anywhere, and delivering on our portability vision," Cloud Foundry said on its blog.
Here's what the various partners are bringing to Cloud Foundry:
- Canonical: The Ubuntu 11.10 Linux distribution will include Cloud Foundry. The company said it has added client and server deployment tools, enabling both single-node and multi-node PaaS environments.
- Dell: Using its software framework called Crowbar, users have a recognized open-source bare-metal installation package. Cloud Foundry will release a Crowbar "barclamp" that will install and configure Cloud Foundry.
- enStratus: The company has added Cloud Foundry to its service catalog, allowing Cloud Foundry to be deployed on any of the 18 clouds supported by the vendor, which will offer a management and automation platform that manages vSphere, vCloud and Cloud Foundry infrastructures.
- Opscode: The creator of the popular Chef open-source cloud automation system said it will simplify the building of complex Cloud Foundry topologies.
Cloud Foundry is predicting that these new integrations and the Cloud Foundry VMC client will show up on "millions" of developer and server operating systems.
Posted by Jeffrey Schwartz on August 17, 2011 at 11:58 AM0 comments
Are your non-technical friends and family members familiar with the term cloud computing? If they're not, don't despair. Most people aren't, either.
Only 22 percent of consumers are familiar with the term cloud computing, according to a study conducted by NPD Group. But that doesn't mean they aren't using cloud computing to conduct various tasks and activities.
"Whether they understand the terminology or not, consumers are actually pretty savvy in their use of cloud-based applications," said Stephen Baker, NPD's vice president of industry analysis, in a statement. "They might not always recognize they are performing activities in the cloud, yet they still rely on and use those services extensively. Even so, they are not yet ready to completely give up on traditional PC-based software applications."
Seventy-six percent of respondents to the researcher firm's "Digital Software and the Cloud Report" said they have used some sort of cloud service over the past year. Top services were e-mail, tax preparation and online gaming, NPD said.
Despite the emergence of cloud services, 24 percent of those surveyed said they purchased a computer-based software app over the past six months.
Posted by Jeffrey Schwartz on August 11, 2011 at 11:58 AM0 comments
Intuit this week said it will "double down" on its partnership with Microsoft to integrate its Intuit Partner Platform (IPP) with the Windows Azure platform for those looking to build Software as a Service (SaaS) apps.
The two companies had jointly announced in January 2010 that Intuit would name Windows Azure as a preferred platform for cloud application development on the IPP. To facilitate that, they developed and released the Windows Azure SDK for the IPP.
In parallel, though, Intuit had its own Platform as a Service (PaaS) cloud aimed at providing an end-to-end service for partners. Alex Chriss, director of the IPP, said in a blog post that while it helped some partners get to market, it lacked the scale that would appeal to its developer community.
"While we've continued to invest in our native platform, other platforms have moved at lightning speed," Chriss wrote. "These platforms have evolved way past anything we could create and honestly provide a better development experience than we could."
As a result, Intuit said it will no longer offer its native development stack for developers. Instead, the company plans to focus on the top of the stack to make data and services for developers a top priority. Hence, the decision to extend its reach with Azure.
"We believe with additional investment in our Windows Azure SDK for IPP, we can make developing an app on Azure, leveraging QuickBooks data and IPP's go-to-market services, and getting into the Intuit App Center Channel, DROP DEAD EASY," Chriss wrote.
A forthcoming 2.0 release of the SDK will provide support for one-click publishing and authentication of federated IPP and Intuit Anywhere apps hosted on Windows Azure, Microsoft said on the Windows Azure Team blog. The companies did not say when they will release the new SDK.
The two companies have launched a program called the Front Runner for Intuit Partner Platform, where developers can get access to SDKs, content and support; test their applications' compatibility; and then publish apps to the Intuit App Center, a marketplace of small-business applications and services.
"If you're a member of the Front Runner Program (for third party developers) then you'll also benefit from technical and business support that simplifies getting your apps hosted onto Windows Azure, while also integrating with QuickBooks data," wrote Liz Ngo, senior business development manager in Microsoft's Global ISV Group, in a blog post. "Once you are on-boarded, Intuit & Microsoft will promote your application to Intuit's 26 million (and growing) SMB customers."
Posted by Jeffrey Schwartz on August 11, 2011 at 11:58 AM0 comments
Cloud service provider Skytap has come up with a way to automate the launching of servers without requiring third-party management tools.
The company has added "cloud orchestration" to its flagship service, which it says will eliminate the need for software from the likes of CA Technologies, Hewlett-Packard and IBM that manage the sequencing of servers.
When provisioning apps to the cloud, often servers need to be brought up in a particular sequence. For example, a company may want to launch Active Directory first to ensure all user credentials are available, then SQL Server where data is stored, then Microsoft's Team Foundation Server followed by an application server, explained Sundar Raghavan, Skytap's VP of marketing.
"When people move to the cloud, they want the same high-fidelity environment but in our opinion they don't have to buy all that expensive software and have an IT person set it all up," Raghavan said. "So what we are doing is allowing customers to use our self-service UI and create rules by which these startup sequences will occur and in which the shutdown and suspend will occur. Being able to schedule those operating dependencies is very crucial for moving to the cloud."
Skytap has also added a network routing feature that lets users consolidate server environments, letting customers recreate a server hub-and-spoke model in the cloud. That would allow, for example, customers to use centralized servers such as Active Directory without having to replicate it to the nodes.
"The benefit of this is if you're a team manager, you don't have to wait around for IT to set up your environment," Raghavan said. "You can go, consolidate, use our UI to route between the environments, and get going immediately. Reduce your setup errors, [avoid your obsolete machines and reduce] your usage cost. So really it's about consolidating your servers and simplifying your architecture yet having the benefits and economics of scale of cloud computing."
Posted by Jeffrey Schwartz on August 10, 2011 at 11:58 AM0 comments
Amazon Web Services' woes continued on Tuesday as the cloud provider worked to recover from a bolt of lighting that caused power outages in its Dublin, Ireland datacenter over the weekend.
The lighting strike brought down Amazon's EC2 and RDS services, as well as Microsoft's Business Productivity Online Services. Microsoft's outage reportedly lasted several hours on Sunday and has since been restored, the company said on its Twitter feed.
But Amazon Tuesday was still trying to restore its Elastic Block Store (EBS) block storage volumes following an explosion and fire that made the company's backup generators unavailable. The company said on its Service Health Dashboard on Sunday:
"Due to the scale of the power disruption, a large number of EBS servers lost power and require manual operations before volumes can be restored. Restoring these volumes requires that we make an extra copy of all data, which has consumed most spare capacity and slowed our recovery process. We've been able to restore EC2 instances without attached EBS volumes, as well as some EC2 instances with attached EBS volumes. We are in the process of installing additional capacity in order to support this process both by adding available capacity currently onsite and by moving capacity from other availability zones to the affected zone. While many volumes will be restored over the next several hours, we anticipate that it will take 24-48 hours until the process is completed."
The company said some EC2 instances and EBS servers lost power before writes to their volumes were completed. "Because of this, in some cases we will provide customers with a recovery snapshot instead of restoring their volume so they can validate the health of their volumes before returning them to service. We will contact those customers with information about their recovery snapshot."
Vincent Partington, co-founder and CTO of XebiaLabs, was among those who received bad news from Amazon. "Just got an email from Amazon AWS saying they lost some of my data," Partington Tweeted. "I'm OK with an outage now and then but this really blows!"
Apparently, the problem extends beyond the lightning strike. Amazon discovered a bug in the software that cleans up unused snapshots. "During a recent run of this EBS software in the EU-West Region [Dublin], one or more blocks in a number of EBS snapshots were incorrectly deleted," the company said. "We've addressed the error in the EBS snapshot system to prevent it from recurring. We have now also disabled all of the snapshots that contain these missing blocks."
As of Tuesday morning, Amazon said it has delivered recovery snapshots for more than half of the volumes that were affected by the power outage. "We are continuing to make steady progress on creation and delivery of the remaining recovery snapshots," the company said.
If that wasn't enough, Amazon suffered a brief outage Monday night at its Northeast U.S. datacenter in Virginia. While only lasting about two hours, it affected the Web sites of customers including Foursquare, Netflix and Reddit. Amazon attributed that to connectivity issues between instances.
Posted by Jeffrey Schwartz on August 10, 2011 at 11:58 AM0 comments
Intel is investing $30 million to expand its research in cloud and embedded computing technologies.
The money will be used by Intel Labs to fund the creation of its Intel Science and Technology Centers (ISTCs) at Carnegie Mellon University. The ISTCs are aimed at providing collaboration between university researchers and Intel. The results of that collaboration are made publicly available via technical journals and open-source software releases, Intel said.
"These new ISTCs are expected to open amazing possibilities," said Justin Rattner, Intel's CTO, in a prepared statement. "Imagine, for example, future cars equipped with embedded sensors and microprocessors to constantly collect and analyze traffic and weather data. That information could be shared and analyzed in the cloud so that drivers could be provided with suggestions for quicker and safer routes."
Intel said researchers will examine technologies that will impact the cloud, such as built-in application optimization, big data analytics and extending cloud capabilities to the edge of the network and client devices.
Posted by Jeffrey Schwartz on August 09, 2011 at 11:58 AM0 comments
CloudBees, which offers a cloud-based Platform as a Service (PaaS) for Java applications, has secured $10.5 million in Series B financing from Lightspeed Venture Partners.
The startup company, founded last year by former JBoss CTO Sacha Labourey, offers a Java development- and deployment-based PaaS.
Waltham, Mass.-based CloudBees currently offers two core services: [email protected], which provides a Java-based development infrastructure for building and testing of apps, and [email protected], which provides an app server in the cloud for deployment.
Among CloudBees' customers are Cisco, Digg, Lawrence Livermore National Laboratory and Sandia National Laboratories.
The company received $4 million in Series A financing in November from Matrix Partners, which also participated in the current round. CloudBees said it will use the funds for product development and to expand sales and marketing efforts.
Posted by Jeffrey Schwartz on August 08, 2011 at 11:58 AM0 comments
Amazon Web Services has rolled out three new features to its portfolio of cloud services that should appeal to enterprise users by making it easier to extend their datacenters to the cloud.
Today marks the general availability of AWS Virtual Private Cloud (VPC), AWS Direct Connect and new identity federation capabilities.
"With today's launch of Amazon VPC worldwide, AWS Direct Connect and the new IAM federated identity capabilities, enterprises have even more flexibility and control over deploying their workloads to the cloud," said Amazon Web Services VP Adam Selipsky in a statement. "These capabilities provide even more privacy, and along with AWS's existing cloud services allow enterprises to choose the environment that is best suited to each of their workloads."
The new AWS Direct Connect feature allows customers to create connections from their datacenters to an AWS location with a dedicated network link. By bypassing the Internet and creating dedicated network connections, users gain improved privacy and better network and data transfer performance, while seeing increased bandwidth and reduced latency between the customer's datacenter and AWS, according to the company.
Currently, AWS Direct Connect is available at one location, Equinix's colocation facility in Virginia, allowing users to connect to services in AWS's East Coast region. In the coming months, Direct Connect locations will be available in San Jose, Los Angeles, London, Tokyo and Singapore.
The new Amazon VPC offering lets customers provision instances on a private, isolated section of the AWS cloud service. "You can now build highly available AWS applications that run in multiple Availability Zones within a VPC, with multiple (redundant) VPN connections if you'd like," wrote AWS evangelist Jeff Barr in a blog post. "You can even create redundant VPCs. And, last but not least, you can do all of this in any AWS Region."
Amazon said users can define a virtual network topology that closely matches a typical network that an organization might run within their own datacenters. Customers have control over the virtual networking environment including IP address ranges, creation of subnets and configuration of route tables and network gateways.
In his blog post, Barr wrote:
- The VPC is available in multiple Availability Zones in every AWS Region.
- A single VPC can now span multiple Availability Zones.
- A single VPC can now support multiple VPN connections.
- You can now create more than one VPC per Region in a single AWS account.
- You can now view the status of each of your VPN connections in the AWS Management Console. You can also access it from the command line and via the EC2 API.
- Windows Server 2008 R2 is now supported within a VPC, as are Reserved Instances for Windows with SQL Server.
Lastly, AWS rolled out Identity Federation to its Identity and Access Management (IAM) service. With Identity Federation, customers can use existing enterprise identities to access AWS without having to create a new identity. It lets users create temporary security credentials for AWS to let identities from an existing directory such as an LDAP server, to use IAM's access controls.
In a separate blog post, Barr explained how the Identity Federation capability can allow organizations to request temporary security credentials. "Identity federation opens up new use cases for our enterprise customers," he noted. "You can provision temporary security credentials for identities you are responsible for managing, with no limits on the number of identities you can represent or the number of credentials you can obtain."
Posted by Jeffrey Schwartz on August 04, 2011 at 11:58 AM0 comments
Looking to bolster its IT monitoring portfolio, CA Technologies has agreed to acquire Web site monitoring company Watchmouse. Terms were not disclosed.
The remote user monitoring service is a subscription-based offering that monitors the performance and availability of Web sites, services and transaction-level applications from 62 monitoring stations worldwide in over 40 countries.
Watchmouse, based in Utrecht, Netherlands, replicates real-time transactions and creates scripts that help identify issues that include slow response times and issues that could cause performance problems.
"The idea is as applications start to become more and more cloud-delivered, it's hard to instrument from inside," said Lokesh Jindal, senior VP of strategy and business development at Nimsoft, a division of CA that provides IT management solutions for midsized enterprises.
Watchmouse will be offered as an adjunct offering to the company's Nimsoft Monitor solution, allowing customers to ascertain how their Web sites and applications are performing from around the world. The service helps provide root causes of slow performing Web sites.
Watchmouse will also be sold as an add-on to CA's higher-end Application Performance Management (APM) solution, targeted at large enterprises. Watchnouse will offer greater visibility into application performance, whether the app is running in a datacenter, in the cloud or from a managed services provider, CA said.
Posted by Jeffrey Schwartz on August 03, 2011 at 11:58 AM0 comments