Feeds

Microsoft brings Hadoop option to SQL Server

Open-source crunchware SQOOPs to conquer

Security for virtualized datacentres

Microsoft customers running SQL Server are getting a taste of really big data processing through an injection of Hadoop.

The company has released early code that will let Microsoft customers plug the open-source Java architecture from Doug Cutting into SQL Server 2008 R2, SQL Server Parallel Data Warehouse for huge data warehouses, as well as the next version of Microsoft's database, which is codenamed Denali.

Hadoop was built by Cutting, who was inspired by Google's MapReduce. It is becoming something of an industry standard for processing huge amounts of data on clustered servers thanks to the fact that its code is open. Hadoop has also been adopted by top-tier web properties including Amazon, Facebook and Twitter.

The industry thinking is that Hadoop can trickle down to customers outside the rarefied circles of serious number-crunchers, where it is used to understand the changing minutiae of millions of users' likes and status updates in order to change services in response. The aim is for Hadoop to find its feet in more mainstream IT.

Microsoft's Research unit has been working on something that sounds remarkably similar to Hadoop, called Dryad, since about 2006. Earlier this year the plan was to "productise" Dryad through integration with SQL Server and its Windows Azure cloud. There have been no updates from Microsoft, but it seems Dryad must now compete for the affections of big-data lovers on SQL Server.

The Microsoft connectors are called Hadoop Connector for SQL Server Parallel Data Warehouse and Hadoop Connector for SQL Server and are available as Community Technology Previews (CTPs).

The connectors are two-way, letting you move data backwards and forwards between Hadoop and Microsoft's database servers

Microsoft said the connectors would let its customers analyse unstructured data in Hadoop and then pull that back into the SQL Server environments for analysis.

Both connectors use SQL to Hadoop (SQOOP) to transfer the data "efficiently" between the Hadoop File System (HDFS) and Microsoft's relational databases. The Parallel Data Warehouse uses PDW Bulk Load/Extract tool for fast import and export of data.

SQL Server PDW customers can get the Hadoop connector from Microsoft while users of the regular SQL Server 2008 R2 can get the code for Hadoop Connector for SQL Server here. ®

New hybrid storage solutions

Whitepapers

Providing a secure and efficient Helpdesk
A single remote control platform for user support is be key to providing an efficient helpdesk. Retain full control over the way in which screen and keystroke data is transmitted.
Top 5 reasons to deploy VMware with Tegile
Data demand and the rise of virtualization is challenging IT teams to deliver storage performance, scalability and capacity that can keep up, while maximizing efficiency.
Reg Reader Research: SaaS based Email and Office Productivity Tools
Read this Reg reader report which provides advice and guidance for SMBs towards the use of SaaS based email and Office productivity tools.
Security for virtualized datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.
Secure remote control for conventional and virtual desktops
Balancing user privacy and privileged access, in accordance with compliance frameworks and legislation. Evaluating any potential remote control choice.