Feeds

IETF plans to NSA-proof all future internet protocols

Standards boffins promise bloody fight for those who seek to sniff private data

Boost IT visibility and business value

The IETF has taken the next small step down the long, long road of protecting user traffic from spooks, snoops and attackers, setting down the basic architectural principle that new protocols should resist monitoring.

It's not going to be a trivial undertaking: practically every layer of the Internet protocol stack has its origins in a more innocent era.

The new document, RFC 7258 (here), formalises the decision reached at the Vancouver IETF plenary in November [video] that pervasive monitoring is an attack on Internet users (and, in fact, “Pervasive Monitoring is an Attack” is the title of the RFC).

Unlike the blithe statements from law enforcement around the world that metadata collection is innocuous, the RFC explicitly includes metadata collection in its list of threats to Internet users, along with the collection of protocol artefacts, application content, active and passive wiretaps, traffic analysis and cryptographic subversion.

The aim of the new RFC, it says, is to record “the IETF community's consensus” and establish “the technical nature of PM.”

However, the RFC also makes the admission that we're never going to beat the spooks and snoops, because traffic necessarily traverses public networks and reveals where it's from and where it's going. “In all cases, there will remain some privacy-relevant information that is inevitably disclosed by protocols.”

Instead, the document states, protocol design in the future should “significantly increase the cost of attacking, force what was covert to be overt, or make the attack more likely to be detected”.

The RFC puts the onus on protocol developers to think about whether they're creating a new risk (or replicating an old one) early in the process: “adequate, early review of architectural decisions including whether appropriate mitigation of PM can be made is important”, because fixing mistakes late in the process is expensive.

The authors also note that the practice of reusing existing technology, while normal developer behaviour, can “significantly impact” how easy it is to monitor traffic in a new protocol. ®

Gartner critical capabilities for enterprise endpoint backup

More from The Register

next story
Microsoft: We plan to CLEAN UP this here Windows Store town
Paid-for apps that provide free downloads? Really
Snowden on NSA's MonsterMind TERROR: It may trigger cyberwar
Plus: Syria's internet going down? That was a US cock-up
Who needs hackers? 'Password1' opens a third of all biz doors
GPU-powered pen test yields more bad news about defences and passwords
e-Borders fiasco: Brits stung for £224m after US IT giant sues UK govt
Defeat to Raytheon branded 'catastrophic result'
Hear ye, young cyber warriors of the realm: GCHQ wants you
Get involved, get a job and then never discuss work ever again
Chinese hackers spied on investigators of Flight MH370 - report
Classified data on flight's disappearance pinched
Microsoft cries UNINSTALL in the wake of Blue Screens of Death™
Cache crash causes contained choloric calamity
prev story

Whitepapers

Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
7 Elements of Radically Simple OS Migration
Avoid the typical headaches of OS migration during your next project by learning about 7 elements of radically simple OS migration.
BYOD's dark side: Data protection
An endpoint data protection solution that adds value to the user and the organization so it can protect itself from data loss as well as leverage corporate data.
Consolidation: The Foundation for IT Business Transformation
In this whitepaper learn how effective consolidation of IT and business resources can enable multiple, meaningful business benefits.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?