Feeds

'Now we understand what's required to explode a supernova' - NASA

Fortunately it isn't anything we've got lying about

Next gen security for virtualised datacentres

Top space boffins say that the latest readings from an orbiting science instrument have unlocked the secrets of mighty "Type Ia" supernovae, events in which stars blow up with such violence that they destroy solar systems and outshine entire galaxies.

The new data come from NASA's Swift satellite, orbiting the Earth and sending back observations of short-wavelength radiation from Type Ia super-mega-explosions afar off in the deeps of space.

“Now, thanks to unprecedented X-ray and ultraviolet data from Swift, we have a clearer picture of what’s required to blow up these stars," says Stefan Immler, NASA astrophysicist involved in the research.

It's well known among astro-boffins that Type Ia supernovae originate with a remnant star called a white dwarf, which detonates when pushed to a critical mass. Just what's required to happen in the run up to the explosion, however, has been harder to pin down.

Two main scenarios had been considered possible: in one, the white dwarf sucks in and gobbles up matter from a companion normal star, so gaining mass until, overstuffed, it blows up with unimaginable violence. Alternatively, two white dwarfs might collide like vast hypermassive billiard balls leading to a cataclysmic blast.

In between sniffing gamma-ray bursts emitted from faraway black holes (its main task) Swift has been used to probe Ia supernovae. In two separate studies, featuring 60-odd of the superviolent blasts, boffins couldn't find any of the types of X-ray or ultraviolet emissions which would indicate that a giant star had been present at the explosion site - the immolation of such a vast star would cause a noticeable amount of these wavelengths to be produced.

Thus, the scientists conclude, if a companion star is present when Type Ia supernovae kick off, it is generally smaller than our own Sun - indeed, these results could suggest that the dwarf-conkers theory is actually the correct one.

Studies based on the data are set to be published in prestigious astroboffinry journals The Astrophysical Journal Letters and The Astrophysical Journal in April. They can be read in advance of press here and here. ®

Secure remote control for conventional and virtual desktops

More from The Register

next story
Our LOHAN spaceplane ballocket Kickstarter climbs through £8000
Through 25 per cent but more is needed: Get your UNIQUE rewards!
LOHAN tunes into ultra long range radio
And verily, Vultures shall speak status unto distant receivers
NASA to reformat Opportunity rover's memory from 125 million miles away
Interplanetary admins will back up data and get to work
SpaceX prototype rocket EXPLODES over Texas. 'Tricky' biz, says Elon Musk
No injuries or near injuries. Flight stayed in designated area
EOS, Lockheed to track space junk from Oz
WA facility gets laser-eyes out of the fog
Volcanic eruption in Iceland triggers CODE RED aviation warning
Lava-spitting Bárðarbunga prompts action from Met Office
LOHAN Kickstarter breaks NINETEEN THOUSAND of your EARTH POUNDS
That's right, OVER 9,000 beer tokens - and counting
prev story

Whitepapers

Endpoint data privacy in the cloud is easier than you think
Innovations in encryption and storage resolve issues of data privacy and key requirements for companies to look for in a solution.
Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
Advanced data protection for your virtualized environments
Find a natural fit for optimizing protection for the often resource-constrained data protection process found in virtual environments.
Boost IT visibility and business value
How building a great service catalog relieves pressure points and demonstrates the value of IT service management.
Next gen security for virtualised datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.