By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
TechgoonduTechgoonduTechgoondu
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Search
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Reading: Scientific data crunching on the cloud
Share
Font ResizerAa
TechgoonduTechgoondu
Font ResizerAa
  • Audio-visual
  • Enterprise
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
  • PC
  • Telecom
Search
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Follow US
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Techgoondu > Blog > Enterprise > Scientific data crunching on the cloud
EnterpriseInternetSoftware

Scientific data crunching on the cloud

Aaron Tan
Last updated: June 13, 2014 at 4:47 PM
Aaron Tan
Published: October 6, 2010
7 Min Read
SHARE

Cloud computing is becoming an attractive option for scientists who require massive computing power to run experiments in high-energy physics and astronomy, among other scientific fields.

In 2008, the U.S. National Aeronautics and Space Administration (NASA) launched its Nebula cloud computing platform comprising existing and newly developed open-source software to provide high performance, instantly available computing, storage, and network resources to NASA’s research community. This eliminates the need to construct expensive data centers whenever scientists require additional data processing.

The National Science Foundation in the U.S. is also partnering with Microsoft to provide American researchers free access to Redmond’s Azure cloud computing platform for scientific data crunching. Announced in February this year, the goal of the project is to help researchers cope with exploding volumes of data that need to be analyzed to yield meaningful results.

Outside the U.S., the Amazon EC2 cloud computing platform is starting to be used by science researchers to generate and process large data sets. In the Asia-Pacific region, at least one group of researchers in a new phase of the Belle project has been using Amazon EC2 to augment its own grid computing infrastructure.

The Belle project, which won a 2008 Physics Nobel Prize for its Japanese researchers, had earlier observed large matter-antimatter asymmetry. The project is run by the KEK High Energy Accelerator Research Organization near Tokyo.

According to a KEK press statement on its Nobel Prize win, “symmetry means that a physical situation will be unchanged under certain transformations.

“One of the important examples of broken symmetry arose immediately after the Big Bang. The most basic form of the Big Bang Theory predicts that equal amounts of matter and antimatter should have been created at the early stage of the universe. They should annihilate each other, but there was some matter remained. This is the broken symmetry thought to be responsible for the visible parts of the universe”.

In Belle II, the next phase of the project, researchers will aim to identify the reasons for this asymmetry. According to Professor Martin Servoir from the University of Melbourne in Australia, the amount of data generated in this phase is likely to be 50 times greater than before.

Servoir is among the 343 global researchers from 13 countries including China, Germany, Korea, India, Russia, United States and Japan who are jointly working on the project.

“What we’ll like to do is to go one step further and win a Nobel Prize for ourselves,” Servoir said. “We need to make measurements that are more precise than what we’ve done so far, and discover something that no one has seen before,” he added.

Generating the data required in this new project phase is no small feat, as the researchers would need computing power that’s 50 times more than what they currently have, Servoir said.

The project team is now preparing for the grand experiment, which won’t be up and running until 2014. “Although Moore’s Law is great, there is not enough time to get over the 50-times-more-computing-power problem,” Servoir said.

“We’re looking to find computing resources as quickly and cheaply as we can, and have turned to Amazon EC2 as a means of providing those resources,” he added.

On Amazon EC2, Servoir is able to replicate the project’s grid computing infrastructure on some 250 instances — or units of compute resources — that are fully accessible to his own in-house systems.

Additionally, “storage elements” were also created on EC2 instances to hold temporary data required to generate more data. This helps to ensure efficient use of cloud computing resources which Amazon charges by the hour.

Currently, all the data generated from Amazon EC2 instances are sent back to the project’s own computing systems and not stored on the cloud. This partly alleviates any security issue that most cloud computing adopters are concerned with.

Besides, “the data generated is sufficiently difficult to analyze that we couldn’t imagine anyone else who might be interested in it,” Servoir said.

In any case, security is ensured by using a secure protocol to move data around as well as “a secure version of Linux that is very hard to hack”, Servoir revealed.

“We were also invited to look at the extra security measures available on EC2 which go well above and beyond what we used to have that’s only available in virtualized environments hosted by Amazon.

“That extra layer totally blew any concern out of the water,” he added.

While generating all that scientific data to answer some of life’s biggest questions is laudable, more work needs to be done to address the long-term preservation of scientific data which is increasingly created in diverse, largely-proprietary formats that may not exist forever.

For now, data curation and preservation remains an afterthought in most scientific endeavors. Servoir, for one, acknowledges the problem, but passes the buck to his colleagues specializing in data preservation when quizzed. Data preservation, in fact, should start right from the beginning with data creation, since only researchers directly involved in the project would know the context in which data was created.

As Carole Palmer, director of the Center for Informatics Research in Science and Scholarship at the University of Illinois, noted in ScienceDaily, data curation and preservation “is about doing good science, and ensuring the science is reproducible, but it’s also about finding ways to bring data together from different sources and using them in new ways.”

“To replicate a study or re-use data you have to know where a data set came from and how it’s been processed. Tracking all the context and transformations is part of the curation process,” she said.

Dell to spend US$1 billion on IoT, takes aim at AWS and Microsoft Azure
IBM predicts the future for 2016
Singtel unveils new logo, service improvements
M1 beefs up Internet TV service with MiBox
New anti-scam measures are good but standards needed instead of “best effort” from banks

Sign up for the TG newsletter

Never miss anything again. Get the latest news and analysis in your inbox.

By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Whatsapp Whatsapp LinkedIn Copy Link Print
Previous Article Singapore Android users get paid apps in two weeks
Next Article HTC Desire HD and Desire Z go on sale in Singapore in November
1 Comment

Leave a ReplyCancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Stay Connected

FacebookLike
XFollow

Latest News

Scammers are so successful they even accidentally scam themselves now
Cybersecurity Internet
June 10, 2025
Doom: The Dark Ages review: Future fantastic demon slaying
Gaming
June 10, 2025
Plaud NotePin review: Note-taking made easy with AI
Internet Mobile
June 9, 2025
Can smart grocery carts, biometric payments boost retailers like FairPrice?
Enterprise Internet
June 6, 2025

Techgoondu.com is published by Goondu Media Pte Ltd, a company registered and based in Singapore.

.

Started in June 2008 by technology journalists and ex-journalists in Singapore who share a common love for all things geeky and digital, the site now includes segments on personal computing, enterprise IT and Internet culture.

banner banner
Everyday DIY
PC needs fixing? Get your hands on with the latest tech tips
READ ON
banner banner
Leaders Q&A
What tomorrow looks like to those at the leading edge today
FIND OUT
banner banner
Advertise with us
Discover unique access and impact with TG custom content
SHOW ME

 

 

POWERED BY READYSPACE
The Techgoondu website is powered by and managed by Readyspace Web Hosting.

TechgoonduTechgoondu
© 2024 Goondu Media Pte Ltd. All Rights Reserved | Privacy | Terms of Use | Advertise | About Us | Contact
Join Us!
Never miss anything again. Get the latest news and analysis in your inbox.

Zero spam, Unsubscribe at any time.
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?