Whenever climate scientists want to analyze data, they need to request it in its messy original format, clean it up and analyzeit. That sort of work takes up valuable time, and so it makes sense that the federal government has started funding efforts to simplify the process.
Speaking at 2013 Hadoop Summit in San Jose on Wednesday, NASA software developer Glenn Tamkin (pictured) explained how he and one of his colleagues have been cooking up a 34-node Hadoop cluster for NASA’s Center for Climate Simulation that can analyze slices of the data in response to end users’ queries. The new architecture could be handy in seeing how the data stacks up in comparison with other data sets used in the U.S. and in other countries.
Tamkin’s team has an 80 TB data set on its hands concerning all kinds of information about climate and atmosphere: winds, clouds, humidity, air and…
View original post 238 more words