Scientific discovery has long been a function of the tools developed and utilised by the pioneers of innovation. While Lippershey was busy patenting his telescope, Galileo pointed his to the heavens, detected craters on the moon, rings around Saturn and moons orbiting Jupiter. In the field of astronomy, observations made with the naked eye became obsolete, and the telescope became the indispensable tool of the trade. Multiple other instruments such as barometers for weather, microscopes for bacteria and cell study, and others, attained similar “vital” status. But in the last 25 or so years, they have all become secondary to the penultimate scientific tool of the 21st century, the supercomputer.
Ever since the mid-1990s, when the use of supercomputers to map the human genome was first proposed, then executed, technology has been the “man behind the curtain” of discovery. Recently the world was captivated by the first ever image of a black hole. To create the image, radio telescopes around the world were linked together to create a virtual single telescope “the size of the earth.” Fascinating science from gifted astrophysicists. However, the data collected took nearly two years to compile, using multiple supercomputers across a number of sites. Such an endeavor could not have been considered without supercomputers.
Today, the ability to connect thousands of inexpensive yet powerful Intel x86 systems together to solve pieces of a problem, then reassemble those pieces into an answer, has introduced High Performance Computing (HPC) to an even greater spectrum of research. Indeed, across the world, Lenovo HPC customers all over the globe are working on solving many of humanity’s greatest challenges in areas such as seismology (Harvard University, US and LRZ, Germany), Lunar surface identification (SciNet, Canada), childhood cancer treatments (LRZ) and continental species distribution modelling (University of Adelaide, Australia). These are but a few of the thousands of projects that rely on the computational muscle of a supercomputer.
Tomorrow, the advent of sensor technology, and its ability to replace multiple scientific instruments, to provide a constant stream of actionable data, and 5G cellular networks that will transmit that data in near-real-time, creates a dilemma not of “how do I get the data?” but of “where do I start?” Data is no longer a boundary, or an obstacle, but a plentiful asset, which presents its own set of challenges.
To deal with this data deluge, scientists will need to adapt. One can easily envision an augmented research environment emerging where Artificial Intelligence (AI) assists in sifting through the data to identify the most compelling pieces upon which researchers focus. HPC systems themselves will evolve. Today, we stand on the doorstep of the exascale era, where systems will be achieve a quintillion (1018) floating point operations per second (FLOPs). That will require some serious technological heft. Compute, storage and networking will all get faster, with greater capacity, but significant advances in cooling technologies, packaging and management are also required. Making those technologies available to HPC customers of all sizes is what Lenovo’s “Exascale to Everyscale” initiative is all about.
Microscopes, telescopes and seismographs will not go away any time soon. There will always be a place in science for tools that have allowed us to “think big.” But now, because of supercomputers, scientists can “think unlimited.” One wonders what Galileo would be able find with a radio telescope and supercomputer.
Product of the Day2 weeks ago
Naspers invests R42-m in public transport
Product of the Day2 weeks ago
Opera launches Hype in SA
People 'n' Privacy2 weeks ago
POPI is NOT coming to get you
People 'n' Issues2 weeks ago
Loyalty points get tax break
Stream of the Day2 weeks ago
E3: What to expect from Ubisoft Forward
Cybersecurity2 weeks ago
Biometrics set to replace passwords
AppDate1 week ago
AppDate: Kaspersky teaches kids digital ethics
Cybersecurity1 week ago
Defend yourself from doxing