When the Vera C. Rubin Observatory comes online in 2025, it will be one of the most powerful tools available to astronomers, capturing huge portions of the sky every night with its 8.4-meter mirror and 3.2-gigapixel camera. Each image will be analyzed within 60 seconds, alerting astronomers to transient events like supernovae. An incredible five petabytes (5,000 terabytes) of new raw images will be recorded each year and made available for astronomers to study.
Not surprisingly, astronomers can’t wait to get their hands on the high-resolution data. A new paper outlines how the huge amounts of data will be processed, organized, and disseminated. The entire process will require several facilities on three continents over the course of the projected ten-year-long survey.
Detailed cut-away render of the telescope model showing the inner workings. Credit: LSST Project/J. Andrew
The Rubin Observatory is a ground-based telescope located high in the Chilean Andes. The observatory’s 8.4-meter Simonyi Survey Telescope will use the highest resolution digital camera in the world, which also includes the world’s largest fish-eye lens. The camera is roughly the size of a small car and weighs almost 2,800 kg (6,200 lbs). This survey telescope is fast-moving and will be able to scan the entire visible sky in the southern hemisphere every four nights.
“Automated detection and classification of celestial objects will be performed by sophisticated algorithms on high-resolution images to progressively produce an astronomical catalog eventually composed of 20 billion galaxies and 17 billion stars and their associated physical properties,” write Fabio Hernandez, George Beckett, Peter Clark and several other astronomers in their preprint paper.
The main project for Rubin Observatory is the Legacy Survey of Space and Time (LSST) and researchers anticipate this project will gather data on more than 5 million asteroid belt objects, 300,000 Jupiter Trojans, 100,000 near-Earth objects, and more than 40,000 Kuiper belt objects. Since Rubin will be able to map the visible night sky every few days, many of these objects will be observed hundreds of times.
Because of the telescope’s repeated observations, the enormous amount of data will help calculate the positions and orbits of all these objects.
Images flow from the Summit Site, where the telescope is located in Chile, to the Base Site and then to the three Rubin Data Facilities which collectively provide the computational capacity for processing the images taken by the Observatory for the duration of the survey. Credit: Vera Rubin Observatory.
Images and data will immediately flow from the telescope to the Base Facility and Chilean Data Access Center in La Serena, Chile and then go to the three Rubin data facilities on dedicated high-speed networks connecting the sites: the French Data Facility CC-IN2P3 in Lyon, France, the UK Data Facility, IRIS network, in the United Kingdom and the US Data Facility and Data Access Center at SLAC National Accelerator Laboratory in California, USA. There is also a Headquarters Site at the Association of Universities for Research in Astronomy (AURA) in Tucson, Arizona, USA.
Once images are taken, they will be processed according to three different timescales: prompt, daily, and annually. The Hernandez et al paper outlines how raw images collected each observing night will be quickly processed (within 60 seconds), and objects that have changed brightness or position will generate and emit alerts for “transient detection.”
Hernandez told Universe Today that for this process, known as Prompt Processing, there will be no proprietary period associated with alerts, and they will be available to the public immediately, since the goal is to quickly transmit nearly everything about any given event, to enable quick classification and decision making. Scientists estimate Prompt Processing could generate millions of alerts per night.
Daily products, released within 24 hours of observation, will include the images from that night. The annual campaigns will reprocess the entire image dataset collected since the beginning of the survey.
Hernandez said that reprocessing the entire image dataset every year has several purposes.
“First, is to use an ever increasing set of images of each patch of the sky to extract information about the celestial objects present in them,” he said via email. “The more images we have, the more information we can extract. Second, with time we will refine our algorithms for extracting that information from images since we will progressively learn about the instrument itself and about the celestial objects.”
The yearly data release will be made available to science collaborations for use in studies in four main science pillars: probing dark matter and dark energy, taking inventory of Solar System objects, exploring the transient optical sky and mapping the Milky Way.
For each data release, there will be raw and calibration images in addition to science-ready images which have been processed with updated scientific algorithms. There will also be catalogs with the properties of all the astrophysical objects detected.
The annual data processing will be run at the three data facilities, with the final dataset assembled at SLAC and made available to astronomers and physicists via the Rubin Science Platform.
“The volume of released data products generated by the annual processing of the accumulated set of raw images is on average 2.3 times the size of the input dataset for that year and is estimated to reach more than one hundred petabytes by the end of the survey,” the astronomers wrote. They also said that over the ten year-long survey the volume of data released for science analysis is estimated to increase by one order of magnitude.
Illustration of the conceptual design of the LSST Science Pipelines for image processing. Credit: Hernandez et al.
The Rubin Observatory will utilize several kinds of data products and services for archiving and dissemination of the data to the various science collaborations. The paper says the Rubin LSST “Science Pipelines” are composed of about 80 different kinds of tasks, which are all implemented on top of a common algorithmic code base and specialized software. There is a feature called the Data Butler, which is the software system that abstracts the data access details (including data location, data format and access protocols).
Hernandez said the Rubin Observatory data will become fully public after two years.
For more details and information, see the Vera Rubin Observatory website.
Paper: Overview of the distributed image processing infrastructure to produce the Legacy Survey of Space and Time