Abstract
Lossy compression is widely used to reduce storage and transmission costs in large-scale scientific data, but it inevitably introduces artifacts that may compromise subsequent analysis. To address this issue, we propose a lightweight 3D convolutional architecture with a fixed-scale batch normalization strategy, ensuring stable training and fast inference. We further analyze the trade-offs related to network size and highlight an empirical relationship between the minimum achievable MSE loss and the corresponding training cost. We also validate the generalizability of the trained network. Experimental results on five representative scientific lossy compressors (SZ1, SZ2, SZP, SZX, and ZFP) and datasets from four diverse scientific domains demonstrate that our method consistently improves reconstruction quality: MSE is reduced by one to four orders of magnitude (10-10000×), corresponding to a PSNR gain of 10-40 dB, while keeping the inference time comparable to the compression runtime. A network trained on a single file generalizes well to other files within the same data set. These results highlight the potential of deep learning as an effective post-processing step to improve compressed scientific data.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of 2025 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, SC 2025 Workshops |
| Pages | 314-323 |
| Number of pages | 10 |
| ISBN (Electronic) | 9798400718717 |
| DOIs | |
| State | Published - Nov 15 2025 |
| Event | 2025 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, SC 2025 Workshops - St. Louis, United States Duration: Nov 16 2025 → Nov 21 2025 |
Publication series
| Name | Proceedings of 2025 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, SC 2025 Workshops |
|---|
Conference
| Conference | 2025 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, SC 2025 Workshops |
|---|---|
| Country/Territory | United States |
| City | St. Louis |
| Period | 11/16/25 → 11/21/25 |
Bibliographical note
Publisher Copyright:© 2025 Copyright held by the owner/author(s).
Funding
This work is supported by the Office of Science, Advanced Scientific Computing Research (ASCR) of the U.S. DOE under Contract No. DE-AC02-06CH11357. This work is also supported by NSF OAC-2514036, OAC-2311875 and OAC-2513768. We also acknowledge the computing resources from Argonne LCRC and ALCF.
| Funders | Funder number |
|---|---|
| Bonnie J Addario Lung Cancer Foundation | |
| Advanced Scientific Computing Research | |
| Office of Science Programs | |
| Laboratory Computing Resource Center | |
| U.S. Department of Energy EPSCoR | DE-AC02-06CH11357 |
| National Science Foundation Arctic Social Science Program | OAC-2513768, OAC-2311875, OAC-2514036 |
Keywords
- Artifact Removal
- Convolutional Neural Networks
- Data Restoration
- Error-bounded Compression
- Scientific Data Compression
ASJC Scopus subject areas
- Computer Science Applications
- Control and Systems Engineering
- Hardware and Architecture
- Computational Mechanics
Fingerprint
Dive into the research topics of 'Lightweight CNN-Based Artifact Reduction for Scientific Error-bounded Lossy Compression'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver