Figures
Abstract
Traditional compression methods struggle to preserve critical mesoscale ocean features like vortices during bandwidth-constrained marine data transmission. To aaddress this limitation, we propose CompressGAN, a novel deep learning framework that transcends conventional approaches reliant on generic image metrics (e.g., peak signal-to-noise ratio, PSNR; structural similarity index, SSIM). The architecture integrates global-local dual discriminators to enforce spatiotemporal coherence of mesoscale vortices, employs dilated convolutions to enhance feature receptive fields without computational overhead, and incorporates vortex recognition rate as a physics-aware evaluation metric. Furthermore, parametric pruning and adaptive quantization strategies are embedded to optimize memory efficiency for shipborne hardware constraints. Validation across multiple ocean reanalysis datasets demonstrates CompressGAN’s superiority at 4 × compression ratios, achieving 91.46% mesoscale eddy identification accuracy (Iden) versus SRGAN (89.71%) and SRResNet (89.82%), while maintaining operational efficiency (148 s/image inference time, 25 GB peak memory). Generalization tests reveal controlled performance degradation: PSNR reduced by 4.2 ± 0.3 dB, SSIM by 0.7126, and Iden by 4.1%, confirming robustness under marine operational scenarios. This work resolves the critical trade-off between vessel-mounted computational limits and real-time ocean data demands, providing a viable pathway for integrated shipboard systems to reconcile multimodal marine data processing with navigation service requirements.
Citation: Ma X, Wan X, Zhang L, Wang D, Dai Z (2025) A lightweight intelligent compression method for fast Sea Level Anomaly data transmission. PLoS One 20(8): e0327220. https://doi.org/10.1371/journal.pone.0327220
Editor: Hui Li, Dalian Maritime University, CHINA
Received: April 5, 2025; Accepted: June 11, 2025; Published: August 18, 2025
Copyright: © 2025 Ma et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All SLA files are available from the AVISO database. AVISO data can be downloaded (open acess) from https://www.aviso.altimetry.fr/en/my-aviso-plus.html.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
1. Introduction
With the rapid development of ocean observation technology, grid ocean data with high spatial and temporal resolutions have become an important basis for modern marine scientific research [1–4]. The wide application of satellite remote sensing, numerical models, and distributed sensor networks has enabled the acquisition frequency and coverage of global marine environmental parameters to reach unprecedented levels. These cubes contain physical parameters such as temperature, salinity, and velocity (in this paper, the sea surface height anomaly is the main research direction). However, the contradiction between the exponential growth of the data volume and the demand for long-distance transmission is becoming increasingly prominent, especially in special scenarios such as deep sea observations and polar research; thus, achieving efficient and reliable data transmission has become a technical bottleneck restricting real-time marine monitoring and collaborative research.
Existing data compression techniques face significant challenges in dealing with grid ocean data. Although conventional lossless compression algorithms (such as GZIP [5] and LZ77 [6]) can maintain data integrity, it is usually difficult for their compression rates to exceed 50%, and the computational complexity increases nonlinearly as the data dimensions increase. Although lossy compression methods (such as the discrete cosine transform and wavelet transform) can improve the compression efficiency [7], they may destroy the unique spatial continuity and physical conservation characteristics of marine data. For example, overcompression of thermohaline profile data leads to spatial gradient distortion of mesoscale vortex structures, which affects the initial field accuracy of ocean circulation models [8]. The special data formats developed in recent years for earth science research (such as NetCDF and HDF5) have improved the storage efficiency through metadata optimization and dimension reorganization [9], but they have not fundamentally solved the bandwidth constraint problem in the transmission stage. The emerging intelligent compression technology provides new possibilities for data volume reduction [10–14]. Moreover, feature extraction methods based on deep learning, such as convolutional autoencoders, can identify nonlinear coupling relationships between ocean parameters and achieve high magnification compression through latent spatial mapping. Studies have shown that such methods can achieve a compression rate of more than 90% for data with strong spatial correlation, such as ocean and atmosphere parameterized grid data, while maintaining the main dynamic characteristics [15–19]. For Marine satellite remote sensing data, Fanelli et al., based on an expanded super-resolution network, achieved an effective resolution improvement for the publicly available abnormal sea surface temperature data and verified the validity of the data from a sub-mesoscale perspective [20]. Nardelli et al. achieved targeted resolution improvement for SLA data by adding an attention mechanism to the existing super-resolution network. They verified it using the MSE metric among multiple baseline models and obtained good model verification effects in the experimental results [21]. Cui et al. used an improved super-resolution model of deep learning gradient constraint embedding method to enhance the resolution of sea surface height data and obtained experimental results with more promising application prospects in Marine engineering [22].
However, data compression requirements and application accuracy in different scenarios are not completely uniform. In current relevant studies, data accuracy evaluation of data compression and decompression processes still utilize image processing values such as the structural similarity index (SSIM), mean squared error (MSE), and peak signal-to-noise ratio (PSNR) as the final evaluation index, rather than a specific parameter of concern to data users. Therefore, in this paper, we propose an intelligent compression framework for lightweight marine grid data for long-distance transmission and ship-based use. By integrating physical prior knowledge and a data-driven model, a multi-scale adaptive quantization mechanism is constructed to effectively control the amount of transmitted data on the premise of ensuring the analyzability of dynamic ocean processes. This method pays special attention to the full-link consistency of the transmission-reconstruction-application of sea level anomaly (SLA) and adopts the forward validation strategy based on the ocean numerical model to ensure that the compressed data can meet the accuracy requirements in different application scenarios. The experimental results show that the comprehensive performance of the framework on typical marine datasets is better than those of the existing mainstream methods and provides a new technical path for building an efficient and reliable marine data transmission system.
2. Data and research area
2.1. Sea level anomaly data
The grid data (mean sea level anomaly, MSLA) were provided by the Archiving, Validation, and Interpretation of Satellite Oceanographic (AVISO) program in France. The dataset combines altimetry data from multiple satellites such as Topex/Poseidon (T/P), Jason-1, European Remote Sensing Satellite (ERS), and Environmental Satellite (ENVISAT). The temporal resolution of the data is 7 days, the spatial resolution is 1° × 1°, and the time span is January 1993 to July 2021. Necessary standard corrections have been made to the data, such as ionospheric delay correction, tropospheric dry and wet component correction, earth and sea tide correction, ocean load tidal correction, polar tide correction, electromagnetic deviation correction, instrument correction, and inverse barometer correction [23–25].
2.2. Ocean reanalysis data
The Japan Coastal Ocean Predictability Experiment 2 Modified (JCOPE2M) is a high-resolution reanalysis dataset for the Northwest Pacific Ocean developed by the Japan Marine Affairs Agency and has a horizontal grid resolution of 1/12° (about 8 km). The vertical full depth is divided into 46 layers, and the temporal resolution is updated daily [26,27]. By assimilating sea surface temperature (SST), sea surface height anomaly (SSHA), and some Argo buoy observations, the dataset achieves accurate reconstruction of the three-dimensional thermohaline structure and flow field of mesoscale vortexes. Previous studies have shown that JCOPE2M has a high reliability in the study of circulation dynamics in the Northwest Pacific Ocean, and it has been widely used in research on the variation mechanism of the Kuroshio extension body and vortex-current interaction.
The hybrid coordinate ocean model (HYCOM) is a numerical model of the global ocean developed by the U.S. Naval Research Laboratory (NRL) that uses a hybrid vertical coordinate system to optimize the simulation of ocean processes at different depths. The base version provides 1/12° (about 9 km) global coverage and a 1/25° (about 4 km) regional enhanced resolution, as well as a time resolution of hourly output and support for operational forecasting of up to 7 days [28–30]. The system can output multi-factor forecast products such as the sea surface height, two-dimensional positive pressure flow field, three-dimensional flow field, temperature, and salinity, and it can integrate historical observation data through reanalysis technology to provide multi-spatiotemporal scale data support for the study of dynamic ocean processes.
The China Ocean Reanalysis and Prediction System (CORA) is a high-precision regional ocean reanalysis dataset developed by the National Oceanic Information Center of China. It focuses on the Northwest Pacific Ocean and the offshore waters of China. This dataset integrates multi-source data such as satellite remote sensing data and buoy observations and adopts advanced data assimilation technology to couple regional ocean models to generate daily three-dimensional marine environment fields from 2000 to 2022. It has a horizontal resolution of 1/12° (about 9 km) and 40 vertical layers and includes core parameters such as the SST, salinity, and ocean currents. Verification has shown that the error of its sea surface temperature is less than 0.5°C, and it has been widely used in the study of ocean dynamic processes and operational forecasting [31].
2.3. Research area
In this study, we focused on the northwest Pacific region (0–60°N, 100–180°E) as the main research area. This area has two high-risk areas that contain the mesoscale vortex [32–35], the Kuroshio extension body area (KE, 30–45°N, 130–180°E), and the zone of the subtropical countercurrent (STCC, 15–28°N, 120–170°E)
3. Methods
3.1. Establishing the sample dataset
In the main research content of this study, data compression was realized using the following super-resolution enhancement process after sampling, so the dataset needed to provide a set of high-resolution images and a set of low-resolution image sample data. To this end, we generated high-resolution images within study area 2.3 with a time resolution of daily SLA data and latitude and longitude spans of 20° (240 × 240 pixels) for the marine reanalysis data and latitude and longitude spans of 40° (40 × 40 pixels) for the AVISO data. The parameters were as follows: × 4 and ×8 data super resolution to enhance the magnification target, corresponding to the formation of low-resolution images (Table 1). The data span was from January 1, 2007, to December 31, 2020. It should be noted that there were a large number of empty values in some of the generated images (corresponding to unmeasurable grid points, usually land or perennial sheltered areas), which greatly affected the training process, we provided a screening mechanism. That is, if the number of null values in a single data grid exceeded 15%, the data were abandoned. A total of 51140 × 2, × 4, and ×8 resolution images were formed in the sample dataset. In the testing stage, the center cropping strategy was adopted to eliminate the boundary effect. For the data preprocessing methods and related threshold Settings of this part, please refer to the reference articles of Zhang et al [36].
In the table, LR represents low-resolution data and HR represents high-resolution data, corresponding to different sampling results of x2, x4 and x8. Additionally, since the initial resolution of the ocean reanalysis data is higher than the original resolution of the SLA data, we only sample the SLA data up to 4 times. This is because after 4 times sampling, the resolution of the SLA data will change to 1° × 1°. This has a significant impact on the identification of mesoscale vortices in the subsequent part of this paper (a considerable number of vortices have radii in the order of 40–100 km).
3.2. Identification of mesoscale eddy
Since the satellite altimetry data were made available to the public in 1992, research on an automatic detection method for mesoscale vortices has attracted much attention. The current mainstream recognition algorithms mainly include the physical parameter method [37], flow field geometry [38], and machine learning-based visual recognition technology [39,40]. However, due to the heterogeneity of the original data pre-processing processes (such as sea surface height anomaly correction standards and gridded interpolation schemes) and the diversity of vortex core criteria (such as the closed contour threshold and streamline curvature parameters), the applications of the different algorithms have significant scene dependence. Therefore, we propose a meso-scale vortex mixing algorithm to reduce the recognition error.
First, the sea surface height field is converted to the geostrophic flow field based on the geostrophic equilibrium relationship (Equation (1), where and
are the meridional and zonal components of the sea surface geostrophic flow field (m/s), and
is the sea surface height (m)). By detecting the symmetric rotation structure of the velocity vector (there are extreme velocity points and the vector is distributed in a clockwise/counterclockwise ring), the candidate vortices are initially screened. Second, the sea surface height field is searched using a closed isoline to exclude the open pseudo-vortex signal. Finally, the joint verification condition is set: the vortex is only judged to be effective (the center location result of the flow field geometry method is preferred) when the space overlap degree (intersection area is ≥ 50% of the area of a single vortex) and the center distance (≤ 1/12°) identified using the two methods meet the standards at the same time. Compared with the original flow field geometry method, contour closure method, and physical parameter method, the expert evaluation rate of the proposed algorithm is 8, 10, and 15% higher, respectively.
3.3. Data super resolution
Super resolution technology aims to reconstruct low-resolution (LR) images into high-resolution (HR) images [37]. Methods for achieving super resolution can generally be divided into two types: image interpolation based methods and deep learning-based methods. This paper primarily proposes a generative adversarial network model that considers both global and local information to improve the super-resolution results of SLA data. The generative adversarial network [41] consists of two main neural networks, a generator and a discriminator. The generator takes a random noise as input and generates a sample that is similar to the real data. The discriminator takes an input sample (which may be a real sample or a generator generated sample) and outputs a probability value, which represents the probability that the sample is a real sample. The core contribution of the super-resolution generative adversarial network (SRGAN) [42] is that it advances the super resolution from pixel recovery to perception enhancement by introducing residual-dense blocks, visual geometry group (VGG) features against loss, and subpixel upsampling to solve the bottlenecks of detail blurring and texture distortion in traditional methods. Its design ideas directly influenced subsequent models (such as the enhanced super-resolution generative adversarial network (ESRGAN) [43] and Real-ESRGAN [44]) and became a milestone work in the field of super resolution. Since the subsequent model needs to multiply the model training computing power and field equipment support for the super-resolution enhancement process of high-resolution images, in this study, we mainly focused on the codec operation built in the lightweight marine ship receiver. Thus, the performance limitation was the most important factor to consider when screening the base model. Accordingly, the base model with a better effect was selected. The original network diagram of the SRGAN is shown in the red dotted box in Fig 1.
The SRGAN replaces the discriminator in the traditional generative adversarial network (GAN) with the VGG-19 network [45–47], changes the pixel loss MSE (Equation (2)) to the perception loss (Equation (3)), and takes the VGG network loss as the Euclidian distance between the features of the generated image and the original image
(Equation (4)). This is done to adjust the edge abrupt effect of the sea surface height field data caused by edge tearing resulting from pure pixel evaluation, which is clearly reflected in the further analysis of the subsequent model evaluation indicators. In addition, considering the characteristics of the sea surface height data and the biased demand for the recognition effect of mesoscale vortices in the scene we are concerned about, the perceptual evaluation range of the regionalized discriminator needs to be further refined and improved. Therefore, based on the SRGAN, in this study, we drew upon the idea of the global and local context codec proposed by Lizuka et al. [48]. A lightweight compression model, namely, CompressGAN, was developed to adapt to the scenarioization. The network consists of one generator and two context discriminators. The model structure of CompressGAN is shown in Fig 2 (inside the red dashed box with additional local discriminators). The SRGAN perception loss is composed of two parts (Equation (3)), namely, content loss
and adversarial loss
.
In Equations (2)–(4), and
describe the dimensions of the respective feature maps within the VGG network,
is the image, and
is the function loss.
Generator loss is a mechanism used to encourage the super-resolution data generated by the generator to reduce the discriminator success rate during adversarial network training. We replaced the adversarial loss function (Equation (5)) of the native GAN network with Equation (6).
In the process of global and local discrimination proposed in this paper, we assign weight to the lost part of the discriminator and dynamically adjust according to the proportion of local blocks in the whole pixel in the optional local discrimination process. In the experiment, we set it to 0.5. Therefore, the perceptive loss of the CompressGAN network can be summarized as the weight sum of the local discrimination loss
and global discrimination loss
(Equation (7)):
In the convolution process of the generator, we introduce an expansive convolution layer (Econv), which increases the receptive field without increasing the number of parameters compared with the normal convolution process, which can improve the local network’s perception of the overall characteristics of the sea surface information during the experiment. The calculation process can be summarized as follows:
where and
are the convolution kernels’ nucleus width and height, η represents the expansion parameters,
and
for the input and output layers of the pixels, respectively,
is a nonlinear transfer function, and
is the convolution kernel matrix
The offset vector is the convolution layer. This becomes the standard convolution formula when η = 1. The distribution of the major convolution layers of the CompressGAN and corresponding characteristic parameters are presented in Table 2.
3.4. Evaluation index
First, we use the mesoscale vortex recognition algorithm described in Section 3.2 as one of the procedural algorithms for the evaluation indexes, and the mesoscale vortex recognition effect after its super resolution improvement is taken as one of the evaluation indexes (Equation 9), where
is the number of successful recognitions, and
is the number of recognitions of mesoscale vortices in the real images. Three conditions must be met when the recognition is successful: (1) the overlap area between the super resolution and the real image recognition scroll exceeds 3/4 of the scope of the scroll contour; (2) the vortex equivalent radius difference does not exceed 20% of the real image; and (3) the vortex center distance does not exceed 10% of the real image radius. Then, we use the PSNR and SSIM indicators, which are commonly used in image super resolution, to improve evaluation indicators [49]. Their calculation formulas are as follows:
In Equation (10), is the mean value of
,
is the mean value of
,
is the variance of
,
is the variance of
, and
is the covariance of
and
.
is the dynamic range of the pixel values. In the case of L = 100,
are set as constant values. In this paper, they are set as 0.01 and 0.03. In Equation (11),
is the total number of data points, and
and
are the SLA values of the pixels corresponding to the true image and super-resolution image, respectively (all normalized to 0–1).
In addition, since we need to deploy relevant equipment on the ship base, there is a huge gap in the performance compared with the shore-based experimental equipment; thus, we utilize the model inference decoding operation speed and the maximum running memory () as evaluation indicators.
4. Experiments
We conducted model training on three different datasets (JCOPE2M, HYCOM, and AVISO) according to the resolution magnification and pixel amount shown in Table 1. To ensure the fairness of the experiment, all of the PSNR and SSIM data were trained in a single channel and were compared with other common super-resolution algorithms and models, including learning class (super-resolution residual neural network (SRResNet), SRGAN) and non-learning class (nearest neighbor, bicubic, AKIMA) models. These models are available for open source downloading at the link presented in the acknowledgements.
4.1. Training details and parameters
In this study, a training environment was constructed based on the NVIDIA GeForce RTX 4060 Ti 16GB graphics card and the i9 13900k central processing unit (CPU). The stratified sampling strategy was adopted to divide the multi-mode super-resolution dataset constructed (Section 3.1). Specifically, the HR and LR image pairs were first paired to ensure the accuracy of the data correspondence, and then, the training set, verification set, and testing set were divided according to the ratio of 8:1:1. In particular, the independent correspondence between the HR and LR images in the three subsets was maintained to avoid the risk of data leakage.
In the model training stage, to adapt to the limitation of the video memory capacity and improve the training efficiency, the dynamic block clipping strategy was adopted: the input image was randomly clipped to a local block of 48 × 48 pixels. The validity of this size selection has been verified by previous experiments, and it can maintain the integrity of local texture details while ensuring that the video memory occupancy of the convolutional operations does not exceed 8 GB. Online data enhancement was performed during the training, including random horizontal/vertical flips (probability of 0.5), ± 15° rotation, and red-green-blue (RGB) channel perturbations (standard deviation of 0.02).
In terms of the optimizer configuration, the Adam [50] optimizer was used and β=(0.9,0.999) dual-momentum parameters were set to implement differentiated learning strategies for the different network components: SRGAN/CompressGAN framework, initial learning rate of 2 × 10−4 (generator), 1 × 10−4 (discriminator), batch size of 32 (HR-LR pairs), and the number of training rounds of 1500 epochs., The iteration strategy was to use an alternate training mechanism. Two rounds of discriminator training were performed after every round of generator update, and the weight coefficients of the adversarial loss and content loss (L1 + L2 Loss) were performed via dynamic balance (initial ratio of 1:100, 10% attenuation per 100 epochs). The SRResNet architecture was utilized [51], the initial learning rate was 1 × 10−5 (determined by grid search), the batch size was 16 (limited by deep residual structure memory requirements), and the number of training rounds was 3000 epochs. The learning scheduling was the cosine annealing strategy. The period length was 500 epochs, the minimum learning rate decayed to 5% of the initial value, and the key hyperparameters were determined via Bayesian optimization for the × 4 super resolution task. When processing the different amplification factors (×2/ × 4/ × 8), parameter adaptive adjustment was implemented, the block size was scaled (amplification factor k corresponds to a block size of 48k × 48k), the learning rate was 1 × 10−5, the batch size was 16, and the training cycle was positive correlation with the amplification factor (×2:1200 epochs, × 4:2000 epochs, and ×4:3000 epochs).
To verify the stability of the training, the early stop mechanism (patience = 50 epochs) was applied to the verification set, and mixed precision training (FP16) was used to accelerate the calculation. In the testing phase, the sliding window strategy was used for the entire graph inference, and weighted fusion of the overlapping regions was used to eliminate the block boundary artifacts. All of the experiments were repeated three times and the average indexes were taken to ensure the reliability of the results. This configuration achieved the inference speed of 12 it/s in the image resolution processing described in this paper, and the peak memory occupancy was controlled to within 7.8 GB.
4.2. Results
4.2.1. Numerical error.
Five hundred sets of data were randomly selected from the testing dataset described in Section 3.1 (resolution in Table 1), and these datasets were used to conduct a numerical error test. The trained SRResNet, SRGAN, CompressGAN, nearest neighbor [52], bicubic [53], and AKIMA [54] (a total of six algorithms) were employed under amplification factors of ×2, × 4, and ×8, and the experimental error fuzzy strategy of averaging three experiments on single data was adopted. Fig 2 presents a comparison of the PSNR, SSIM, and Iden indexes for the × 4 amplification factor. The test results for the × 2 and ×8 amplification factors are presented in Table 3 based on the × 4 amplification factor.
The numerical error gap between the models in Fig 2 under the × 4 amplification factor is mainly between the learning class and the non-learning class, which indicates the superiority of the super-resolution enhancement process of the learning class. Compared with the three deep learning models assessed in this paper, SRResNet was demonstrated to have certain advantages (32.1524) in terms of the PSNR index, which is 0.3633 higher than that of the CompressGAN (31.7891) proposed in this paper and 1.2260 higher than that of the native SRGAN (30.9264). This advantage is also reflected in the SSIM index. The SSIM of the SRResNet is 0.9132, which is 0.0048 higher than that of the CompressGAN (0.9084) proposed in this paper and 0.0231 higher than that of the native SRGAN (0.8901). The reason for this is that the CompressGAN is based on the improvement of the SRGAN, and both of them adjust the loss function in the adversarial network, and thus, it not only has an excellent numerical error performance but also optimizes the overall perception effect of the image. Therefore, the PSNR and SSIM indicators are slightly affected. However, the results show that the gap is not huge, so it has a small disadvantage, and can be acceptable for use in marine applications. Regarding the Iden index, the SRResNet exhibits a significant disadvantage (89.82%). Its Iden value is 0.11% higher than that of the SRGAN (89.71%) and 1.64% lower than that of the CompressGAN (91.46%). This indicates that the local discriminator added by the CompressGAN model proposed in this paper achieves better resolution improvement and detail retention of vortex continuity, and its recognition rate is 1.75% higher than that of the native SRGAN without a local discriminator.
In addition, we carried out ×2, × 4, and ×8 amplification factor tests with a resolution different from that in Table 1 and adopted random size data with a resolution of (128–512 px) × (128–512 px). Three rounds of single data test were conducted, and the average value was taken as the test result (table). The reasoning equipment still used the 4.1 hardware. The overall test results exhibit the same fixed resolution trend, which demonstrates the scalability of the model in terms of the resolution adjustment (Fig 3).
4.2.2. Comparison of transplant decoding effect.
Due to the limitation of the communication capability of ships at sea and the computing capacity of on-board equipment, algorithms with lower performance requirements will be accepted by more small and medium-sized ships. Therefore, for ship-end testing, we adopted a set of general performance on-board computing equipment (the detailed parameters are shown in Table 4) and adopted the same test process described in Section 4.2.2 to test the proposed method on this equipment. However, the data resolution used was adjusted to the needs of the business (project team): 128 × 128 px, and only the × 4 resolution increase ratio was tested.
As can be seen from Table 4, compared with the test equipment in Section 4.1, there is a big difference in the overall performance, and the equipment is the highest value of the average computing power of the on-board equipment in the data collected by the project team after multiple investigations. Therefore, the test results presented in this section may still not meet the requirements of some equipment with too low a performance. For the applied comparison of the test results, we adopted two evaluation methods: (1) whether the inference of 1500 sets of single images can be completed independently and successfully; and (2) the length of the inference decoding time. We do not test the indicators discussed in Section 3.4. The relevant test results are referred to in Section 4.2.1. The testing process was consistent with that in Section 4.2.1, but only the × 4 amplification factor (without cropping) was tested. The testing data were obtained from 1000 pairs of images with a resolution of 1024 × 1024 px, which were randomly selected from the HYCOM and JCOPE2M testing sets. The test results are presented in Table 5.
As can be seen from Table 5, among the five deep learning models tested in this paper, only three tested in Section 4.1 completed the independence test. Although the ESRGAN and Real ESRGAN models have been shown to have super-resolution performances superior to that of SRGAN-based models in a large number of studies, their required hardware conditions have been greatly improved, and this hardware cannot be widely used in shipborne low performance equipment.
4.3. Verification and generalization of the model
The generalization performance is crucial for deep learning network models. Therefore, in this section, we test the generalization of the model from two perspectives of different data sources and different sea area data. First, we test the generalization of the data in different sea areas. We use the sea surface height reanalysis data from the HYCOM data in other mesoscale vortex high occurrence regions of the global ocean as the material for the production of the sample dataset. According to the method described in Section 3.1, 1000 groups of test samples were randomly selected, and 3 rounds of repeated tests were carried out on each single group of data under a × 4 amplification factor. The test results are shown in Fig 4. The parameters of the on-board hardware equipment were the same as those in Section 4.2.2.
Fig 4 shows that the overall effect of each deep learning model in the process of the generalization test decreased. Compared with the results presented in Section 4.2.1, the PSNR index decreased by 3–4 db on average. That of the SRResNet decreased the most, followed by CompressGAN, and that of the SRGAN decreased the least. Compared with the results presented in Section 4.2.1, the SSIM index decreased by 0.5426 on average. That of the SRResNet decreased the most, followed by the SRGAN, and that of the CompressGAN decreased the least. Compared with the results presented in Section 4.2.1, the Iden index decreased by about 2% on average. That of the SRResNet decreased the most, followed by the SRGAN, and that of the CompressGAN decreased the least.
In addition, we conducted generalization tests from the perspective of different data sources. According to the same method in 3.1, we also reanalyzed the sea surface height data of the mesoscale vortex high-incidence area in the CORA dataset, randomly selected 1000 groups of test samples, and conducted 3 rounds of repeated tests on each single group of data under the condition of ×4 amplification factor. The test results are shown in Fig 5.
Fig 5 shows that the overall effect of each deep learning model decreased during the generalization test. Compared with the results presented in Section 4.2.1, the PSNR index decreased by 4–5 db on average. That of the SRResNet decreased the most, followed by the CompressGAN, and that of the SRGAN decreased the least. Compared with the results presented in Section 4.2.1, the SSIM index decreased by 0.7126 on average. That of the SRResNet decreased the most, followed by the SRGAN, and that of the CompressGAN decreased the least. Compared with the results presented in Section 4.2.1, the Iden index decreased by about 4% on average. That of the SRResNet decreased the most, followed by the SRGAN, and that of the CompressGAN decreased the least. These results show that the CompressGAN model has good generalization and portability performances.
5. Conclusions
In this study, a lightweight intelligent compression framework, namely, the CompressGAN, for ship-based applications, was developed to solve the bandwidth limitation problem in ocean grid data transmission. By combining the adversarial network and physical prior knowledge, a global and local dual discriminator mechanism was designed to improve the compression efficiency of SLA data and to optimize the reconstruction accuracy of mesoscale vortex features. Experiments demonstrated that the model achieves a better scroll recognition rate (91.46%) on the JCOPE2M, HYCOM, and AVISO datasets compared to traditional methods (e.g., SRGAN and SRResNet), and it has a good portability for low-performance on-board equipment (reasoning time of 148 s/image and peak memory of 25 GB). Through the introduction of a forward validation strategy based on an ocean numerical model, the practicability of the compressed data in ocean dynamic process analysis was verified, and a new technical path was provided for remote ocean real-time monitoring.
Although the CompressGAN performs well on specific datasets, there are limitations to its ability to generalize. Significant degradation of the model performance (4–5 dB reduction in the PSNR and 4% reduction in the scroll recognition rate) was observed across sea areas (e.g., the Atlantic Ocean and Indian Ocean) and across data sources (e.g., the CORTA reanalysis data), indicating that the current training data have insufficient geographic coverage and physical diversity of. In addition, the adaptability of the model to a high compression ratio (such as x8) is weak, which makes it easy for edge artifacts and vortex distortion to occur in extreme compression scenarios. In addition, there is still a certain dependence on the hardware computing power during shipboard deployment (16 GB video memory is required). Finally, although the vortex recognition rate index is introduced in the evaluation system, the impact of the compressed data on downstream tasks such as three-dimensional warm salt field reconstruction and marine numerical prediction is not fully quantified, which limits the comprehensive evaluation of its application value.
Future research can optimize the model architecture and validation system from multiple dimensions. First, a cross-sea, multi-source heterogeneous ocean dataset needs to be constructed, and the generalization ability of the model can be improved by combining transfer learning and domain adaptive technology. Second, the lightweight network design (such as knowledge distillation and neural architecture search) should be utilized to further reduce the computational complexity to meet the real-time requirements of lower end shipborne equipment. In terms of the physical consistency, dynamic constraints such as vorticity conservation and he geostrophic equilibrium equation can be introduced as loss functions to enhance the physical interpretability of the reconstructed data. In addition, it is necessary to establish an end-to-end evaluation link to quantify the impact of the compressed data on business scenarios such as ocean circulation simulation and disaster warning and to promote the deep coupling of algorithms with ocean numerical models. Finally, an adaptive compression strategy needs to be developed to dynamically adjust the compression ratio and feature retention priority according to the transmission environment and application scenarios so as to achieve intelligent balance between the precision and efficiency.
Acknowledgments
We thank JAMEST for the JCOPE2M data support (https://www.Jamstec.go.jp/jcope/htdocs/distribution/index.html).
We thank AVISO for providing the mesoscale eddy dataset (https://www.aviso.altimetry.fr/en/data/products/value-added-products/global-mesoscale-eddy-trajectory-product. html).
We thank the National Centers for Environmental Information (NCEI) for providing the World Ocean Atlas (WOA) data (https://www.ncei.noaa.gov/products/world-ocean-atlas)
We also thank the National Marine Data Center (https://mds.nmdis.org.cn/) the Japan Oceanographic Data Center (https://www.jodc.go.jp/).
Other scholars and organizations that helped in the research process are also acknowledged.
We thank LetPub (www.letpub.com.cn) for its linguistic assistance during the preparation of this manuscript.
References
- 1. Chelton D, Schlax MG, Samelson RM. Global observations of nonlinear mesoscale eddies. Prog Oceanograph. 2011;91:167–216.
- 2. Zhang Z, et al. Universal structure of mesoscale eddies in the ocean. Geophysical Res Letters. 2013;40(14):3677–81.
- 3. Ji J, et al. Oceanic eddy characteristics and generation mechanisms in the Kuroshio Extension region. J Geophysical Res Oceans. 2018;123(11):8548–67.
- 4. Zhao D, et al. Global chlorophyll distribution induced by mesoscale eddies. Remote Sensing Environ. 2021;254:112245.
- 5. Kösters M, Leufken J, Schulze S, Sugimoto K, Klein J, Zahedi RP, et al. pymzML v2.0: introducing a highly compressed and seekable gzip format. Bioinformatics. 2018;34(14):2513–4. pmid:29394323
- 6.
Kreft S, Navarro G. LZ77-like compression with fast random access. in 2010 Data Compression Conference. IEEE; 2010.
- 7. Said A, Pearlman WA. An image multiresolution representation for lossless and lossy compression. IEEE Trans Image Process. 1996;5(9):1303–10. pmid:18285219
- 8. Tao D, et al. Significantly improving lossy compression for scientific data sets based on multidimensional prediction and error-controlled quantization. In: 2017 IEEE International Parallel and Distributed Processing Symposium (IPDPS). IEEE; 2017.
- 9. Delaunay X, Courtois A, Gouillon F. Evaluation of lossless and lossy algorithms for the compression of scientific datasets in netCDF-4 or HDF5 files. Geoscientific Model Development. 2019;12(9):4099–113.
- 10. Cheng Z, et al. Deep convolutional autoencoder-based lossy image compression. In: 2018 Picture Coding Symposium (PCS). IEEE; 2018.
- 11. Chen Z, et al. Lossy intermediate deep learning feature compression and evaluation. In: Proceedings of the 27th ACM International Conference on Multimedia, 2019.
- 12. Jin S, et al. DeepSZ: A novel framework to compress deep neural networks by using error-bounded lossy compression. In: Proceedings of the 28th International Symposium on High-Performance Parallel and Distributed Computing, 2019.
- 13. Tschannen M, Agustsson E, Lucic M. Deep generative models for distribution-preserving lossy compression. Adv Neural Informat Process Syst. 2018;31.
- 14. Reagan B, et al. Weightless: lossy weight encoding for deep neural network compression. In: International Conference on Machine Learning. PMLR; 2018.
- 15. Baker AH, et al. Evaluating lossy data compression on climate simulation data within a large ensemble. Geoscientific Model Develop. 2016;9(12):4381–403.
- 16. Garcia DC, Queiroz RLd, Fonseca LENd. Lossless and lossy compression of water-column profile data from multibeam echosounders based on image predictions and multiple-context entropy-encoding. Boletim de Ciências Geodésicas. 2023;29(04):e2023010.
- 17. Le H, Tao J. Hierarchical autoencoder-based lossy compression for large-scale high-resolution scientific data. arXiv preprint. 2023:arXiv.2307.04216.
- 18. Kong L, et al. Data loss and reconstruction in sensor networks. In: 2013 Proceedings IEEE INFOCOM. IEEE; 2013.
- 19. Hu C, et al. Secure and efficient data collection and storage of IoT in smart ocean. IEEE Internet Thing J. 2020;7(10):9980–94.
- 20. Fanelli C, Ciani D, Pisano A, Buongiorno Nardelli B. Deep learning for the super resolution of Mediterranean sea surface temperature fields. Ocean Sci. 2024;20(4):1035–50.
- 21. Buongiorno Nardelli B, et al. Super-resolving ocean dynamics from space with computer vision algorithms. Remote Sensing. 2022;14(5):1159.
- 22. Cui J, et al. High-resolution sea surface height reconstruction method based on deep learning gradient constraint embedding method. Ocean Eng. 2025;324:120671.
- 23. Fu L-L, Chelton DB, Zlotnicki V. Satellite altimetry: Observing ocean variability from space. Oceanography. 1988;1(2):4–58.
- 24. Cazenave A, Palanisamy H, Ablain M. Contemporary sea level changes from satellite altimetry: What have we learned? What are the new challenges? Adv Space Res. 2018;62(7):1639–53.
- 25. Abdalla S, et al. Altimetry for the future: Building on 25 years of progress. Adv Space Res. 2021;68(2):319–63.
- 26. Miyazawa Y. The JCOPE ocean forecast system. In: First ARGO Science Workshop, November 12-14, 2003, Tokyo, Japan. 2003.
- 27. Ma X. A mesoscale eddy reconstruction method based on generative adversarial networks. Front Marine Sci. 2024;11:1411779.
- 28. Wallcraft A, et al. Global ocean prediction using HYCOM. In: 2007 DoD High Performance Computing Modernization Program Users Group Conference. IEEE; 2007.
- 29. Chassignet EP, et al. US GODAE: global ocean prediction with the HYbrid Coordinate Ocean Model (HYCOM). Oceanography. 2009;22(2):64–75.
- 30. Trott CB, Metzger EJ, Yu Z. Luzon strait mesoscale eddy characteristics in HYCOM reanalysis, simulation, and forecasts. J Oceanograph. 2023;79(4):423–41.
- 31. Cabanes C, et al. The CORA dataset: validation and diagnostics of ocean temperature and salinity in situ measurements. Ocean Sci Discussions. 2012;9(2):1273–312.
- 32. Qiu B. Kuroshio and Oyashio currents. Ocean Currents: A Derivative of the Encyclopedia of Ocean Sciences. 2001;2:61–72.
- 33. Qiu B, Chen S. Concurrent decadal mesoscale eddy modulations in the western North Pacific subtropical gyre. J Phys Oceanograph. 2013;43(2):344–58.
- 34. Yang G, Wang F, Li Y, Lin P. Mesoscale eddies in the northwestern subtropical Pacific Ocean: Statistical characteristics and three‐dimensional structures. JGR Oceans. 2013;118(4):1906–25.
- 35. Amores A, Melnichenko O, Maximenko N. Coherent mesoscale eddies in the North Atlantic subtropical gyre: 3‐D structure and transport with application to the salinity maximum. J Geophys Res. 2017;122(1):23–41.
- 36. Zhang L, et al. Three-dimensional thermohaline reconstruction of mesoscale eddies under remote sensing observation: from the perspective of deep learning of layer depth sequences with fusion of physical mechanisms. J Sea Res. 2025:102593.
- 37.
Okubo A. Horizontal dispersion of floatable particles in the vicinity of velocity singularities such as convergences. In: Deep sea research and oceanographic abstracts. Elsevier; 1970.
- 38. Nencioli F, et al. A vector geometry–based eddy detection algorithm and its application to a high-resolution numerical model product and high-frequency radar surface velocities in the Southern California Bight. J Atmospheric Ocean Technol. 2010;27(3):564–79.
- 39. Xu G, et al. Oceanic eddy identification using an AI scheme. Remote Sensing. 2019;11(11):1349.
- 40. Nian R, et al. The identification and prediction of mesoscale eddy variation via memory in memory with scheduled sampling for sea level anomaly. Front Marine Sci. 2021;8:753942.
- 41. Goodfellow I, et al. Generative adversarial nets. Adv Neural Information Process Syst. 2014;27.
- 42. Ledig C, et al. Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017.
- 43. Wang X, et al. Esrgan: Enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops, 2018.
- 44. Wang X, et al. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
- 45. Mascarenhas S, Agarwal M. A comparison between VGG16, VGG19 and ResNet50 architecture frameworks for image classification. In: 2021 International conference on disruptive technologies for multi-disciplinary research and applications (CENTCON). IEEE; 2021.
- 46. Mateen M, et al. Fundus image classification using VGG-19 architecture with PCA and SVD. Symmetry. 2018;11(1):1.
- 47. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint. 2014:arXiv:1409.1556.
- 48. Iizuka S, Simo-Serra E, Ishikawa H. Globally and locally consistent image completion. ACM Trans Graph. 2017;36(4):1–14.
- 49. Hore A, Ziou D. Image quality metrics: PSNR vs. SSIM. In: 2010 20th International Conference on Pattern Recognition. IEEE; 2010.
- 50. Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint. 2014:arXiv.1412.6980.
- 51. Lim B, et al. Enhanced deep residual networks for single image super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017.
- 52.
Beyer K, et al. When is “nearest neighbor” meaningful? In: Database Theory—ICDT’99: 7th International Conference Jerusalem, Israel, January 10–12, 1999 Proceedings 7. Springer; 1999.
- 53. Carlson RE, Fritsch FN. Monotone piecewise bicubic interpolation. SIAM J Numeric Analy. 1985;22(2):386–400.
- 54. Akima H. A new method of interpolation and smooth curve fitting based on local procedures. J ACM. 1970;17(4):589–602.