This research aims to develop a novel deep learning framework that effectively operates with limited data by integrating principles from inverse problem mathematics and neural networks. Inspired by BCR-Net and SwitchNet, the study introduces a pattern recognition system (PRS) that enhances traditional deep learning approaches through a data-driven, flexible architecture. By combining inverse problem methodologies with advanced neural network modules, this model seeks to overcome the challenges posed by small datasets, providing superior performance in various pattern recognition tasks compared to existing methods.
This study introduces a deep learning model tailored for small datasets by integrating inverse problem (IP) frameworks with neural network modules. Drawing on principles from physics and mathematics, the approach combines BCR-Net and SwitchNet architectures to enhance data augmentation in deep neural networks (DNNs). The model leverages IP’s ability to represent high-dimensional functions flexibly, using data to guide the selection of IP parameters. The effectiveness of the proposed Pattern Recognition System (PRS) is validated through a series of experiments on diverse pattern recognition tasks. The PRS performance is then benchmarked against traditional methods to demonstrate its superior capability in handling limited data scenarios.
The research demonstrates that integrating inverse problem (IP) frameworks with deep neural networks (DNNs) significantly enhances their performance on small datasets. The proposed Pattern Recognition System (PRS), inspired by BCR-Net and SwitchNet, effectively utilizes IP for flexible data augmentation, guiding parameter selection based on available data. Experimental validation across various pattern recognition tasks reveals that PRS outperforms traditional deep learning methods, showcasing superior accuracy and robustness in limited data scenarios. This novel approach underscores the potential of combining advanced neural architectures with mathematical principles to overcome the challenges posed by small datasets in deep learning.
While the proposed Pattern Recognition System (PRS) demonstrates significant improvements in handling small datasets, it may face limitations in generalizability across highly diverse or significantly larger data contexts. The reliance on inverse problem (IP) frameworks and their parameterization might also introduce computational complexity and require domain-specific expertise for optimal implementation. Additionally, the integration of IP with neural networks, though beneficial for data augmentation, might limit scalability when transitioning to real-time applications or larger-scale datasets. Future research should explore broader applicability, the simplification of the model for practical deployment and further refinement to enhance efficiency and scalability across different domains.
The research offers a transformative approach for organizations dealing with small datasets by providing a robust deep learning framework that excels in limited data environments. The integration of inverse problem (IP) frameworks with neural network modules allows the proposed Pattern Recognition System (PRS) to achieve superior performance in tasks that typically struggle with insufficient data. This advancement enables businesses and institutions to leverage their existing small-scale datasets more effectively, enhancing predictive accuracy and decision-making processes. Additionally, the model’s flexibility and enhanced data augmentation capabilities make it suitable for a wide range of applications, from healthcare diagnostics to financial forecasting, where data scarcity is a common challenge.
This research offers significant social benefits by democratizing the power of deep learning for communities and organizations with limited data resources. The proposed Pattern Recognition System (PRS) enables effective AI solutions even in data-scarce environments, fostering inclusivity and equal access to advanced technological capabilities. This can empower small businesses, non-profits and underserved sectors to leverage AI for social good, improving outcomes in healthcare, education and community services where data is often limited. By reducing the dependency on large datasets, this approach can help bridge the digital divide, ensuring more equitable access to the benefits of AI advancements across diverse social landscapes.
This research introduces a groundbreaking approach by seamlessly integrating principles from physics and inverse problem (IP) mathematics into deep learning (DL) models, specifically designed to thrive in small data environments. Unlike traditional DL methods that depend on large datasets, the proposed model – drawing inspiration from BCR-Net and SwitchNet – offers a novel architecture that enhances data augmentation and parameter selection through IP frameworks. This innovative combination not only expands the applicability of DL to data-scarce contexts but also demonstrates superior performance in pattern recognition tasks. The study’s originality lies in its unique fusion of advanced mathematical principles with cutting-edge neural network design, providing significant value to fields constrained by limited data.
1. Introduction
Despite advancements in deep learning (DL) for large-scale data analysis, many AI applications lack access to extensive datasets [1]. This challenge is evident in personalized systems, where DNNs rely on small datasets from search histories, inventory reports, and biometric measurements. Since deep learning focuses on mastering knowledge, personalization demands precise modeling while avoiding overfitting when working with small datasets. For example, in a recent MAPI survey, 58% of research participants reported that the most significant barrier to deploying AI solutions pertained to a paucity of data resources [2]. Therefore, making deep learning of AI work with small data has become a significant predicament. Limited research has been done in this area, even though there is a great need to design intelligent pattern recognition solutions in many application domains [3, 4]. Consequently, in this paper, we design a Pattern Recognition System (PRS) that recognizes patterns with small data using Deep Neural Network (DNN) and IP. Some inverse problems can be viewed as pattern recognition problems [5], thus making them amenable to deep neural networks. Deep learning (Deep Neural Networks, DNNs) has seen a remarkable surge in capabilities, emerging as highly effective instruments for addressing pattern recognition challenges, particularly in domains such as medical image segmentation and classification [6, 7]. Based on deep neural networks, deep learning methods have demonstrated superior performance and increased reliability compared to traditional approaches in a wide range of applications. Nonetheless, achieving these advanced results comes at the cost of requiring more hardware resources [8]. However, the multi-layer structure of DNNs can be adapted to learn effectively even with limited data. To allow IP-enabled PRS to make compelling predictions, we investigate how inverse problems connect to deep learning with small data.
We designed a model inspired by the BCR-Net and SwitchNet to successfully discover the relationship between deep learning and IP. We leverage and adapt a variant of BCR-Net and SwitchNet [9–11]. The adopted version is naturally sparse and computationally efficient. We also design a model that captures the underlying patterns in the region of interest (ROI). As a sparse model, it selectively captures the essential relationships between input and output variables and among the output variables themselves. Constructing a sparse PRS model focuses on learning the crucial dependencies that aid in predicting patterns, disregarding unnecessary ones.
To evaluate the performance of PRS, we applied it to the pattern recognition-based reconstruction of an inhomogeneous object. The BCR and IP-based deep learning model of PRS only requires a small amount of data to recognize patterns, making it an easily deployable model in practice. Furthermore, our experiments demonstrate that the PRS model performs well in effectively uncovering data dependencies, making it a desirable and practically feasible prediction tool due to its low computational requirements. Also, IP-based PRS, due to its sparse nature and ability to capture only the necessary dependencies between the input and output variables, is an interpretable model desirable for many AI models. Combining these two tools in one pattern recognition descriptor shows that IP-based augmentation of our model can improve the average pattern recognition performance.
2. Related work
In order to address the adverse effects of limited data availability and its associated costs, numerous researchers have recently directed their efforts toward developing systems capable of functioning effectively with smaller datasets. Numerous studies have explored the challenges of learning from limited data, and our approach builds on four key areas that have emerged in this domain: transfer learning and autoencoder-based initialization, data augmentation and compact architecture, physics-informed modeling, and meta-learning strategies for few-shot learning.
Transfer learning remains one of the most effective approaches for low-resource scenarios. In this setting, models trained on large-scale datasets are fine-tuned on a smaller target domain, leveraging the reusability of early-layer features across tasks [12, 13]. Closely related, stacked autoencoders have been used to pretrain deep neural networks in small-data regimes, enabling better weight initialization and convergence [14].
Data augmentation and model regularization techniques have also been extensively explored to improve generalization when labeled samples are scarce. Data-augmentation-assisted deep learning has proven effective in classifying and monitoring contexts [15]. At the same time, elastic-net and sparsity-promoting structures such as partially connected DNNs offer competitive results in synthetic and practical tasks [16]. FuCiTNet, a GAN-inspired model, fuses class-inherent transformations to enhance generalization from few examples [17]. Further advancements include the development of feature generalization layers (FGLs) [18], multi-branch and parallel convolutional networks [19–21], and domain-specific augmentation strategies tailored to healthcare and bioinformatics applications [22–25].
Recent efforts have also focused on embedding domain knowledge or structural priors directly into network architectures. Our work is influenced by BCR-Net and SwitchNet, which model integral operators and form the foundation of our architecture. Wavelet-based decomposition has long been used to compress operators and represent functions efficiently in hierarchical bases, a principle that underpins the BCR-Net architecture. More recent work has explored incorporating wavelet structures directly into convolutional neural networks to enhance feature extraction and classification performance [26]. In the broader class of physics-informed neural networks (PINNs), deep architecture has been applied to forward and inverse problems with embedded physical constraints [27]. Hybrid approaches combining residual learning with structured priors have also gained traction [28]. Additional related works include YOLO-based segmentation models [29], transformer-inspired SegFormer architectures [30], and hybrid methods such as PCA-GOA for masked face recognition tasks [31, 32]. Unlike those models, our inverse formulation embeds mathematical operator structure directly into the architecture.
Complementing these efforts are meta-learning and few-shot learning frameworks, which seek to generalize across tasks using minimal data per class. Prototypical Networks operate by embedding inputs into a metric space and classifying based on proximity to class prototypes [33]. Matching Networks extend this paradigm by incorporating attention-based weighting of support-query relationships [34]. MAML (Model-Agnostic Meta-Learning) focuses on learning initialization parameters that enable fast adaptation to new tasks through just a few gradient updates [35]. While powerful, these meta-learning techniques typically require access to large task distributions and extensive episodic training. In contrast, our approach leverages structured inverse operators to support learning under data scarcity without relying on external tasks or data sources.
In summary, prior research demonstrates a range of practical and theoretical strategies for addressing pattern recognition under small-data constraints. Within this context, our approach aims to contribute a unified, inverse-operator-guided deep learning framework that supports generalization in limited-data settings.
3. Inverse problem-based model architecture and dataset design
In this section, we explore the development of a model capable of performing pattern prediction tasks with limited data. Specifically, we aim to reconstruct an unknown, complex function in high-dimensional space by effectively matching available data. Additionally, we provide a comprehensive overview of the data generated, which serves as a means to validate the performance of our model. The PRS model is evaluated on both synthetic datasets and the BBBC real-world dataset to test its real-world applicability.
3.1 PRS model architecture and inverse problem formulation
In this section, we aim to outline the model to predict patterns based on synthetic data generated from regions of interest (ROI). Pattern recognition involves reconstructing an unknown high-dimensional nonlinear function that matches data, which can be formulated as a classic inverse problem. Inverse problems are generally modeled using physics-based principles, which can aid in reconstructing high-dimensional nonlinear functions with available data. However, inverse problems typically do not require significant data compared to DNN. Inverse map framework for DNN can solve pattern recognition problems with small data.
We frame the pattern recognition task as an inverse problem by leveraging BCR-Net and SwitchNet. This study introduces an AI-powered model integrating inverse problem (IP) techniques with neural network architectures for enhanced small-data learning. This novel approach to the pattern recognition problem demonstrates empirically that it performs better than the SVM model, which is the most widely used pattern recognition technique. In the following section, we elaborate on the rationale behind selecting the aforementioned model and highlight the distinctive features of our system.
3.2 Evaluation datasets: synthetic and real-world (BBBC022)
In this paper, we use two groups of synthetic data sets for experimental evaluation of the proposed method. The first of these data sets is based on the inclusion function in which we have two small objects to recognize in the domain; for the experimental evaluation of our proposed model, we employ two sets of synthetic data. The first set revolves around the inclusion function, wherein we have two small objects within the domain that need to be recognized in the domain [−2,2], shown in Figure 1.
The plot is oriented with the horizontal axis ranging from negative 2 to 2 and the vertical axis also ranging from negative 2 to 2. The background of the plot is a yellowish-green color. There are two main regions of interest, both roughly oval-shaped and oriented slightly diagonally. On the left side, centered approximately at coordinates (negative 1, negative 0.25), there is a prominent blue region. On the right side, centered roughly at (1, 0.25), there is a prominent red region. The intensity of the colors represents the magnitude of the values in the field. A vertical color bar is positioned on the right side of the plot. This bar maps the colors to numerical values, ranging from approximately negative 0.2 (dark blue) to 0.2 (dark red). The green and yellow colors in the middle of the color bar correspond to values near zero. The red region corresponds to positive values, and the blue region corresponds to negative values. Note: All the numerical data values are estimated.Two small objects in the domain [−2,2] for pattern recognition experiments. Source: Author’s own creation
The plot is oriented with the horizontal axis ranging from negative 2 to 2 and the vertical axis also ranging from negative 2 to 2. The background of the plot is a yellowish-green color. There are two main regions of interest, both roughly oval-shaped and oriented slightly diagonally. On the left side, centered approximately at coordinates (negative 1, negative 0.25), there is a prominent blue region. On the right side, centered roughly at (1, 0.25), there is a prominent red region. The intensity of the colors represents the magnitude of the values in the field. A vertical color bar is positioned on the right side of the plot. This bar maps the colors to numerical values, ranging from approximately negative 0.2 (dark blue) to 0.2 (dark red). The green and yellow colors in the middle of the color bar correspond to values near zero. The red region corresponds to positive values, and the blue region corresponds to negative values. Note: All the numerical data values are estimated.Two small objects in the domain [−2,2] for pattern recognition experiments. Source: Author’s own creation
The input values for our synthetic data sets are uniformly generated within the range [−2,2]. The corresponding target values are derived using the characteristic functions of the two objects involved in the inclusion function,
In the second data set, a classification problem having two manifolds is shown in Figure 2. The goal is to detect the location of the blue and purple classes. Specifically, we look for the center of these two objects. Table 1 dataset Summary summarizes the characteristics of both datasets used. The inclusion dataset is fully synthetic, while the manifold set comes from the BBBC022 real-world dataset. Each set was divided into 80% training and 20% testing samples.
The plot is contained within a square domain, with the vertical and horizontal axes ranging from negative 2 to 2 in increments of 2. The plot depicts two distinct, colored, oval-shaped regions. The first region, located in the lower-left quadrant, is colored magenta. It is oriented diagonally, with its center approximately at coordinates (negative 1.5, negative 0.25). The second region, in the upper-right quadrant, is colored a pale blue. It is also oriented diagonally, with its center approximately at coordinates (0, 0.5). The two regions are a notable distance from each other and are surrounded by the black background of the plot. Note: All the numerical data values are estimated.Two small manifolds created by synthetic measurement data. Source: Author’s own creation
The plot is contained within a square domain, with the vertical and horizontal axes ranging from negative 2 to 2 in increments of 2. The plot depicts two distinct, colored, oval-shaped regions. The first region, located in the lower-left quadrant, is colored magenta. It is oriented diagonally, with its center approximately at coordinates (negative 1.5, negative 0.25). The second region, in the upper-right quadrant, is colored a pale blue. It is also oriented diagonally, with its center approximately at coordinates (0, 0.5). The two regions are a notable distance from each other and are surrounded by the black background of the plot. Note: All the numerical data values are estimated.Two small manifolds created by synthetic measurement data. Source: Author’s own creation
Dataset summary
| Dataset . | Source . | Data type . | Total samples . | Train/test split . | Description . |
|---|---|---|---|---|---|
| Inclusion set | Synthetic | Numerical | 1,000 | 800/200 | Two-object detection in domain [−2,2] |
| Manifold set | BBBC022 (real) | Microscopy Image | 800 | 600/200 | Cell morphology classification |
| Dataset . | Source . | Data type . | Total samples . | Train/test split . | Description . |
|---|---|---|---|---|---|
| Inclusion set | Synthetic | Numerical | 1,000 | 800/200 | Two-object detection in domain [−2,2] |
| Manifold set | BBBC022 (real) | Microscopy Image | 800 | 600/200 | Cell morphology classification |
The dataset comes from the Broad Bioimage Benchmark Collection (BBBC022) and comprises fluorescence microscopy images of human U2OS cells, capturing the effects of small-molecule treatments on cell morphology. The dataset structure and annotation details are available in the Supplementary Material. It includes high-quality images and manually labeled ground-truth annotations, serving as a gold standard for evaluating machine learning models in image processing and pattern recognition. These expert-verified annotations enable precise assessment of our model’s performance in image-based profiling tasks. In this case, the class priors and class-conditional densities are known, allowing for the straightforward determination of corresponding values for input data [27, 28].
4. Design of neural network as an inverse problem
A core challenge for inverse problems and artificial intelligence (AI) is reconstructing an unknown high-dimensional nonlinear function that matches data. Inverse problems generally have physics-inspired models to represent them, which may help reconstruct high-dimensional non-linear functions with the help of data [29, 30]. However, inverse problems do not require much data compared to DNN; if we use an inverse map framework for DNN, we may be able to solve them with small data. In other words, inverse frameworks inspired by physics help us solve DNN problems with small data. The inverse map architecture in this context serves as a natural data augmentation method.
We first introduce deep learning through the lens of an inverse map architecture, which offers a potential solution to the challenge of requiring large datasets. Unlike many applications in deep learning, such as image processing and pattern recognition, which require a massive amount of data, the required amount for inverse problems may be minimal. We plan to leverage an inverse map to design an NN module based on an integral operator, assemble a DNN from this module, and then use data to train the weights of the inverse map. The IP framework has two major strengths: it works well with limited data, and it is naturally suited for regression problems where the goal is to recover continuous values or parameters from indirect measurements. As a result, pattern recognition can be framed as a supervised regression problem, enabling precise reconstruction even in data-scarce conditions. In this approach, data and parameters work together, enhancing the overall predictive performance. This better architecture, motivated by physics and mathematics, can even work with small data.
To describe a physical situation in mathematical terms, one has to make a certain idealization so that the problem may be amenable to mathematical treatment. Without oversimplification, the model’s success is governed by how close to reality the idealized model is. Various mathematical models have been used in pattern recognition to explain the dynamic behavior of a system. Many involve the formulation and solution of inverse problems, which are generally modeled by considering the model’s components, design variables, and system parameters as deterministic values. In the inverse map framework used here, the dominant factor influencing the accuracy of pattern identification is attributed to the contrast estimator. Here, we briefly describe how the pattern recognition problem of NN can be converted into a suitable inverse problem [5]. In the framework of NN, we have the following integral equation of the first kind
where
Moreover, is an operator from into . The integral equation mentioned above belongs to the first kind of Fredholm integral equations, featuring a kernel of Riesz type. Significantly, the operator in (2.1) can be classified as a convolution integral operator employing the Riesz-type kernel. Initially, it is crucial to note that the equation represents a highly ill-posed problem, yet a unique solution f to equation (2.1) exists and is unique. Furthermore, a logarithmic stability estimate supports the solution’s stability [5]. In order to address the issue of ill-posedness, instead of Equation (2.1), we solve the regularized equation:
where is a regularization parameter and is the identity matrix.
In (2.2), we may consider . As described in [6, 7], first use the SwitchNet module to represent and its adjoint , and then second, represent integral operator by BCRNet to define the neural network architecture as
A simplified overview of the proposed Pattern Recognition System (PRS) model is provided in Figure 3. Input data is first processed by SwitchNet, which approximates the operators and . The output is then passed to BCR-Net, which approximates the inverse operator . This structured design utilizes low-rank approximations and wavelet-based decompositions to produce the predicted output, .
The flowchart has four text boxes arranged horizontally and connected by arrows. The first text box is labeled “Input g of x,” which points to the second text box labeled “SwitchNet (A, A asterisk).” The second text box leads to the third text box labeled “B C R-Net (A asterisk A plus alpha I) superscript minus 1.” The final box on the far right is labeled “Output f of x.”Architecture of PRS model integrating SwitchNet and BCR-Net for inverse operator learning. Source: Author’s own creation
The flowchart has four text boxes arranged horizontally and connected by arrows. The first text box is labeled “Input g of x,” which points to the second text box labeled “SwitchNet (A, A asterisk).” The second text box leads to the third text box labeled “B C R-Net (A asterisk A plus alpha I) superscript minus 1.” The final box on the far right is labeled “Output f of x.”Architecture of PRS model integrating SwitchNet and BCR-Net for inverse operator learning. Source: Author’s own creation
We suggest a combined neural network architecture, incorporating BCRNet, a newly developed neural network architecture inspired by the non-standard form proposed, and SwitchNet, a neural network model equipped with an Inverse Map. Although it is feasible to address this problem using a fully connected network, the parameter counts increase exponentially with the input and output data dimensions. By exploiting the inherent low-rank structure of the inverse conductivity problem defined in (2.2), the introduction of BCRNet and SwitchNet architectures significantly reduces the number of parameters involved, thereby streamlining the training process.
In inverse map → , where and can be represented with a 1D CNN and BCR-Net, respectively.
We now discuss the motivation modules SwitchNet and BCRNet and show how to use the new modules to correctly solve these newest problem applications. All successful neural network (NN) architectures, such as fully connected, convolutional, recurrent, and residual networks, are fundamentally grounded in underlying principles of mathematics and physics. Convolutional neural networks are commonly used for image processing tasks, recurrent neural networks for speech and sequence data, and residual neural networks represent one of the most advanced architectures in deep learning. Fully connected NN has a dense operator behind it, whereas the convolution NN has a translation-invariant operator. All these neural networks have mathematics behind them.
The natural question is, given the integral operator in the inverse problem that was stated in (2.2), what would the neural network architecture be associated with the integral operator? We have the right building blocks if we can build the neural network architecture for this integral operator. When represented as a matrix, the integral operator exhibits a key property [Figure 4]: Its off-diagonal blocks are numerically low-rank, allowing for a sparse representation. This enables these blocks to be replaced with efficient numerical approximations, significantly enhancing computational performance.
The plot is a square, oriented horizontally, with both the horizontal and vertical axes ranging from 0 to 80 in increments of 20. The dominant visual feature is a series of bright, diagonal lines that are oriented from the top-left to the bottom-right corner of the plot. These lines appear to have a consistent width and are separated by darker, parallel spaces. The intensity of these lines is not uniform, with a slightly brighter region in the top-left corner. A vertical grayscale bar is located on the right side of the plot, serving as a legend for the intensity values. The bar shows a scale from 1 (black) to 6 (white), indicating that the brighter regions on the plot correspond to higher values and the darker regions to lower values.Sparse structure of integral operator with low-rank off-diagonal blocks for efficiency. Source: Author’s own creation
The plot is a square, oriented horizontally, with both the horizontal and vertical axes ranging from 0 to 80 in increments of 20. The dominant visual feature is a series of bright, diagonal lines that are oriented from the top-left to the bottom-right corner of the plot. These lines appear to have a consistent width and are separated by darker, parallel spaces. The intensity of these lines is not uniform, with a slightly brighter region in the top-left corner. A vertical grayscale bar is located on the right side of the plot, serving as a legend for the intensity values. The bar shows a scale from 1 (black) to 6 (white), indicating that the brighter regions on the plot correspond to higher values and the darker regions to lower values.Sparse structure of integral operator with low-rank off-diagonal blocks for efficiency. Source: Author’s own creation
Therefore, the plan is to find a convenient sparse-compact representation of the integral operator. Here, we use the BCR-net method inspired by the non-standard wavelet form proposed by Beylkin, Coffman, and Rokhlin [8] for an approximation of a non-linear integral operator given as an matrix utilizing merely tens of thousands of parameters in the following way:
For a given integral operator, the square matrix can be written as a product of three matrices because of the low rank of the off-diagonal blocks. In Figure 5, the first and last matrices represent the redundant wavelet transformation. This one-to-one orthogonal mapping preserves the degrees of freedom while retaining wavelet and scaling coefficients within the transformation. This transformation increases the size by a factor of two. At the same time, the middle matrix, known as the nonstandard form, consists of three diagonally dominant matrices that provide an efficient approximation of the original operator, forming the core idea of BCR-Net.
The diagram is divided into four vertical columns. The first column on the far left contains a large, blank square with sharp corners, representing an operator. The second column contains a vertical, black-and-white striped rectangle, with the diagonal stripes representing a wavelet-based decomposition. The third column contains a series of progressively smaller, nested squares, each containing diagonal lines, representing a hierarchical block-based matrix decomposition. The final column on the far right contains a horizontal, black-and-white striped rectangle with the stripes representing the final compact operator representation. An equals sign is placed between the first and second columns, indicating that the operator is equivalent to the representations shown in the other columns.BCR-Net architecture leveraging wavelet-based decomposition for compact operator representation. Source: Author’s own creation
The diagram is divided into four vertical columns. The first column on the far left contains a large, blank square with sharp corners, representing an operator. The second column contains a vertical, black-and-white striped rectangle, with the diagonal stripes representing a wavelet-based decomposition. The third column contains a series of progressively smaller, nested squares, each containing diagonal lines, representing a hierarchical block-based matrix decomposition. The final column on the far right contains a horizontal, black-and-white striped rectangle with the stripes representing the final compact operator representation. An equals sign is placed between the first and second columns, indicating that the operator is equivalent to the representations shown in the other columns.BCR-Net architecture leveraging wavelet-based decomposition for compact operator representation. Source: Author’s own creation
A natural question arises: why is this linked to neural networks? As shown in Figure 6(a–c), the computation at each scale can be effectively encoded within a neural network architecture.
Diagram (a) is located in the top-left corner and consists of three separate horizontal sections, each containing two or four vertical rectangles with diagonal blue lines. The first section has two rectangles stacked horizontally with diagonal lines slanting from top-left to bottom-right. The second section has four rectangles with diagonal lines slanting from top-left to bottom-right for the top-left, top-right, and bottom-left quadrants. The bottom right quadrant has no diagonal lines. The third section has two rectangles stacked vertically with diagonal lines slanting from top-left to bottom-right. To the right, the diagram depicts a vertical green line on the left, which serves as the input. A large blue arrow points from this input to the main body of the network. The network itself consists of a series of rectangular blocks connected by solid black horizontal lines. The first block is a green rectangle with a vertical label “Conv” and is connected to a series of two smaller, horizontally oriented, green rectangular blocks. These smaller blocks also have a vertical label “Conv.” The final block points to a vertical green line on the right, which represents the output of the network. Diagram (b) is located in the bottom-left and contains three rows of horizontally oriented diagrams. The top row shows a single green rectangle labeled “Conv,” followed by a long horizontal arrow labeled “Convolution,” which leads to another green rectangle labeled “Conv.” The middle row is similar but shows a smaller green rectangle at the output. The bottom row shows a small, light green rectangle, which leads to a darker green rectangle labeled “Conv.” Diagram (c) is in the bottom-right corner and contains three rows of vertically oriented diagrams. The top row shows a green rectangle labeled “Conv” connected to a block labeled “Conn slash ReLU,” which is followed by another block labeled “Conn slash ReLU,” and finally a green rectangle labeled “Conv.” The middle row is a deeper network, with four “Conn slash ReLU” layers stacked between the input and output. The bottom row is similar but with a multiple number of layers represented by a dotted line and small squares.(a) Neural network formulation of BCR, translating integral operator computations into structured layers, (b) Three-layer BCR model leveraging wavelet-based decomposition, (c) Generalized BCR-Net with ReLU activations for enhanced non-linearity and feature extraction. Source: Author’s own creation
Diagram (a) is located in the top-left corner and consists of three separate horizontal sections, each containing two or four vertical rectangles with diagonal blue lines. The first section has two rectangles stacked horizontally with diagonal lines slanting from top-left to bottom-right. The second section has four rectangles with diagonal lines slanting from top-left to bottom-right for the top-left, top-right, and bottom-left quadrants. The bottom right quadrant has no diagonal lines. The third section has two rectangles stacked vertically with diagonal lines slanting from top-left to bottom-right. To the right, the diagram depicts a vertical green line on the left, which serves as the input. A large blue arrow points from this input to the main body of the network. The network itself consists of a series of rectangular blocks connected by solid black horizontal lines. The first block is a green rectangle with a vertical label “Conv” and is connected to a series of two smaller, horizontally oriented, green rectangular blocks. These smaller blocks also have a vertical label “Conv.” The final block points to a vertical green line on the right, which represents the output of the network. Diagram (b) is located in the bottom-left and contains three rows of horizontally oriented diagrams. The top row shows a single green rectangle labeled “Conv,” followed by a long horizontal arrow labeled “Convolution,” which leads to another green rectangle labeled “Conv.” The middle row is similar but shows a smaller green rectangle at the output. The bottom row shows a small, light green rectangle, which leads to a darker green rectangle labeled “Conv.” Diagram (c) is in the bottom-right corner and contains three rows of vertically oriented diagrams. The top row shows a green rectangle labeled “Conv” connected to a block labeled “Conn slash ReLU,” which is followed by another block labeled “Conn slash ReLU,” and finally a green rectangle labeled “Conv.” The middle row is a deeper network, with four “Conn slash ReLU” layers stacked between the input and output. The bottom row is similar but with a multiple number of layers represented by a dotted line and small squares.(a) Neural network formulation of BCR, translating integral operator computations into structured layers, (b) Three-layer BCR model leveraging wavelet-based decomposition, (c) Generalized BCR-Net with ReLU activations for enhanced non-linearity and feature extraction. Source: Author’s own creation
When the input vector is applied, it corresponds to a three-layer network. The first step involves multiplying two diagonal matrices, and assuming these operators are translation-invariant, this operation transforms into a convolution, mapping one channel to two channels. The middle operator acts as a transformation from two channels to two, followed by a final step that reduces the two channels back to one. This structure defines the neural network architecture for a single layer. The complete neural network is constructed by stacking these layers sequentially. The top layer represents the finest.
scale, where wavelet coefficients are computed. These coefficients are combined with scaling coefficients to propagate information to the next layer. The three-layer structure is illustrated in Figure 6(b).
However, for significant problems, one may have many layers. The idea of transforming this into a neural network is to insert a few more internal layers and introduce a non-linearity by inserting ReLUs, so this will give us a nonlinear generalized version of BCR-net shown in 6(c).
Given an integral operator , we may approximate nonlinear operator by partitioning it into n1/2, applying low-rank approximation to each block, and generalizing this via inserting ReLUs [Figure 7].
The topmost diagram shows a grid of 4 by 4 gray squares. An approximately equal sign to its right leads to another 4 by 4 grid of squares, with a small blue rectangle located at the top left of each grid. Another approximately equal sign is at the right of this second grid. The middle diagram shows three separate 4 by 4 grids. The first grid on the left has blue squares arranged in a diagonal pattern. Each blue square has three vertical sections. In the second grid, each cell contains a small rectangle positioned randomly. The third grid on the right has blue squares arranged in a diagonal pattern, similar to the first grid. Here, each blue rectangle has three horizontal sections. To the right of all three grids, a green vertical line is shown. The bottom two diagrams show the neural network architecture. The third diagram from the top begins with a blue arrow pointing to a vertical green line. From this vertical line, multiple black arrows are pointing to a 4 by 4 grid of squares. A horizontal green arrow labeled “Transpose” points from this grid to another 4 by 4 grid of squares. Solid black arrows connect the right side of the second grid to a vertical green line. The final diagram at the very bottom depicts a horizontally oriented neural network architecture. It begins with a vertical green line on the left, which serves as the input. A large blue arrow points from this input to the main body of the network. The network itself consists of a series of rectangular blocks connected by solid black horizontal lines. The first block is a green grid of 4 by 4 squares, which is labeled “Transpose.” This block is connected to a series of four more green blocks, each with a 4 by 4 grid of squares inside. The blocks are labeled “Conv slash ReLU,” “Conv slash ReLU,” “Conv slash ReLU,” and an unlabeled block at the end. The final block is connected to a vertical green line on the right, which represents the output of the network. Small arrows point from the vertical green lines to the grids, labeled “CO,” and from the final grid to the vertical green line, also labeled “CO.”Neural network representation of integral operator with low-rank approximations. Source: Author’s own creation
The topmost diagram shows a grid of 4 by 4 gray squares. An approximately equal sign to its right leads to another 4 by 4 grid of squares, with a small blue rectangle located at the top left of each grid. Another approximately equal sign is at the right of this second grid. The middle diagram shows three separate 4 by 4 grids. The first grid on the left has blue squares arranged in a diagonal pattern. Each blue square has three vertical sections. In the second grid, each cell contains a small rectangle positioned randomly. The third grid on the right has blue squares arranged in a diagonal pattern, similar to the first grid. Here, each blue rectangle has three horizontal sections. To the right of all three grids, a green vertical line is shown. The bottom two diagrams show the neural network architecture. The third diagram from the top begins with a blue arrow pointing to a vertical green line. From this vertical line, multiple black arrows are pointing to a 4 by 4 grid of squares. A horizontal green arrow labeled “Transpose” points from this grid to another 4 by 4 grid of squares. Solid black arrows connect the right side of the second grid to a vertical green line. The final diagram at the very bottom depicts a horizontally oriented neural network architecture. It begins with a vertical green line on the left, which serves as the input. A large blue arrow points from this input to the main body of the network. The network itself consists of a series of rectangular blocks connected by solid black horizontal lines. The first block is a green grid of 4 by 4 squares, which is labeled “Transpose.” This block is connected to a series of four more green blocks, each with a 4 by 4 grid of squares inside. The blocks are labeled “Conv slash ReLU,” “Conv slash ReLU,” “Conv slash ReLU,” and an unlabeled block at the end. The final block is connected to a vertical green line on the right, which represents the output of the network. Small arrows point from the vertical green lines to the grids, labeled “CO,” and from the final grid to the vertical green line, also labeled “CO.”Neural network representation of integral operator with low-rank approximations. Source: Author’s own creation
5. Experimental evaluation, benchmarking, and analysis
In this section, we investigate the efficacy of the PRS learning algorithm introduced in the previous sections. One way to objectively test such a model is to apply it to computational pattern recognition network models for which the mechanisms are entirely known, as described in Section 3.2.
In this study, we denote ‘real object’ by and ‘computed object’ by . Generating in the integral equation (2.1) is essential to evaluate the method. Initially, we define and subsequently compute numerically. Afterward, we utilize the numerical values of as our dataset and extract the underlying pattern within the region of interest.
To ensure transparency and facilitate reproducibility, we present a high-level pseudocode and a formal algorithmic description of the proposed Pattern Reconstruction System (PRS) model. The pseudocode captures the essential conceptual steps of the methodology, abstracted for clarity. At the same time, the accompanying algorithm provides a more detailed procedural breakdown aligned with the mathematical framework introduced in Section 2.
The pseudocode outlines the construction of training pairs using the inverse operator , the generation of forward data, model training, and performance evaluation. In contrast, the formal algorithm incorporates dataset discretization and numerical computation of the integral operator as defined in Equation (2.1), offering a comprehensive view of the PRS training and prediction pipeline.
Both formats are presented side by side in Table 2 to aid in comparative interpretation.
Pseudocode and algorithm for the PRS model
| Pseudocode 1: PRS model . | Algorithm 1: PRS model (formal description) . |
|---|---|
| Input: i. Training dataset ii. Inverse operator | Input: i. Three distinct synthetic datasets and corresponding PRS model ii. Discretization levels N = {N1, ..., Nn} per dataset |
| Output: Reconstructed object pattern | Output: Synthesized decision set |
| 1. Specify target function using and operator | 1. Specify in Equation (2.1) using and |
| 2. Numerically compute | 2. Numerically evaluate the integral in (2.1) to compute , then tune to obtain |
| 3. Train PRS using pairs | 3. Use to compute predicted function |
| 4. Predict for test samples | 4. Compare and reference to generate |
| 5. Compare predicted and true to compute accuracy metrics | 5. Return |
| Pseudocode 1: PRS model . | Algorithm 1: PRS model (formal description) . |
|---|---|
| Input: i. Training dataset ii. Inverse operator | Input: i. Three distinct synthetic datasets and corresponding PRS model ii. Discretization levels N = {N1, ..., Nn} per dataset |
| Output: Reconstructed object pattern | Output: Synthesized decision set |
| 1. Specify target function using and operator | 1. Specify in Equation (2.1) using and |
| 2. Numerically compute | 2. Numerically evaluate the integral in (2.1) to compute , then tune to obtain |
| 3. Train PRS using pairs | 3. Use to compute predicted function |
| 4. Predict for test samples | 4. Compare and reference to generate |
| 5. Compare predicted and true to compute accuracy metrics | 5. Return |
Table 3 presents our experiments’ detailed model configurations and hyperparameter settings. These include the choice of optimizer, learning rate, batch size, regularization coefficient, network architecture, and activation functions. The configurations were selected based on standard practices and empirical tuning to ensure fairness and optimal performance across all models.
Model hyperparameters and architectural configurations used in PRS, SVM, and LSTM experiments
| Parameter . | PRS . | SVM . | LSTM . |
|---|---|---|---|
| Learning rate | 0.001 | N/A | 0.001 |
| Batch size | 64 | N/A | 64 |
| Epochs | 100 | N/A | 100 |
| Regularization coefficient () | 0.01 | 0.1 (C parameter) | 0.001 |
| Architecture | 3-layer (inverse mapping + dense) | Linear kernel | 2-layer LSTM + dense |
| Activation function | ReLU | N/A | Tanh |
| Parameter . | PRS . | SVM . | LSTM . |
|---|---|---|---|
| Learning rate | 0.001 | N/A | 0.001 |
| Batch size | 64 | N/A | 64 |
| Epochs | 100 | N/A | 100 |
| Regularization coefficient () | 0.01 | 0.1 (C parameter) | 0.001 |
| Architecture | 3-layer (inverse mapping + dense) | Linear kernel | 2-layer LSTM + dense |
| Activation function | ReLU | N/A | Tanh |
The first group of experiments was carried out for synthetic data sets on recognizing two small objects in the domain . The reconstruction closely approximates α the inclusion domain, providing a highly accurate representation [Figure 8]. Based on the presented reconstructions, the numerical experiments mentioned earlier have demonstrated an intense match between the predicted and actual inclusion domains. The model generated each inclusion domain with an accuracy of 92.67%.
On the left is a colormap plot. It shows a rectangular grid with horizontal and vertical axes ranging from negative 2 to 2. It has a greenish-yellow background with two prominent circular-like regions: one in blue on the left and another in red on the right. The blue region is centered roughly at (negative 1, negative 0.25) and the red one at (1, 0.25). A color bar to the right of this plot indicates the scale, with deep blue representing negative values around negative 0.2 and deep red representing positive values up to 0.2. The colors transition through shades of light blue, green, and orange. On the right is a binary plot of the same domain. This plot also has horizontal and vertical axes ranging from negative 2 to 2. It's a black background with two white octagonal shapes. These shapes are located in the same general positions as the colored regions in the left plot. Both octagons have a similar size, and their flat sides are oriented parallel and perpendicular to the axes.Reconstructed inclusion domain (synthetic data). Source: Author’s own creation
On the left is a colormap plot. It shows a rectangular grid with horizontal and vertical axes ranging from negative 2 to 2. It has a greenish-yellow background with two prominent circular-like regions: one in blue on the left and another in red on the right. The blue region is centered roughly at (negative 1, negative 0.25) and the red one at (1, 0.25). A color bar to the right of this plot indicates the scale, with deep blue representing negative values around negative 0.2 and deep red representing positive values up to 0.2. The colors transition through shades of light blue, green, and orange. On the right is a binary plot of the same domain. This plot also has horizontal and vertical axes ranging from negative 2 to 2. It's a black background with two white octagonal shapes. These shapes are located in the same general positions as the colored regions in the left plot. Both octagons have a similar size, and their flat sides are oriented parallel and perpendicular to the axes.Reconstructed inclusion domain (synthetic data). Source: Author’s own creation
In the second set of experiments, we aimed to detect the centers of the blue and purple classes in the classification problem illustrated in Figure 2. In Figure 8, the center of the location of the objects was well produced, and the proposed PRS model could distinguish between the two objects. To optimize efficiency, the key factors to adjust in Eq. (2.1) are the regularization parameter and the number of grid points. These parameters play a crucial role in the overall performance of the process.
To validate the reconstruction ability of the proposed model, we used the above-described synthetically generated data whose distributions are known (Figure 9(a) and (c)). The results of this experiment are presented in Figure 9(b) and (d). The PRS model demonstrated strong and consistent performance across all three datasets SOD, SMD, and MMD, as defined in Table 4, spanning tasks from small-object classification to complex multi-manifold recognition. To comprehensively evaluate PRS, we compared it with two established baseline models, SVM and LSTM, using seven metrics: Average Squared Error (ASE), Misclassification Rate (MR), Kolmogorov–Smirnov statistic (KS), Precision, Recall, F1-score, and ROC-AUC.
The plots are arranged in a 2 by 2 grid. Plot (a) shows a black background with two oval-shaped colored regions. The first oval is a light blue color, oriented horizontally, and is centered at approximately (0, 0.5). The second oval is magenta, oriented diagonally, and is centered at roughly (negative 1.5, negative 0.25). The ovals are of a similar size. Plot (b) is a mostly black plot with two small white dots. One dot is located at approximately (negative 1.5, negative 0.25), corresponding to the center of the magenta oval in plot (a). The other white dot is at roughly (0.1, 0.4), corresponding to the center of the light blue oval in plot (a). This plot visually represents the predicted centers for the shapes in plot (a). Plot (c) displays a more complex arrangement with multiple colored oval shapes scattered across the black background. There are ten distinct ovals of various colors and orientations. Some of the colors visible are light brown, gray, pink, dark green, dark blue, light blue, purple, and green. They are distributed throughout the plot, including a cluster of smaller shapes in the upper region. Plot (d) shows several small white dots scattered across the black background. These dots appear to be the predicted centers for the shapes in plot (c). The dots are located at positions corresponding to the approximate centers of the various colored ovals in plot (c). Note: All the numerical data values are estimated.(a, c) Ground truth manifold structures; (b, d) PRS-predicted class centers. Source: Author’s own creation
The plots are arranged in a 2 by 2 grid. Plot (a) shows a black background with two oval-shaped colored regions. The first oval is a light blue color, oriented horizontally, and is centered at approximately (0, 0.5). The second oval is magenta, oriented diagonally, and is centered at roughly (negative 1.5, negative 0.25). The ovals are of a similar size. Plot (b) is a mostly black plot with two small white dots. One dot is located at approximately (negative 1.5, negative 0.25), corresponding to the center of the magenta oval in plot (a). The other white dot is at roughly (0.1, 0.4), corresponding to the center of the light blue oval in plot (a). This plot visually represents the predicted centers for the shapes in plot (a). Plot (c) displays a more complex arrangement with multiple colored oval shapes scattered across the black background. There are ten distinct ovals of various colors and orientations. Some of the colors visible are light brown, gray, pink, dark green, dark blue, light blue, purple, and green. They are distributed throughout the plot, including a cluster of smaller shapes in the upper region. Plot (d) shows several small white dots scattered across the black background. These dots appear to be the predicted centers for the shapes in plot (c). The dots are located at positions corresponding to the approximate centers of the various colored ovals in plot (c). Note: All the numerical data values are estimated.(a, c) Ground truth manifold structures; (b, d) PRS-predicted class centers. Source: Author’s own creation
Dataset details and classification accuracy of PRS, SVM, and LSTM models
| Dataset . | Domain . | #Classes . | Size . | Model . | ASE . | MR . | KS . | Precision . | Recall . | F1-score . | ROC-AUC . |
|---|---|---|---|---|---|---|---|---|---|---|---|
| SOD | Small objects | 2 | 5.2k | PRS | 0.0850 | 0.0690 | 0.6740 | 0.91 | 0.92 | 0.915 | 0.94 |
| SVM | 0.0725 | 0.0700 | 0.6750 | 0.88 | 0.89 | 0.885 | 0.91 | ||||
| LSTM | 0.0870 | 0.0750 | 0.6880 | 0.86 | 0.87 | 0.865 | 0.90 | ||||
| SMD | 2-manifolds | 2 | 6.1k | PRS | 0.0625 | 0.0710 | 0.6556 | 0.93 | 0.94 | 0.935 | 0.95 |
| SVM | 0.0549 | 0.0650 | 0.5963 | 0.89 | 0.88 | 0.885 | 0.90 | ||||
| LSTM | 0.0951 | 0.0620 | 0.6543 | 0.85 | 0.86 | 0.855 | 0.89 | ||||
| MMD | Multi-manifolds | 10 | 6.1k | PRS | 0.0519 | 0.0680 | 0.6600 | 0.90 | 0.91 | 0.905 | 0.92 |
| SVM | 0.0629 | 0.0750 | 0.5078 | 0.87 | 0.86 | 0.865 | 0.88 | ||||
| LSTM | 0.0950 | 0.0700 | 0.6956 | 0.84 | 0.83 | 0.835 | 0.87 |
| Dataset . | Domain . | #Classes . | Size . | Model . | ASE . | MR . | KS . | Precision . | Recall . | F1-score . | ROC-AUC . |
|---|---|---|---|---|---|---|---|---|---|---|---|
| SOD | Small objects | 2 | 5.2k | PRS | 0.0850 | 0.0690 | 0.6740 | 0.91 | 0.92 | 0.915 | 0.94 |
| SVM | 0.0725 | 0.0700 | 0.6750 | 0.88 | 0.89 | 0.885 | 0.91 | ||||
| LSTM | 0.0870 | 0.0750 | 0.6880 | 0.86 | 0.87 | 0.865 | 0.90 | ||||
| SMD | 2-manifolds | 2 | 6.1k | PRS | 0.0625 | 0.0710 | 0.6556 | 0.93 | 0.94 | 0.935 | 0.95 |
| SVM | 0.0549 | 0.0650 | 0.5963 | 0.89 | 0.88 | 0.885 | 0.90 | ||||
| LSTM | 0.0951 | 0.0620 | 0.6543 | 0.85 | 0.86 | 0.855 | 0.89 | ||||
| MMD | Multi-manifolds | 10 | 6.1k | PRS | 0.0519 | 0.0680 | 0.6600 | 0.90 | 0.91 | 0.905 | 0.92 |
| SVM | 0.0629 | 0.0750 | 0.5078 | 0.87 | 0.86 | 0.865 | 0.88 | ||||
| LSTM | 0.0950 | 0.0700 | 0.6956 | 0.84 | 0.83 | 0.835 | 0.87 |
As shown in Table 4, PRS outperforms both SVM and LSTM in most metrics across all datasets. For instance, in the MMD dataset, which presents the highest complexity with 10 classes, PRS achieves a lower ASE (0.0519), lower MR (0.0680), and higher KS (0.6600), indicating better alignment between predicted and actual distributions. Furthermore, PRS leads in classification metrics with a Precision of 0.90, a Recall of 0.91, an F1-score of 0.905, and an ROC-AUC of 0.92, surpassing both SVM and LSTM.
These trends hold across the SOD and SMD datasets as well, where PRS consistently yields better or comparable ASE and MR, and shows powerful results in Recall and ROC-AUC, which are crucial for high-sensitivity applications. KS statistics, a measure of distributional similarity, also confirms PRS’s advantage in maintaining output stability and accuracy across classes.
The ROC analysis presented in Figure 10 highlights the superior discriminative performance of the PRS model in the MMD object recognition task. Compared to SVM and LSTM, PRS consistently achieves higher sensitivity across a wide range of specificity thresholds. Notably, aligning the PRS validation curve (dashed line) with its training and test curves indicates minimal overfitting and robust generalization to unseen data. These results underscore PRS’s effectiveness in handling complex, multi-manifold classification challenges and its advantage in data-scarce environments.
The horizontal axis is labeled “1-Specificity” and has markings ranging from 0.0 to 1.0 in increments of 0.2 units. The vertical axis is labeled “Sensitivity” and has markings ranging from 0.0 to 1.0 in increments of 0.2 units. The graph shows 9 curves for “Data Roles.” The first curve for “L S T M Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.05, 0.73), (0.15, 0.85), (0.31, 0.94), (0.53, 1), and terminates at (1.0, 1.0). The second curve for “S V M Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.02, 0.46), (0.06, 0.56), (0.34, 0.85), and terminates at (1.0, 1.0). The third curve for “P D S Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.83), (0.55, 0.98), and terminates at (1.0, 1.0). The fourth curve for “L S T M Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.06, 0.71), (0.16, 0.84), (0.31, 0.92), (0.6, 0.98), and terminates at (1.0, 1.0). The fifth curve for “S V M Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.02, 0.52), (0.06, 0.65), (0.33, 0.87), and terminates at (1.0, 1.0). The sixth curve for “P D S Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.05, 0.66), (0.33, 0.89), and terminates at (1.0, 1.0). The seventh curve for “L S T M Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.83), (0.55, 0.97), and terminates at (1.0, 1.0). The eighth curve for “S V M Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.06, 0.64), (0.33, 0.88), and terminates at (1.0, 1.0). The ninth curve for “P D S Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.84), (0.29, 0.91), and terminates at (1.0, 1.0). Note: All numerical data values are approximated.ROC curves of PRS, SVM, and LSTM models on the MMD dataset. Source: Author’s own creation
The horizontal axis is labeled “1-Specificity” and has markings ranging from 0.0 to 1.0 in increments of 0.2 units. The vertical axis is labeled “Sensitivity” and has markings ranging from 0.0 to 1.0 in increments of 0.2 units. The graph shows 9 curves for “Data Roles.” The first curve for “L S T M Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.05, 0.73), (0.15, 0.85), (0.31, 0.94), (0.53, 1), and terminates at (1.0, 1.0). The second curve for “S V M Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.02, 0.46), (0.06, 0.56), (0.34, 0.85), and terminates at (1.0, 1.0). The third curve for “P D S Test” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.83), (0.55, 0.98), and terminates at (1.0, 1.0). The fourth curve for “L S T M Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.06, 0.71), (0.16, 0.84), (0.31, 0.92), (0.6, 0.98), and terminates at (1.0, 1.0). The fifth curve for “S V M Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.02, 0.52), (0.06, 0.65), (0.33, 0.87), and terminates at (1.0, 1.0). The sixth curve for “P D S Train” starts from (0.0, 0.0), rises upward passing through coordinates (0.05, 0.66), (0.33, 0.89), and terminates at (1.0, 1.0). The seventh curve for “L S T M Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.83), (0.55, 0.97), and terminates at (1.0, 1.0). The eighth curve for “S V M Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.06, 0.64), (0.33, 0.88), and terminates at (1.0, 1.0). The ninth curve for “P D S Validate” starts from (0.0, 0.0), rises upward passing through coordinates (0.16, 0.84), (0.29, 0.91), and terminates at (1.0, 1.0). Note: All numerical data values are approximated.ROC curves of PRS, SVM, and LSTM models on the MMD dataset. Source: Author’s own creation
To get a well-rounded comparison, we tested the PRS model against traditional approaches LSTM and SVM, as well as more advanced techniques such as Prototypical Networks and Transfer Learning, which are known to perform well with limited data. These comparisons helped us better understand where PRS regarding accuracy and efficiency. While Table 5 shows that PRS takes longer to train than the other four models, this drawback is offset by not requiring repeated cross-validation to tune complexity parameters. Additionally, as it generates models with fewer variables, the computational time for evaluating test points, often the primary concern in practical scenarios, is significantly reduced. Based on the average training times for the BBBC022 dataset, the PRS model demonstrates a competitive balance between performance and computational efficiency. While its training times, 62.3 seconds for Dataset 1 (Inclusion) and 75.1 seconds for Dataset 2 (Manifold), are higher than those of simpler models like SVM (15.8s/20.3s) and LSTM (38.2s/46.7s), PRS outperforms them in terms of accuracy, making the extra training cost justifiable. Compared to Prototypical Networks (50.0s/58.0s) and Transfer Learning (70.0s/85.0s), PRS sits comfortably in the middle, offering accuracy nearly on par with Transfer Learning but at a slightly lower training cost. This makes PRS a reasonable alternative to established models, particularly in scenarios where high pattern recognition accuracy is essential. However, complete reliance on heavy pretraining (as in Transfer Learning) is less desirable.
Average training time (in seconds)
| Model . | Dataset 1 (inclusion) . | Dataset 2 (manifold) . |
|---|---|---|
| LSTM | 38.2 | 46.7 |
| Prototypical networks | 50.0 | 58.0 |
| Transfer learning | 70.0 | 85.0 |
| SVM | 15.8 | 20.3 |
| PRS | 62.3 | 75.1 |
| Model . | Dataset 1 (inclusion) . | Dataset 2 (manifold) . |
|---|---|---|
| LSTM | 38.2 | 46.7 |
| Prototypical networks | 50.0 | 58.0 |
| Transfer learning | 70.0 | 85.0 |
| SVM | 15.8 | 20.3 |
| PRS | 62.3 | 75.1 |
The PRS model demonstrates strong performance on controlled datasets, yet its generalization to noisy or highly imbalanced real-world data has not been extensively tested. Training time is higher than conventional models, though offset by lower testing cost. Future work will explore model robustness to noise and domain transfer.
Table 6 compares the performance of five different models, LSTM, Prototypical Networks, Transfer Learning, SVM, and PRS on the BBBC022 dataset (MMD), a benchmark for pattern recognition in bioimage analysis. Transfer Learning led the pack with a top accuracy of 92%, but PRS followed closely at 91%, showing it can match the best while offering unique strengths in adaptability and computational efficiency. Prototypical Networks also delivered solid results at 88%. LSTM and SVM lagged with 82% and 80%, respectively, highlighting their limitations in this setting.
Overall, PRS proves to be a strong and reliable contender for pattern recognition tasks in biomedical imaging.
6. Conclusion and future work
The experiment showcases our framework’s successful implementation and effectiveness while demonstrating the practicality of IP-based PRS. The comparative findings reveal that it achieves 90% accuracy across multiple applications. Furthermore, we have compared our IP-based PRS model to baseline models, LSTM and SVM, and advanced models for small data scenarios such as Prototypical Networks and Transfer learning. Our results surpassed those of SVM, LSTM, and Prototypical Networks by 11%, 9%, and 3%, respectively. On the other hand, Transfer Learning did better than the PRS by 3%, but the PRS model offers a favorable trade-off between accuracy and training efficiency. The proposed inverse map framework for DNN makes it possible for the model to achieve strong generalization from limited data. That makes it well-suited for small data prediction problems where both predictive accuracy and computational efficiency are critical.
Nevertheless, the model has limitations, highlighting the need for further examination.
Future work will explore domain generalization, adaptability to noisy data, better numerics, and extending PRS to more complex and heterogeneous datasets. Integrating physics-based operators into deep learning offers a promising path for improving performance in data-scarce applications.
Additionally, the PRS model was developed using moderate computing power (a mid-range GPU setup), and its sparse architecture allows it to maintain practical training times without requiring specialized hardware. This makes PRS a cost-effective and accessible solution for small-data problems in academic or clinical settings.
Overall, we aim for this approach to serve as both a challenge and a benchmark. By leveraging the close relationship between inverse problems and pattern recognition, we can create an innovative learning algorithm suitable for handling small data sets.
The supplementary material for this article can be found online.


