SCIENTIFIC COMPUTING AND IMAGING INSTITUTE
at the University of Utah

An internationally recognized leader in visualization, scientific computing, and image analysis

SCI Publications

2025


B. Charoenwong, R.M. Kirby, J. Reiter. “Tradeoffs in automated financial regulation of decentralized finance due to limits on mutable turing machines,” In Scientific Reports, Vol. 15, No. 3016, 2025.
DOI: https://doi.org/10.1038/s41598-024-84612-9

ABSTRACT

We examine which decentralized finance architectures enable meaningful regulation by combining financial and computational theory. We show via deduction that a decentralized and permissionless Turing-complete system cannot provably comply with regulations concerning anti-money laundering, know-your-client obligations, some securities restrictions and forms of exchange control. Any system that claims to follow regulations must choose either a form of permission or a less-than-Turing-complete update facility. Compliant decentralized systems can be constructed only by compromising on the richness of permissible changes. Regulatory authorities must accept new tradeoffs that limit their enforcement powers if they want to approve permissionless platforms formally. Our analysis demonstrates that the fundamental constraints of computation theory have direct implications for financial regulation. By mapping regulatory requirements onto computational models, we characterize which types of automated compliance are achievable and which are provably impossible. This framework allows us to move beyond traditional debates about regulatory effectiveness to establish concrete boundaries for automated enforcement.



N. Gorski, X. Liang, H. Guo, L. Yan, B. Wang. “A General Framework for Augmenting Lossy Compressors with Topological Guarantees,” Subtitled “ arXiv:2502.14022,” 2025.

ABSTRACT

Topological descriptors such as contour trees are widely utilized in scientific data analysis and visualization, with applications from materials science to climate simulations. It is desirable to preserve topological descriptors when data compression is part of the scientific workflow for these applications. However, classic error-bounded lossy compressors for volumetric data do not guarantee the preservation of topological descriptors, despite imposing strict pointwise error bounds. In this work, we introduce a general framework for augmenting any lossy compressor to preserve the topology of the data during compression. Specifically, our framework quantifies the adjustments (to the decompressed data) needed to preserve the contour tree and then employs a custom variable-precision encoding scheme to store these adjustments. We demonstrate the utility of our framework in augmenting classic compressors (such as SZ3, TTHRESH, and ZFP) and deep learning-based compressors (such as Neurcomp) with topological guarantees.



J.K. Holmen, M. Garcia, A. Sanderson, A. Bagusetty, M. Berzins. “Lessons Learned and Scalability Achieved when Porting Uintah to DOE Exascale Systems,” In Proceedings of the AMTE workshop (accepted), 2025.

ABSTRACT

A key challenge faced when preparing codes for Department of Energy (DOE) exascale systems was designing scalable applications for systems featuring hardware and software not yet available at leadership class scale. With such systems now available, it is important to evaluate scalability of the resulting software solutions on these target systems. One such code designed with the exascale DOE Aurora and DOE Frontier systems in mind is the Uintah Computational Framework, an open-source asynchronous many-task (AMT) runtime system. To prepare for exascale, Uintah adopted a portable MPI+X hybrid parallelism approach using the Kokkos performance portability library (i.e., MPI+Kokkos). This paper complements recent work with additional details and an evaluation of the resulting approach on Aurora and Frontier. Results are shown for a challenging benchmark demonstrating interoperability of 3 portable codes essential to Uintah-related combustion research. These results demonstrate single-source portability across Aurora and Frontier with scaling characteristics shown to 3,072 Aurora nodes and 9,216 Frontier nodes. In addition to showing results run to new scales on new systems, this paper also discusses lessons learned through efforts preparing Uintah for exascale systems.



X. Huang, W. Usher, V. Pascucci. “Approximate Puzzlepiece Compositing,” Subtitled “arXiv:2501.12581,” 2025.

ABSTRACT

The increasing demand for larger and higher fidelity simulations has made Adaptive Mesh Refinement (AMR) and unstructured mesh techniques essential to focus compute effort and memory cost on just the areas of interest in the simulation domain. The distribution of these meshes over the compute nodes is often determined by balancing compute, memory, and network costs, leading to distributions with jagged nonconvex boundaries that fit together much like puzzle pieces. It is expensive, and sometimes impossible, to re-partition the data posing a challenge for in situ and post hoc visualization as the data cannot be rendered using standard sort-last compositing techniques that require a convex and disjoint data partitioning. We present a new distributed volume rendering and compositing algorithm, Approximate Puzzlepiece Compositing, that enables fast and high-accuracy in-place rendering of AMR and unstructured meshes. Our approach builds on Moment-Based Ordered-Independent Transparency to achieve a scalable, order-independent compositing algorithm that requires little communication and does not impose requirements on the data partitioning. We evaluate the image quality and scalability of our approach on synthetic data and two large-scale unstructured meshes on HPC systems by comparing to state-of-the-art sort-last compositing techniques, highlighting our approach’s minimal overhead at higher core counts. We demonstrate that Approximate Puzzlepiece Compositing provides a scalable, high-performance, and high-quality distributed rendering approach applicable to the complex data distributions encountered in large-scale CFD simulations.



B. Hunt, E. Kwan, J. Bergquist, J. Brundage, B. Orkild, J. Dong, E. Paccione, K. Yazaki, R.S. MacLeod, D. Dosdall, T. Tasdizen, R. Ranjan. “Contrastive Pretraining Improves Deep Learning Classification of Endocardial Electrograms in a Preclinical Model,” In Heart Rhythm O2, Elsevier, 2025.
ISSN: 2666-5018
DOI: https://doi.org/10.1016/j.hroo.2025.01.008

ABSTRACT

Background

Rotors and focal ectopies, or “drivers,” are hypothesized mechanisms of persistent atrial fibrillation (AF). Machine learning algorithms have been employed to identify these drivers, but the limited size of current driver datasets constrains their performance.

Objective

We proposed that pretraining using unsupervised learning on a substantial dataset of unlabeled electrograms could enhance classifier accuracy when applied to a smaller driver dataset.

Methods

We utilized a SimCLR-based framework to pretrain a residual neural network on 113,000 unlabeled 64-electrode measurements from a canine model of AF. The network was then fine-tuned to identify drivers from intra-cardiac electrograms. Various augmentations, including cropping, Gaussian blurring, and rotation, were applied during pretraining to improve the robustness of the learned representations.

Results

Pretraining significantly improved driver detection accuracy compared to a non-pretrained network (80.8% vs. 62.5%). The pretrained network also demonstrated greater resilience to reductions in training dataset size, maintaining higher accuracy even with a 30% reduction in data. Grad-CAM analysis revealed that the network’s attention aligned well with manually annotated driver regions, suggesting that the network learned meaningful features for driver detection.

Conclusion

This study demonstrates that contrastive pretraining can enhance the accuracy of driver detection algorithms in AF. The findings support the broader application of transfer learning to other electrogram-based tasks, potentially improving outcomes in clinical electrophysiology.



K. Iyer, M.S.T. Karanam, S. Elhabian. “Mesh2SSM++: A Probabilistic Framework for Unsupervised Learning of Statistical Shape Model of Anatomies from Surface Meshes,” Subtitled “arXiv:2502.07145,” 2025.

ABSTRACT

Anatomy evaluation is crucial for understanding the physiological state, diagnosing abnormalities, and guiding medical interventions. Statistical shape modeling (SSM) is vital in this process, particularly in medical image analysis and computational anatomy. By enabling the extraction of quantitative morphological shape descriptors from medical imaging data such as MRI and CT scans, SSM provides comprehensive descriptions of anatomical variations within a population. However, the effectiveness of SSM in anatomy evaluation hinges on the quality and robustness of the shape models, which face challenges due to substantial nonlinear variability in human anatomy. While deep learning techniques show promise in addressing these challenges by learning complex nonlinear representations of shapes, existing models still have limitations and often require pre-established shape models for training. To overcome these issues, we propose Mesh2SSM++, a novel approach that learns to estimate correspondences from meshes in an unsupervised manner. This method leverages unsupervised, permutation-invariant representation learning to estimate how to deform a template point cloud into subject-specific meshes, forming a correspondence-based shape model. Additionally, our probabilistic formulation allows learning a population-specific template, reducing potential biases associated with template selection. A key feature of Mesh2SSM++ is its ability to quantify aleatoric uncertainty, which captures inherent data variability and is essential for ensuring reliable model predictions and robust decision-making in clinical tasks, especially under challenging imaging conditions. Through extensive validation across diverse anatomies, evaluation metrics, and downstream tasks, we demonstrate that Mesh2SSM++ outperforms existing methods. Its ability to operate directly on meshes, combined with computational efficiency and interpretability through its probabilistic framework, makes it an attractive alternative to traditional and deep learning-based SSM approaches. Github: https://github.com/iyerkrithika21/Mesh2SSMJournal



M.S.T. Karanam, K. Iyer, S. Joshi, S. Elhabian. “Log-Euclidean Regularization for Population-Aware Image Registration,” Subtitled “arXiv:2502.02029,” 2025.

ABSTRACT

Spatial transformations that capture population-level morphological statistics are critical for medical image analysis. Commonly used smoothness regularizers for image registration fail to integrate population statistics, leading to anatomically inconsistent transformations. Inverse consistency regularizers promote geometric consistency but lack population morphometrics integration. Regularizers that constrain deformation to low-dimensional manifold methods address this. However, they prioritize reconstruction over interpretability and neglect diffeomorphic properties, such as group composition and inverse consistency. We introduce MORPH-LER, a Log-Euclidean regularization framework for population-aware unsupervised image registration. MORPH-LER learns population morphometrics from spatial transformations to guide and regularize registration networks, ensuring anatomically plausible deformations. It features a bottleneck autoencoder that computes the principal logarithm of deformation fields via iterative square-root predictions. It creates a linearized latent space that respects diffeomorphic properties and enforces inverse consistency. By integrating a registration network with a diffeomorphic autoencoder, MORPH-LER produces smooth, meaningful deformation fields. The framework offers two main contributions: (1) a data-driven regularization strategy that incorporates population-level anatomical statistics to enhance transformation validity and (2) a linearized latent space that enables compact and interpretable deformation fields for efficient population morphometrics analysis. We validate MORPH-LER across two families of deep learning-based registration networks, demonstrating its ability to produce anatomically accurate, computationally efficient, and statistically meaningful transformations on the OASIS-1 brain imaging dataset.



A. C. Peterson, M. R. Requist, J. C. Benna, J. R. Nelson, S. Elhabian, C. de Cesar Netto, T. C. Beals, A. L. Lenz. “Talar Morphology of Charcot-Marie-Tooth Patients With Cavovarus Feet,” In Foot & Ankle International, Sage Publications, 2025.
DOI: 10.1177/10711007241309915
PubMed ID: 39937093

ABSTRACT

Background:

Charcot-Marie-Tooth disease (CMT), a common inherited neurologic disorder, significantly impacts the morphology of foot bones, particularly the talus. The disease has been classified into types based on specific mutations, with the most common being CMT type 1 (CMT1; demyelinating) and CMT type 2 (CMT2; axonal). However, the specific osseous morphologic variations in CMT patients and their major genetic subgroups remain insufficiently understood, posing challenges in clinical management and surgical intervention.

Methods:

This study analyzed talar morphology in individuals with CMT compared with a healthy control group, employing a single-bone statistical shape model and talar neck offset angle measurements. Participants included 18 CMT individuals (yielding 29 tali) and 43 healthy controls. For individuals with CMT, the average age at diagnosis was 36.5 ± 19.8 years, with a mean interval of 8.6 years between diagnosis and imaging. Talar morphology was evaluated using weightbearing computed tomography and subsequent morphologic and angular analysis.

Results:

Differences were observed in talar morphology between CMT and healthy individuals. Notably, CMT1 and CMT2 tali exhibited a flatter talar dome and more medial talar head and neck compared with controls. Additionally, the CMT1 and CMT2 subgroups both had a more medially oriented talar neck based on the talar neck offset angle compared with the controls.

Conclusion:

The findings illustrate significant morphologic variations in the talus of CMT patients, indicating the need for type-specific clinical approaches in treating CMT-related foot deformities. Understanding these talar variations is crucial for tailoring surgical techniques and orthotic designs, and developing effective rehabilitation protocols for individuals with CMT, potentially improving patient care and outcomes.



S. Saha, S. Joshi, R. Whitaker. “ARD-VAE: A Statistical Formulation to Find the Relevant Latent Dimensions of Variational Autoencoders,” Subtitled “arXiv:2501.10901,” 2025.

ABSTRACT

The variational autoencoder (VAE) is a popular, deep, latent-variable model (DLVM) due to its simple yet effective formulation for modeling the data distribution. Moreover, optimizing the VAE objective function is more manageable than other DLVMs. The bottleneck dimension of the VAE is a crucial design choice, and it has strong ramifications for the model’s performance, such as finding the hidden explanatory factors of a dataset using the representations learned by the VAE. However, the size of the latent dimension of the VAE is often treated as a hyperparameter estimated empirically through trial and error. To this end, we propose a statistical formulation to discover the relevant latent factors required for modeling a dataset. In this work, we use a hierarchical prior in the latent space that estimates the variance of the latent axes using the encoded data, which identifies the relevant latent dimensions. For this, we replace the fixed prior in the VAE objective function with a hierarchical prior, keeping the remainder of the formulation unchanged. We call the proposed method the automatic relevancy detection in the variational autoencoder (ARD-VAE). We demonstrate the efficacy of the ARD-VAE on multiple benchmark datasets in finding the relevant latent dimensions and their effect on different evaluation metrics, such as FID score and disentanglement analysis.



S. Saha, S. Joshi, R. Whitaker. “Disentanglement Analysis in Deep Latent Variable Models Matching Aggregate Posterior Distributions,” Subtitled “arXiv:2501.15705,” 2025.

ABSTRACT

Deep latent variable models (DLVMs) are designed to learn meaningful representations in an unsupervised manner, such that the hidden explanatory factors are interpretable by independent latent variables (aka disentanglement). The variational autoencoder (VAE) is a popular DLVM widely studied in disentanglement analysis due to the modeling of the posterior distribution using a factorized Gaussian distribution that encourages the alignment of the latent factors with the latent axes. Several metrics have been proposed recently, assuming that the latent variables explaining the variation in data are aligned with the latent axes (cardinal directions). However, there are other DLVMs, such as the AAE and WAE-MMD (matching the aggregate posterior to the prior), where the latent variables might not be aligned with the latent axes. In this work, we propose a statistical method to evaluate disentanglement for any DLVMs in general. The proposed technique discovers the latent vectors representing the generative factors of a dataset that can be different from the cardinal latent axes. We empirically demonstrate the advantage of the method on two datasets.



S. Saha, R. Whitaker. “AdaSemSeg: An Adaptive Few-shot Semantic Segmentation of Seismic Facies,” Subtitled “arXiv:2501.16760,” 2025.

ABSTRACT

Automated interpretation of seismic images using deep learning methods is challenging because of the limited availability of training data. Few-shot learning is a suitable learning paradigm in such scenarios due to its ability to adapt to a new task with limited supervision (small training budget). Existing few-shot semantic segmentation (FSSS) methods fix the number of target classes. Therefore, they do not support joint training on multiple datasets varying in the number of classes. In the context of the interpretation of seismic facies, fixing the number of target classes inhibits the generalization capability of a model trained on one facies dataset to another, which is likely to have a different number of facies. To address this shortcoming, we propose a few-shot semantic segmentation method for interpreting seismic facies that can adapt to the varying number of facies across the dataset, dubbed the AdaSemSeg. In general, the backbone network of FSSS methods is initialized with the statistics learned from the ImageNet dataset for better performance. The lack of such a huge annotated dataset for seismic images motivates using a self-supervised algorithm on seismic datasets to initialize the backbone network. We have trained the AdaSemSeg on three public seismic facies datasets with different numbers of facies and evaluated the proposed method on multiple metrics. The performance of the AdaSemSeg on unseen datasets (not used in training) is better than the prototype-based few-shot method and baselines.



C. Scully-Allison, K. Williams, S. Brink, O. Pearce, K. Isaacs. “A Tale of Two Models: Understanding Data Workers' Internal and External Representations of Complex Data,” Subtitled “arXiv:2501.09862v2,” 2025.

ABSTRACT

Data workers may have a different mental model of their data than the one reified in code. Understanding the organization of their data is necessary for analyzing data, be it through scripting, visualization, or abstract thought. More complicated organizations, such as tables with attached hierarchies, may tax people’s ability to think about and interact with data. To better understand and ultimately design for these situations, we conduct a study across a team of ten people working with the same reified data model. Through interviews and sketching, we probed their conception of the data model and developed themes through reflexive data analysis. Participants had diverse data models that differed from the reified data model, even among team members who had designed the model, resulting in parallel hazards limiting their ability to reason about the data. From these observations, we suggest potential design interventions for data analysis processes and tools.



C. You, H. Dai, Y. Min, J.S. Sekhon, S. Joshi, J.S. Duncan. “The Silent Majority: Demystifying Memorization Effect in the Presence of Spurious Correlations,” Subtitled “arXiv:2501.00961v2,” 2025.

ABSTRACT

Machine learning models often rely on simple spurious features – patterns in training data that correlate with targets but are not causally related to them, like image backgrounds in foreground classification. This reliance typically leads to imbalanced test performance across minority and majority groups. In this work, we take a closer look at the fundamental cause of such imbalanced performance through the lens of memorization, which refers to the ability to predict accurately on atypical examples (minority groups) in the training set but failing in achieving the same accuracy in the testing set. This paper systematically shows the ubiquitous existence of spurious features in a small set of neurons within the network, providing the first-ever evidence that memorization may contribute to imbalanced group performance. Through three experimental sources of converging empirical evidence, we find the property of a small subset of neurons or channels in memorizing minority group information. Inspired by these findings, we articulate the hypothesis: the imbalanced group performance is a byproduct of “noisy” spurious memorization confined to a small set of neurons. To further substantiate this hypothesis, we show that eliminating these unnecessary spurious memorization patterns via a novel framework during training can significantly affect the model performance on minority groups. Our experimental results across various architectures and benchmarks offer new insights on how neural networks encode core and spurious knowledge, laying the groundwork for future research in demystifying robustness to spurious correlation.


2024


J. Adams, K. Iyer, S. Elhabian. “Weakly Supervised Bayesian Shape Modeling from Unsegmented Medical Images,” Subtitled “arXiv:2405.09697v1,” 2024.

ABSTRACT

Anatomical shape analysis plays a pivotal role in clinical research and hypothesis testing, where the relationship between form and function is paramount. Correspondence-based statistical shape modeling (SSM) facilitates population-level morphometrics but requires a cumbersome, potentially bias-inducing construction pipeline. Recent advancements in deep learning have streamlined this process in inference by providing SSM prediction directly from unsegmented medical images. However, the proposed approaches are fully supervised and require utilizing a traditional SSM construction pipeline to create training data, thus inheriting the associated burdens and limitations. To address these challenges, we introduce a weakly supervised deep learning approach to predict SSM from images using point cloud supervision. Specifically, we propose reducing the supervision associated with the state-of-the-art fully Bayesian variational information bottleneck DeepSSM (BVIB-DeepSSM) model. BVIB-DeepSSM is an effective, principled framework for predicting probabilistic anatomical shapes from images with quantification of both aleatoric and epistemic uncertainties. Whereas the original BVIB-DeepSSM method requires strong supervision in the form of ground truth correspondence points, the proposed approach utilizes weak supervision via point cloud surface representations, which are more readily obtainable. Furthermore, the proposed approach learns correspondence in a completely data-driven manner without prior assumptions about the expected variability in shape cohort. Our experiments demonstrate that this approach yields similar accuracy and uncertainty estimation to the fully supervised scenario while substantially enhancing the feasibility of model training for SSM construction.



J. Adams, S. Elhabian. “Point2SSM++: Self-Supervised Learning of Anatomical Shape Models from Point Clouds,” Subtitled “arXiv:2405.09707v1,” 2024.

ABSTRACT

Correspondence-based statistical shape modeling (SSM) stands as a powerful technology for morphometric analysis in clinical research. SSM facilitates population-level characterization and quantification of anatomical shapes such as bones and organs, aiding in pathology and disease diagnostics and treatment planning. Despite its potential, SSM remains under-utilized in medical research due to the significant overhead associated with automatic construction methods, which demand complete, aligned shape surface representations. Additionally, optimization-based techniques rely on bias-inducing assumptions or templates and have prolonged inference times as the entire cohort is simultaneously optimized. To overcome these challenges, we introduce Point2SSM++, a principled, self-supervised deep learning approach that directly learns correspondence points from point cloud representations of anatomical shapes. Point2SSM++ is robust to misaligned and inconsistent input, providing SSM that accurately samples individual shape surfaces while effectively capturing population-level statistics. Additionally, we present principled extensions of Point2SSM++ to adapt it for dynamic spatiotemporal and multi-anatomy use cases, demonstrating the broad versatility of the Point2SSM++ framework. Furthermore, we present extensions of Point2SSM++ tailored for dynamic spatiotemporal and multi-anatomy scenarios, showcasing the broad versatility of the framework. Through extensive validation across diverse anatomies, evaluation metrics, and clinically relevant downstream tasks, we demonstrate Point2SSM++’s superiority over existing state-of-the-art deep learning models and traditional approaches. Point2SSM++ substantially enhances the feasibility of SSM generation and significantly broadens its array of potential clinical applications.



S.I. Adams-Tew, H. Odéen, D.L. Parker, C.C. Cheng, B. Madore, A. Payne, S. Joshi. “Physics Informed Neural Networks for Estimation of Tissue Properties from Multi-echo Configuration State MRI,” In Medical Image Computing and Computer Assisted Intervention -- MICCAI 2024, Springer Nature Switzerland, pp. 502--511. 2024.

ABSTRACT

This work investigates the use of configuration state imaging together with deep neural networks to develop quantitative MRI techniques for deployment in an interventional setting. A physics modeling technique for inhomogeneous fields and heterogeneous tissues is presented and used to evaluate the theoretical capability of neural networks to estimate parameter maps from configuration state signal data. All tested normalization strategies achieved similar performance in estimating T2 and T2*. Varying network architecture and data normalization had substantial impacts on estimated flip angle and T1, highlighting their importance in developing neural networks to solve these inverse problems. The developed signal modeling technique provides an environment that will enable the development and evaluation of physics-informed machine learning techniques for MR parameter mapping and facilitate the development of quantitative MRI techniques to inform clinical decisions during MR-guided treatments.



T. M. Athawale, B. Triana, T. Kotha, D. Pugmire, P. Rosen. “A Comparative Study of the Perceptual Sensitivity of Topological Visualizations to Feature Variations,” In IEEE Transactions on Visualization and Computer Graphics, Vol. 30, No. 1, pp. 1074-1084. Jan, 2024.
DOI: 10.1109/TVCG.2023.3326592

ABSTRACT

Color maps are a commonly used visualization technique in which data are mapped to optical properties, e.g., color or opacity. Color maps, however, do not explicitly convey structures (e.g., positions and scale of features) within data. Topology-based visualizations reveal and explicitly communicate structures underlying data. Although we have a good understanding of what types of features are captured by topological visualizations, our understanding of people’s perception of those features is not. This paper evaluates the sensitivity of topology-based isocontour, Reeb graph, and persistence diagram visualizations compared to a reference color map visualization for synthetically generated scalar fields on 2-manifold triangular meshes embedded in 3D. In particular, we built and ran a human-subject study that evaluated the perception of data features characterized by Gaussian signals and measured how effectively each visualization technique portrays variations of data features arising from the position and amplitude variation of a mixture of Gaussians. For positional feature variations, the results showed that only the Reeb graph visualization had high sensitivity. For amplitude feature variations, persistence diagrams and color maps demonstrated the highest sensitivity, whereas isocontours showed only weak sensitivity. These results take an important step toward understanding which topology-based tools are best for various data and task scenarios and their effectiveness in conveying topological variations as compared to conventional color mapping.



T.M. Athawale, Z. Wang, D. Pugmire, K. Moreland, Q. Gong, S. Klasky, C.R. Johnson, P. Rosen. “Uncertainty Visualization of Critical Points of 2D Scalar Fields for Parametric and Nonparametric Probabilistic Models,” In IEEE Transactions on Visualization and Computer Graphics, IEEE, pp. 1--11. 2024.

ABSTRACT

This paper presents a novel end-to-end framework for closed-form computation and visualization of critical point uncertainty in 2D uncertain scalar fields. Critical points are fundamental topological descriptors used in the visualization and analysis of scalar fields. The uncertainty inherent in data (e.g., observational and experimental data, approximations in simulations, and compression), however, creates uncertainty regarding critical point positions. Uncertainty in critical point positions, therefore, cannot be ignored, given their impact on downstream data analysis tasks. In this work, we study uncertainty in critical points as a function of uncertainty in data modeled with probability distributions. Although Monte Carlo (MC) sampling techniques have been used in prior studies to quantify critical point uncertainty, they are often expensive and are infrequently used in production-quality visualization software. We, therefore, propose a new end-to-end framework to address these challenges that comprises a threefold contribution. First, we derive the critical point uncertainty in closed form, which is more accurate and efficient than the conventional MC sampling methods. Specifically, we provide the closed-form and semianalytical (a mix of closed-form and MC methods) solutions for parametric (e.g., uniform, Epanechnikov) and nonparametric models (e.g., histograms) with finite support. Second, we accelerate critical point probability computations using a parallel implementation with the VTK-m library, which is platform portable. Finally, we demonstrate the integration of our implementation with the ParaView software system to demonstrate near-real-time results for real datasets.



B. Aubert, N. Khan, F. Toupin, M. Pacheco, A. Morris. “Deformable Vertebra 3D/2D Registration from Biplanar X-Rays Using Particle-Based Shape Modelling,” In Shape in Medical Imaging, Springer Nature Switzerland, pp. 33--47. 2024.
ISSN: 978-3-031-75291-9

ABSTRACT

Patient-specific 3D vertebra models are essential for accurately assessing the spinal deformities quantitatively in 3D and for surgical planning, including determining the optimal implant size and 3D positioning. Calibrated biplanar X-rays serve as an alternative to CT scans to generate the 3D models in a weight-bearing standing position. This paper presents an intensity-based 3D/2D registration method for vertebra statistical shape model (VSSM), incorporating two key elements: the particle-based shape modeling and an image domain transfer for efficient image matching. In the 3D/3D setting, the VSSMs reach a surface reconstruction error of less than 0.5 mm. For 3D reconstruction from biplanar X-rays, the root mean square point-to-surface are 1.05 mm for L1 to L4 vertebrae and 1.6 mm for the L5 vertebra. The particle-based VSSMs offer a significant balance between the model compactness and the reconstruction error, which is advantageous for deformable 3D/2D registration.



A.Z.B. Aziz, M.S.T. Karanam, T. Kataria, S.Y. Elhabian. “EfficientMorph: Parameter-Efficient Transformer-Based Architecture for 3D Image Registration,” Subtitled “arXiv preprint arXiv:2403.11026,” 2024.

ABSTRACT

Transformers have emerged as the state-of-the-art architecture in medical image registration, outperforming convolutional neural networks (CNNs) by addressing their limited receptive fields and overcoming gradient instability in deeper models. Despite their success, transformer-based models require substantial resources for training, including data, memory, and computational power, which may restrict their applicability for end users with limited resources. In particular, existing transformer-based 3D image registration architectures face three critical gaps that challenge their efficiency and effectiveness. Firstly, while mitigating the quadratic complexity of full attention by focusing on local regions, window-based attention mechanisms often fail to adequately integrate local and global information. Secondly, feature similarities across attention heads that were recently found in multi-head attention architectures indicate a significant computational redundancy, suggesting that the capacity of the network could be better utilized to enhance performance. Lastly, the granularity of tokenization, a key factor in registration accuracy, presents a trade-off; smaller tokens improve detail capture at the cost of higher computational complexity, increased memory demands, and a risk of overfitting. Here, we propose EfficientMorph, a transformer-based architecture for unsupervised 3D image registration. It optimizes the balance between local and global attention through a plane-based attention mechanism, reduces computational redundancy via cascaded group attention, and captures fine details without compromising computational efficiency, thanks to a Hi-Res tokenization strategy complemented by merging operations. We compare the effectiveness of EfficientMorph on two public datasets, OASIS and IXI, against other state-of-the-art models. Notably, EfficientMorph sets a new benchmark for performance on the OASIS dataset with ∼16-27× fewer parameters.