• Complex
  • Title
  • Author
  • Keyword
  • Abstract
  • Scholars
Search
High Impact Results & Cited Count Trend for Year Keyword Cloud and Partner Relationship

Query:

学者姓名:孟德宇

Refining:

Source

Submit Unfold

Co-Author

Submit Unfold

Language

Submit

Clean All

Export Sort by:
Default
  • Default
  • Title
  • Year
  • WOS Cited Count
  • Impact factor
  • Ascending
  • Descending
< Page ,Total 20 >
RCDNet: An Interpretable Rain Convolutional Dictionary Network for Single Image Deraining SCIE Scopus
期刊论文 | 2023 | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
SCOPUS Cited Count: 7
Abstract&Keyword Cite

Abstract :

As common weather, rain streaks adversely degrade the image quality and tend to negatively affect the performance of outdoor computer vision systems. Hence, removing rains from an image has become an important issue in the field. To handle such an ill-posed single image deraining task, in this article, we specifically build a novel deep architecture, called rain convolutional dictionary network (RCDNet), which embeds the intrinsic priors of rain streaks and has clear interpretability. In specific, we first establish a rain convolutional dictionary (RCD) model for representing rain streaks and utilize the proximal gradient descent technique to design an iterative algorithm only containing simple operators for solving the model. By unfolding it, we then build the RCDNet in which every network module has clear physical meanings and corresponds to each operation involved in the algorithm. This good interpretability greatly facilitates an easy visualization and analysis of what happens inside the network and why it works well in the inference process. Moreover, taking into account the domain gap issue in real scenarios, we further design a novel dynamic RCDNet, where the rain kernels can be dynamically inferred corresponding to input rainy images and then help shrink the space for rain layer estimation with few rain maps, so as to ensure a fine generalization performance in the inconsistent scenarios of rain types between training and testing data. By end-to-end training such an interpretable network, all involved rain kernels and proximal operators can be automatically extracted, faithfully characterizing the features of both rain and clean background layers and, thus, naturally leading to better deraining performance. Comprehensive experiments implemented on a series of representative synthetic and real datasets substantiate the superiority of our method, especially on its well generality to diverse testing scenarios and good interpretability for all its modules, compared with state-of-the-art single image derainers both visually and quantitatively.

Keyword :

Dictionary learning generalization performance interpretable deep learning (DL) single image rain removal

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Wang, Hong , Xie, Qi , Zhao, Qian et al. RCDNet: An Interpretable Rain Convolutional Dictionary Network for Single Image Deraining [J]. | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS , 2023 .
MLA Wang, Hong et al. "RCDNet: An Interpretable Rain Convolutional Dictionary Network for Single Image Deraining" . | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2023) .
APA Wang, Hong , Xie, Qi , Zhao, Qian , Li, Yuexiang , Liang, Yong , Zheng, Yefeng et al. RCDNet: An Interpretable Rain Convolutional Dictionary Network for Single Image Deraining . | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS , 2023 .
Export to NoteExpress RIS BibTex
InDuDoNet plus : A deep unfolding dual domain network for metal artifact reduction in CT images EI SCIE Scopus
期刊论文 | 2023 , 85 | MEDICAL IMAGE ANALYSIS
SCOPUS Cited Count: 26
Abstract&Keyword Cite

Abstract :

During the computed tomography (CT) imaging process, metallic implants within patients often cause harmful artifacts, which adversely degrade the visual quality of reconstructed CT images and negatively affect the subsequent clinical diagnosis. For the metal artifact reduction (MAR) task, current deep learning based methods have achieved promising performance. However, most of them share two main common limitations: (1) the CT physical imaging geometry constraint is not comprehensively incorporated into deep network structures; (2) the entire framework has weak interpretability for the specific MAR task; hence, the role of each network module is difficult to be evaluated. To alleviate these issues, in the paper, we construct a novel deep unfolding dual domain network, termed InDuDoNet+, into which CT imaging process is finely embedded. Concretely, we derive a joint spatial and Radon domain reconstruction model and propose an optimization algorithm with only simple operators for solving it. By unfolding the iterative steps involved in the proposed algorithm into the corresponding network modules, we easily build the InDuDoNet+ with clear interpretability. Furthermore, we analyze the CT values among different tissues, and merge the prior observations into a prior network for our InDuDoNet+, which significantly improve its generalization performance. Comprehensive experiments on synthesized data and clinical data substantiate the superiority of the proposed methods as well as the superior generalization performance beyond the current state-of-the-art (SOTA) MAR methods. Code is available at https://github.com/hongwang01/InDuDoNet_plus.

Keyword :

CT imaging geometry Generalization ability Metal artifact reduction Physical interpretability

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Wang, Hong , Li, Yuexiang , Zhang, Haimiao et al. InDuDoNet plus : A deep unfolding dual domain network for metal artifact reduction in CT images [J]. | MEDICAL IMAGE ANALYSIS , 2023 , 85 .
MLA Wang, Hong et al. "InDuDoNet plus : A deep unfolding dual domain network for metal artifact reduction in CT images" . | MEDICAL IMAGE ANALYSIS 85 (2023) .
APA Wang, Hong , Li, Yuexiang , Zhang, Haimiao , Meng, Deyu , Zheng, Yefeng . InDuDoNet plus : A deep unfolding dual domain network for metal artifact reduction in CT images . | MEDICAL IMAGE ANALYSIS , 2023 , 85 .
Export to NoteExpress RIS BibTex
Uncertainty-guided hierarchical frequency domain Transformer for image restoration EI SCIE Scopus
期刊论文 | 2023 , 263 | KNOWLEDGE-BASED SYSTEMS
SCOPUS Cited Count: 16
Abstract&Keyword Cite

Abstract :

Existing convolutional neural network (CNN)-based and vision Transformer (ViT)-based image restora-tion methods are usually explored in the spatial domain. However, we employ Fourier analysis to show that these spatial domain models cannot perceive the entire frequency spectrum of images, i.e., mainly focus on either high-frequency (CNN-based models) or low-frequency components (ViT-based models). This intrinsic limitation results in the partial missing of semantic information and the appearance of artifacts. To address this limitation, we propose a novel uncertainty-guided hierarchical frequency domain Transformer named HFDT to effectively learn both high and low-frequency information while perceiving local and global features. Specifically, to aggregate semantic information from various fre-quency levels, we propose a dual-domain feature interaction mechanism, in which the global frequency information and local spatial features are extracted by corresponding branches. The frequency domain branch adopts the Fast Fourier Transform (FFT) to convert the features from the spatial domain to the frequency domain, where the global low and high-frequency components are learned with Log -linear complexity. Complementarily, an efficient convolution group is employed in the spatial domain branch to capture local high-frequency details. Moreover, we introduce an uncertainty degradation -guided strategy to efficiently represent degraded prior information, rather than simply distinguishing degraded/non-degraded regions in binary form. Our approach achieves competitive results in several degraded scenarios, including rain streaks, raindrops, motion blur, and defocus blur.(c) 2023 Elsevier B.V. All rights reserved.

Keyword :

Frequency-domain Transformer Image restoration Log-linear complexity Uncertainty-guided

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Shao, Mingwen , Qiao, Yuanjian , Meng, Deyu et al. Uncertainty-guided hierarchical frequency domain Transformer for image restoration [J]. | KNOWLEDGE-BASED SYSTEMS , 2023 , 263 .
MLA Shao, Mingwen et al. "Uncertainty-guided hierarchical frequency domain Transformer for image restoration" . | KNOWLEDGE-BASED SYSTEMS 263 (2023) .
APA Shao, Mingwen , Qiao, Yuanjian , Meng, Deyu , Zuo, Wangmeng . Uncertainty-guided hierarchical frequency domain Transformer for image restoration . | KNOWLEDGE-BASED SYSTEMS , 2023 , 263 .
Export to NoteExpress RIS BibTex
Plenty is Plague: Fine-Grained Learning for Visual Question Answering SCIE
期刊论文 | 2022 , 44 (2) , 697-709 | IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Abstract&Keyword Cite

Abstract :

Visual Question Answering (VQA) has attracted extensive research focus recently. Along with the ever-increasing data scale and model complexity, the enormous training cost has become an emerging challenge for VQA. In this article, we show such a massive training cost is indeed plague. In contrast, a fine-grained design of the learning paradigm can be extremely beneficial in terms of both training efficiency and model accuracy. In particular, we argue that there exist two essential and unexplored issues in the existing VQA training paradigm that randomly samples data in each epoch, namely, the "difficulty diversity" and the "label redundancy". Concretely, "difficulty diversity" refers to the varying difficulty levels of different question types, while "label redundancy" refers to the redundant and noisy labels contained in individual question type. To tackle these two issues, in this article we propose a fine-grained VQA learning paradigm with an actor-critic based learning agent, termed FG-A1C. Instead of using all training data from scratch, FG-A1C includes a learning agent that adaptively and intelligently schedules the most difficult question types in each training epoch. Subsequently, two curriculum learning based schemes are further designed to identify the most useful data to be learned within each inidividual question type. We conduct extensive experiments on the VQA2.0 and VQA-CP v2 datasets, which demonstrate the significant benefits of our approach. For instance, on VQA-CP v2, with less than 75 percent of the training data, our learning paradigms can help the model achieves better performance than using the whole dataset. Meanwhile, we also shows the effectivenesss of our method in guiding data labeling. Finally, the proposed paradigm can be seamlessly integrated with any cutting-edge VQA models, without modifying their structures.

Keyword :

Data models Feature extraction Fine-grained learning Knowledge discovery Redundancy Training Training data Visualization visual question answering

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Zhou, Yiyi , Ji, Rongrong , Sun, Xiaoshuai et al. Plenty is Plague: Fine-Grained Learning for Visual Question Answering [J]. | IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE , 2022 , 44 (2) : 697-709 .
MLA Zhou, Yiyi et al. "Plenty is Plague: Fine-Grained Learning for Visual Question Answering" . | IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 44 . 2 (2022) : 697-709 .
APA Zhou, Yiyi , Ji, Rongrong , Sun, Xiaoshuai , Su, Jinsong , Meng, Deyu , Gao, Yue et al. Plenty is Plague: Fine-Grained Learning for Visual Question Answering . | IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE , 2022 , 44 (2) , 697-709 .
Export to NoteExpress RIS BibTex
Infrared Action Detection in the Dark via Cross-Stream Attention Mechanism EI SCIE Scopus
期刊论文 | 2022 , 24 , 288-300 | IEEE TRANSACTIONS ON MULTIMEDIA
SCOPUS Cited Count: 17
Abstract&Keyword Cite

Abstract :

Action detection plays an important role in video understanding and attracts considerable attention in the last decade. However, current action detection methods are mainly based on visible videos, and few of them consider scenes with low-light, where actions are difficult to be detected by existing methods, or even by human eyes. Compared with visible videos, infrared videos are more suitable for the dark environment and resistant to background clutter. In this paper, we investigate the temporal action detection problem in the dark by using infrared videos, which is, to the best of our knowledge, the first attempt in the action detection community. Our model takes the whole video as input, a Flow Estimation Network (FEN) is employed to generate the optical flow for infrared data, and it is optimized with the whole network to obtain action-related motion representations. After feature extraction, the infrared stream and flow stream are fed into a Selective Cross-stream Attention (SCA) module to narrow the performance gap between infrared and visible videos. The SCA emphasizes informative snippets and focuses on the more discriminative stream automatically. Then we adopt a snippet-level classifier to obtain action scores for all snippets and link continuous snippets into final detections. All these modules are trained in an end-to-end manner. We collect an Infrared action Detection (InfDet) dataset obtained in the dark and conduct extensive experiments to verify the effectiveness of the proposed method. Experimental results show that our proposed method surpasses state-of-the-art temporal action detection methods designed for visible videos, and it also achieves the best performance compared with other infrared action recognition methods on both InfAR and Infrared-Visible datasets.

Keyword :

Feature extraction Image recognition Infrared video Optical imaging Proposals selective cross-stream attention Streaming media Task analysis temporal action detection Three-dimensional displays

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Chen, Xu , Gao, Chenqiang , Li, Chaoyu et al. Infrared Action Detection in the Dark via Cross-Stream Attention Mechanism [J]. | IEEE TRANSACTIONS ON MULTIMEDIA , 2022 , 24 : 288-300 .
MLA Chen, Xu et al. "Infrared Action Detection in the Dark via Cross-Stream Attention Mechanism" . | IEEE TRANSACTIONS ON MULTIMEDIA 24 (2022) : 288-300 .
APA Chen, Xu , Gao, Chenqiang , Li, Chaoyu , Yang, Yi , Meng, Deyu . Infrared Action Detection in the Dark via Cross-Stream Attention Mechanism . | IEEE TRANSACTIONS ON MULTIMEDIA , 2022 , 24 , 288-300 .
Export to NoteExpress RIS BibTex
Orientation-Shared Convolution Representation for CT Metal Artifact Learning CPCI-S Scopus
期刊论文 | 2022 , 13436 , 665-675 | MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VI
SCOPUS Cited Count: 8
Abstract&Keyword Cite

Abstract :

During X-ray computed tomography (CT) scanning, metallic implants carrying with patients often lead to adverse artifacts in the captured CT images and then impair the clinical treatment. Against this metal artifact reduction (MAR) task, the existing deep-learning-based methods have gained promising reconstruction performance. Nevertheless, there is still some room for further improvement of MAR performance and generalization ability, since some important prior knowledge underlying this specific task has not been fully exploited. Hereby, in this paper, we carefully analyze the characteristics of metal artifacts and propose an orientation-shared convolution representation strategy to adapt the physical prior structures of artifacts, i.e., rotationally symmetrical streaking patterns. The proposed method rationally adopts Fourier-series-expansion-based filter parametrization in artifact modeling, which can better separate artifacts from anatomical tissues and boost the model generalizability. Comprehensive experiments executed on synthesized and clinical datasets show the superiority of our method in detail preservation beyond the current representative MAR methods. Code will be available at https://github.com/hongwang01/OSCNet.

Keyword :

Fourier series expansion Metal artifact reduction Model generalizability Orientation-shared convolution Rotation prior

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Wang, Hong , Xie, Qi , Li, Yuexiang et al. Orientation-Shared Convolution Representation for CT Metal Artifact Learning [J]. | MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VI , 2022 , 13436 : 665-675 .
MLA Wang, Hong et al. "Orientation-Shared Convolution Representation for CT Metal Artifact Learning" . | MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VI 13436 (2022) : 665-675 .
APA Wang, Hong , Xie, Qi , Li, Yuexiang , Huang, Yawen , Meng, Deyu , Zheng, Yefeng . Orientation-Shared Convolution Representation for CT Metal Artifact Learning . | MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VI , 2022 , 13436 , 665-675 .
Export to NoteExpress RIS BibTex
Sparsity-Enhanced Convolutional Decomposition: A Novel Tensor-Based Paradigm for Blind Hyperspectral Unmixing EI SCIE Scopus
期刊论文 | 2022 , 60 | IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING
SCOPUS Cited Count: 55
Abstract&Keyword Cite

Abstract :

Blind hyperspectral unmixing (HU) has long been recognized as a crucial component in analyzing the hyperspectral imagery (HSI) collected by airborne and spaceborne sensors. Due to the highly ill-posed problems of such a blind source separation scheme and the effects of spectral variability in hyperspectral imaging, the ability to accurately and effectively unmixing the complex HSI still remains limited. To this end, this article presents a novel blind HU model, called sparsity-enhanced convolutional decomposition (SeCoDe), by jointly capturing spatial-spectral information of HSI in a tensor-based fashion. SeCoDe benefits from two perspectives. On the one hand, the convolutional operation is employed in SeCoDe to locally model the spatial relation between the targeted pixel and its neighbors, which can be well explained by spectral bundles that are capable of addressing spectral variabilities effectively. It maintains, on the other hand, physically continuous spectral components by decomposing the HSI along with the spectral domain. With sparsity-enhanced regularization, an alternative optimization strategy with alternating direction method of multipliers (ADMM)-based optimization algorithm is devised for efficient model inference. Extensive experiments conducted on three different data sets demonstrate the superiority of the proposed SeCoDe compared to previous state-of-the-art methods. We will also release the code at https://github.com/danfenghong/IEEE_TGRS_SeCoDe to encourage the reproduction of the given results.

Keyword :

Blind hyperspectral unmixing (HU) Context modeling Convolutional codes convolutional sparse coding (CSC) Encoding Hyperspectral imaging Optimization spectral bundles spectral variability (SV) Task analysis tensor decomposition Tensors

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Yao, Jing , Hong, Danfeng , Xu, Lin et al. Sparsity-Enhanced Convolutional Decomposition: A Novel Tensor-Based Paradigm for Blind Hyperspectral Unmixing [J]. | IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING , 2022 , 60 .
MLA Yao, Jing et al. "Sparsity-Enhanced Convolutional Decomposition: A Novel Tensor-Based Paradigm for Blind Hyperspectral Unmixing" . | IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING 60 (2022) .
APA Yao, Jing , Hong, Danfeng , Xu, Lin , Meng, Deyu , Chanussot, Jocelyn , Xu, Zongben . Sparsity-Enhanced Convolutional Decomposition: A Novel Tensor-Based Paradigm for Blind Hyperspectral Unmixing . | IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING , 2022 , 60 .
Export to NoteExpress RIS BibTex
STEIN VARIATIONAL GRADIENT DESCENT ON INFINITE-DIMENSIONAL SPACE AND APPLICATIONS TO STATISTICAL INVERSE PROBLEMS EI SCIE Scopus
期刊论文 | 2022 , 60 (4) , 2225-2252 | SIAM JOURNAL ON NUMERICAL ANALYSIS
SCOPUS Cited Count: 3
Abstract&Keyword Cite

Abstract :

In this paper, we propose an infinite-dimensional version of the Stein variational gradient descent (iSVGD) method for solving Bayesian inverse problems. The method can generate approximate samples from posteriors efficiently. Based on the concepts of operator-valued kernels and vector-valued reproducing kernel Hilbert spaces, a rigorous definition is given for the infinite-dimensional objects, e.g., the Stein operator, which are proved to be the limit of finite-dimensional ones. Moreover, a more efficient iSVGD with preconditioning operators is constructed by generalizing the change of variables formula and introducing a regularity parameter. The proposed algorithms are applied to an inverse problem of the steady state Darcy flow equation. Numerical results confirm our theoretical findings and demonstrate the potential applications of the proposed approach in the posterior sampling of large-scale nonlinear statistical inverse problems.

Keyword :

Bayes? method machine learning statistical inverse problems Stein variational gradient descent variational inference method

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Jia, Junxiong , LI, Peijun , Meng, Deyu . STEIN VARIATIONAL GRADIENT DESCENT ON INFINITE-DIMENSIONAL SPACE AND APPLICATIONS TO STATISTICAL INVERSE PROBLEMS [J]. | SIAM JOURNAL ON NUMERICAL ANALYSIS , 2022 , 60 (4) : 2225-2252 .
MLA Jia, Junxiong et al. "STEIN VARIATIONAL GRADIENT DESCENT ON INFINITE-DIMENSIONAL SPACE AND APPLICATIONS TO STATISTICAL INVERSE PROBLEMS" . | SIAM JOURNAL ON NUMERICAL ANALYSIS 60 . 4 (2022) : 2225-2252 .
APA Jia, Junxiong , LI, Peijun , Meng, Deyu . STEIN VARIATIONAL GRADIENT DESCENT ON INFINITE-DIMENSIONAL SPACE AND APPLICATIONS TO STATISTICAL INVERSE PROBLEMS . | SIAM JOURNAL ON NUMERICAL ANALYSIS , 2022 , 60 (4) , 2225-2252 .
Export to NoteExpress RIS BibTex
Two-Stream Graph Convolutional Network for Intra-Oral Scanner Image Segmentation EI SCIE Scopus
期刊论文 | 2022 , 41 (4) , 826-835 | IEEE TRANSACTIONS ON MEDICAL IMAGING
SCOPUS Cited Count: 14
Abstract&Keyword Cite

Abstract :

Precise segmentation of teeth from intra-oral scanner images is an essential task in computer-aided orthodontic surgical planning. The state-of-the-art deep learning-based methods often simply concatenate the raw geometric attributes (i.e., coordinates and normal vectors) of mesh cells to train a single-stream network for automatic intra-oral scanner image segmentation. However, since different raw attributes reveal completely different geometric information, the naive concatenation of different raw attributes at the (low-level) input stage may bring unnecessary confusion in describing and differentiating between mesh cells, thus hampering the learning of high-level geometric representations for the segmentation task. To address this issue, we design a two-stream graph convolutional network (i.e., TSGCN), which can effectively handle inter-view confusion between different raw attributes to more effectively fuse their complementary information and learn discriminative multi-view geometric representations. Specifically, our TSGCN adopts two input-specific graph-learning streams to extract complementary high-level geometric representations from coordinates and normal vectors, respectively. Then, these single-view representations are further fused by a self-attention module to adaptively balance the contributions of different views in learning more discriminative multi-view representations for accurate and fully automatic tooth segmentation. We have evaluated our TSGCN on a real-patient dataset of dental (mesh) models acquired by 3D intraoral scanners. Experimental results show that our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.

Keyword :

Dentistry Feature extraction graph convolutional network Image segmentation Intra-oral scanner image segmentation Shape Task analysis Teeth Three-dimensional displays

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Zhao, Yue , Zhang, Lingming , Liu, Yang et al. Two-Stream Graph Convolutional Network for Intra-Oral Scanner Image Segmentation [J]. | IEEE TRANSACTIONS ON MEDICAL IMAGING , 2022 , 41 (4) : 826-835 .
MLA Zhao, Yue et al. "Two-Stream Graph Convolutional Network for Intra-Oral Scanner Image Segmentation" . | IEEE TRANSACTIONS ON MEDICAL IMAGING 41 . 4 (2022) : 826-835 .
APA Zhao, Yue , Zhang, Lingming , Liu, Yang , Meng, Deyu , Cui, Zhiming , Gao, Chenqiang et al. Two-Stream Graph Convolutional Network for Intra-Oral Scanner Image Segmentation . | IEEE TRANSACTIONS ON MEDICAL IMAGING , 2022 , 41 (4) , 826-835 .
Export to NoteExpress RIS BibTex
DICDNet: Deep Interpretable Convolutional Dictionary Network for Metal Artifact Reduction in CT Images EI SCIE Scopus
期刊论文 | 2022 , 41 (4) , 869-880 | IEEE TRANSACTIONS ON MEDICAL IMAGING
SCOPUS Cited Count: 23
Abstract&Keyword Cite

Abstract :

Computed tomography (CT) images are often impaired by unfavorable artifacts caused by metallic implants within patients, which would adversely affect the subsequent clinical diagnosis and treatment. Although the existing deep-learning-based approaches have achieved promising success on metal artifact reduction (MAR) for CT images, most of them treated the task as a general image restoration problem and utilized off-the-shelf network modules for image quality enhancement. Hence, such frameworks always suffer from lack of sufficient model interpretability for the specific task. Besides, the existing MAR techniques largely neglect the intrinsic prior knowledge underlying metal-corrupted CT images which is beneficial for the MAR performance improvement. In this paper, we specifically propose a deep interpretable convolutional dictionary network (DICDNet) for the MAR task. Particularly, we first explore that the metal artifacts always present non-local streaking and star-shape patterns in CT images. Based on such observations, a convolutional dictionary model is deployed to encode the metal artifacts. To solve the model, we propose a novel optimization algorithm based on the proximal gradient technique. With only simple operators, the iterative steps of the proposed algorithm can be easily unfolded into corresponding network modules with specific physical meanings. Comprehensive experiments on synthesized and clinical datasets substantiate the effectiveness of the proposed DICDNet as well as its superior interpretability, compared to current state-of-the-art MAR methods. Code is available at https://github.com/hongwang01/DICDNet.

Keyword :

Computed tomography CT metal artifact reduction Dictionaries generalization performance Image reconstruction interpretable dictionary learning Mars Metals Optimization Task analysis

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Wang, Hong , Li, Yuexiang , He, Nanjun et al. DICDNet: Deep Interpretable Convolutional Dictionary Network for Metal Artifact Reduction in CT Images [J]. | IEEE TRANSACTIONS ON MEDICAL IMAGING , 2022 , 41 (4) : 869-880 .
MLA Wang, Hong et al. "DICDNet: Deep Interpretable Convolutional Dictionary Network for Metal Artifact Reduction in CT Images" . | IEEE TRANSACTIONS ON MEDICAL IMAGING 41 . 4 (2022) : 869-880 .
APA Wang, Hong , Li, Yuexiang , He, Nanjun , Ma, Kai , Meng, Deyu , Zheng, Yefeng . DICDNet: Deep Interpretable Convolutional Dictionary Network for Metal Artifact Reduction in CT Images . | IEEE TRANSACTIONS ON MEDICAL IMAGING , 2022 , 41 (4) , 869-880 .
Export to NoteExpress RIS BibTex
10| 20| 50 per page
< Page ,Total 20 >

Export

Results:

Selected

to

Format:
FAQ| About| Online/Total:1490/188895687
Address:XI'AN JIAOTONG UNIVERSITY LIBRARY(No.28, Xianning West Road, Xi'an, Shaanxi Post Code:710049) Contact Us:029-82667865
Copyright:XI'AN JIAOTONG UNIVERSITY LIBRARY Technical Support:Beijing Aegean Software Co., Ltd.