Multimodal medical image fusion can help physicians provide more accurate treatment plans for patients, as unimodal images provide limited valid information. To address the insufficient ability of traditional medical ...Multimodal medical image fusion can help physicians provide more accurate treatment plans for patients, as unimodal images provide limited valid information. To address the insufficient ability of traditional medical image fusion solutions to protect image details and significant information, a new multimodality medical image fusion method(NSST-PAPCNNLatLRR) is proposed in this paper. Firstly, the high and low-frequency sub-band coefficients are obtained by decomposing the source image using NSST. Then, the latent low-rank representation algorithm is used to process the low-frequency sub-band coefficients;An improved PAPCNN algorithm is also proposed for the fusion of high-frequency sub-band coefficients. The improved PAPCNN model was based on the automatic setting of the parameters, and the optimal method was configured for the time decay factor αe. The experimental results show that, in comparison with the five mainstream fusion algorithms, the new algorithm has significantly improved the visual effect over the comparison algorithm,enhanced the ability to characterize important information in images, and further improved the ability to protect the detailed information;the new algorithm has achieved at least four firsts in six objective indexes.展开更多
Medical image fusion has been developed as an efficient assistive technology in various clinical applications such as medical diagnosis and treatment planning.Aiming at the problem of insufficient protection of image ...Medical image fusion has been developed as an efficient assistive technology in various clinical applications such as medical diagnosis and treatment planning.Aiming at the problem of insufficient protection of image contour and detail information by traditional image fusion methods,a new multimodal medical image fusion method is proposed.This method first uses non-subsampled shearlet transform to decompose the source image to obtain high and low frequency subband coefficients,then uses the latent low rank representation algorithm to fuse the low frequency subband coefficients,and applies the improved PAPCNN algorithm to fuse the high frequency subband coefficients.Finally,based on the automatic setting of parameters,the optimization method configuration of the time decay factorαe is carried out.The experimental results show that the proposed method solves the problems of difficult parameter setting and insufficient detail protection ability in traditional PCNN algorithm fusion images,and at the same time,it has achieved great improvement in visual quality and objective evaluation indicators.展开更多
User representation learning is crucial for capturing different user preferences,but it is also critical challenging because user intentions are latent and dispersed in complex and different patterns of user-generated...User representation learning is crucial for capturing different user preferences,but it is also critical challenging because user intentions are latent and dispersed in complex and different patterns of user-generated data,and thus cannot be measured directly.Text-based data models can learn user representations by mining latent semantics,which is beneficial to enhancing the semantic function of user representations.However,these technologies only extract common features in historical records and cannot represent changes in user intentions.However,sequential feature can express the user’s interests and intentions that change time by time.But the sequential recommendation results based on the user representation of the item lack the interpretability of preference factors.To address these issues,we propose in this paper a novel model with Dual-Layer User Representation,named DLUR,where the user’s intention is learned based on two different layer representations.Specifically,the latent semantic layer adds an interactive layer based on Transformer to extract keywords and key sentences in the text and serve as a basis for interpretation.The sequence layer uses the Transformer model to encode the user’s preference intention to clarify changes in the user’s intention.Therefore,this dual-layer user mode is more comprehensive than a single text mode or sequence mode and can effectually improve the performance of recommendations.Our extensive experiments on five benchmark datasets demonstrate DLUR’s performance over state-of-the-art recommendation models.In addition,DLUR’s ability to explain recommendation results is also demonstrated through some specific cases.展开更多
Low-Rank and Sparse Representation(LRSR)method has gained popularity in Hyperspectral Image(HSI)processing.However,existing LRSR models rarely exploited spectral-spatial classification of HSI.In this paper,we proposed...Low-Rank and Sparse Representation(LRSR)method has gained popularity in Hyperspectral Image(HSI)processing.However,existing LRSR models rarely exploited spectral-spatial classification of HSI.In this paper,we proposed a novel Low-Rank and Sparse Representation with Adaptive Neighborhood Regularization(LRSR-ANR)method for HSI classification.In the proposed method,we first represent the hyperspectral data via LRSR since it combines both sparsity and low-rankness to maintain global and local data structures simultaneously.The LRSR is optimized by using a mixed Gauss-Seidel and Jacobian Alternating Direction Method of Multipliers(M-ADMM),which converges faster than ADMM.Then to incorporate the spatial information,an ANR scheme is designed by combining Euclidean and Cosine distance metrics to reduce the mixed pixels within a neighborhood.Lastly,the predicted labels are determined by jointly considering the homogeneous pixels in the classification rule of the minimum reconstruction error.Experimental results based on three popular hyperspectral images demonstrate that the proposed method outperforms other related methods in terms of classification accuracy and generalization performance.展开更多
The conventional sparse representation-based image classification usually codes the samples independently,which will ignore the correlation information existed in the data.Hence,if we can explore the correlation infor...The conventional sparse representation-based image classification usually codes the samples independently,which will ignore the correlation information existed in the data.Hence,if we can explore the correlation information hidden in the data,the classification result will be improved significantly.To this end,in this paper,a novel weighted supervised spare coding method is proposed to address the image classification problem.The proposed method firstly explores the structural information sufficiently hidden in the data based on the low rank representation.And then,it introduced the extracted structural information to a novel weighted sparse representation model to code the samples in a supervised way.Experimental results show that the proposed method is superiority to many conventional image classification methods.展开更多
针对潜在低秩表示学习的投影矩阵不能解释提取特征重要程度和保持数据的局部几何结构的问题,提出了一种基于双邻域和特征选择的潜在低秩稀疏投影算法(LLRSP:Latent Low-Rank And Sparse Projection)。该算法首先融合低秩约束和正交重构...针对潜在低秩表示学习的投影矩阵不能解释提取特征重要程度和保持数据的局部几何结构的问题,提出了一种基于双邻域和特征选择的潜在低秩稀疏投影算法(LLRSP:Latent Low-Rank And Sparse Projection)。该算法首先融合低秩约束和正交重构保持数据的主要能量,然后对投影矩阵施加行稀疏约束进行特征选择,使特征更加紧凑和具有可解释性。此外引入l_(2,1)范数对误差分量进行正则化使模型对噪声更具健壮性。最后在低维数据和低秩表示系数矩阵上施加邻域保持正则化以保留数据的局部几何结构。公开数据集上的大量实验结果表明,所提方法与其他先进算法相比具有更好的性能。展开更多
Indoor environment quality(IEQ)is one of the most concerned building performances during the operation stage.The non-uniform spatial distribution of various IEQ parameters in large-scale public buildings has been demo...Indoor environment quality(IEQ)is one of the most concerned building performances during the operation stage.The non-uniform spatial distribution of various IEQ parameters in large-scale public buildings has been demonstrated to be an essential factor affecting occupant comfort and building energy consumption.Currently,IEQ sensors have been widely employed in buildings to monitor thermal,visual,acoustic and air quality.However,there is a lack of effective methods for exploring the typical spatial distribution of indoor environmental quality parameters,which is crucial for assessing and controlling non-uniform indoor environments.In this study,a novel clustering method for extracting IEQ spatial distribution patterns is proposed.Firstly,representation vectors reflecting IEQ distributions in the concerned space are generated based on the low-rank sparse representation.Secondly,a multi-step clustering method,which addressed the problems of the“curse of dimensionality”,is designed to obtain typical IEQ distribution patterns of the entire indoor space.The proposed method was applied to the analysis of indoor thermal environment in Beijing Daxing international airport terminal.As a result,four typical temperature spatial distribution patterns of the terminal were extracted from a four-month monitoring,which had been validated for their good representativeness.These typical patterns revealed typical environmental issues in the terminal,such as long-term localized overheating and temperature increases due to a sudden influx of people.The extracted typical IEQ spatial distribution patterns could assist building operators in effectively assessing the uneven distribution of IEQ space under current environmental conditions,facilitating targeted environmental improvements,optimization of thermal comfort levels,and application of energy-saving measures.展开更多
The task of dividing corrupted-data into their respective subspaces can be well illustrated,both theoretically and numerically,by recovering low-rank and sparse-column components of a given matrix.Generally,it can be ...The task of dividing corrupted-data into their respective subspaces can be well illustrated,both theoretically and numerically,by recovering low-rank and sparse-column components of a given matrix.Generally,it can be characterized as a matrix and a 2,1-norm involved convex minimization problem.However,solving the resulting problem is full of challenges due to the non-smoothness of the objective function.One of the earliest solvers is an 3-block alternating direction method of multipliers(ADMM)which updates each variable in a Gauss-Seidel manner.In this paper,we present three variants of ADMM for the 3-block separable minimization problem.More preciously,whenever one variable is derived,the resulting problems can be regarded as a convex minimization with 2 blocks,and can be solved immediately using the standard ADMM.If the inner iteration loops only once,the iterative scheme reduces to the ADMM with updates in a Gauss-Seidel manner.If the solution from the inner iteration is assumed to be exact,the convergence can be deduced easily in the literature.The performance comparisons with a couple of recently designed solvers illustrate that the proposed methods are effective and competitive.展开更多
The rapid expansion of online content and big data has precipitated an urgent need for efficient summarization techniques to swiftly comprehend vast textual documents without compromising their original integrity.Curr...The rapid expansion of online content and big data has precipitated an urgent need for efficient summarization techniques to swiftly comprehend vast textual documents without compromising their original integrity.Current approaches in Extractive Text Summarization(ETS)leverage the modeling of inter-sentence relationships,a task of paramount importance in producing coherent summaries.This study introduces an innovative model that integrates Graph Attention Networks(GATs)with Transformer-based Bidirectional Encoder Representa-tions from Transformers(BERT)and Latent Dirichlet Allocation(LDA),further enhanced by Term Frequency-Inverse Document Frequency(TF-IDF)values,to improve sentence selection by capturing comprehensive topical information.Our approach constructs a graph with nodes representing sentences,words,and topics,thereby elevating the interconnectivity and enabling a more refined understanding of text structures.This model is stretched to Multi-Document Summarization(MDS)from Single-Document Summarization,offering significant improvements over existing models such as THGS-GMM and Topic-GraphSum,as demonstrated by empirical evaluations on benchmark news datasets like Cable News Network(CNN)/Daily Mail(DM)and Multi-News.The results consistently demonstrate superior performance,showcasing the model’s robustness in handling complex summarization tasks across single and multi-document contexts.This research not only advances the integration of BERT and LDA within a GATs but also emphasizes our model’s capacity to effectively manage global information and adapt to diverse summarization challenges.展开更多
Extracting discriminative speaker-specific representations from speech signals and transforming them into fixed length vectors are key steps in speaker identification and verification systems.In this study,we propose ...Extracting discriminative speaker-specific representations from speech signals and transforming them into fixed length vectors are key steps in speaker identification and verification systems.In this study,we propose a latent discriminative representation learning method for speaker recognition.We mean that the learned representations in this study are not only discriminative but also relevant.Specifically,we introduce an additional speaker embedded lookup table to explore the relevance between different utterances from the same speaker.Moreover,a reconstruction constraint intended to learn a linear mapping matrix is introduced to make representation discriminative.Experimental results demonstrate that the proposed method outperforms state-of-the-art methods based on the Apollo dataset used in the Fearless Steps Challenge in INTERSPEECH2019 and the TIMIT dataset.展开更多
基金funded by the National Natural Science Foundation of China,grant number 61302188.
文摘Multimodal medical image fusion can help physicians provide more accurate treatment plans for patients, as unimodal images provide limited valid information. To address the insufficient ability of traditional medical image fusion solutions to protect image details and significant information, a new multimodality medical image fusion method(NSST-PAPCNNLatLRR) is proposed in this paper. Firstly, the high and low-frequency sub-band coefficients are obtained by decomposing the source image using NSST. Then, the latent low-rank representation algorithm is used to process the low-frequency sub-band coefficients;An improved PAPCNN algorithm is also proposed for the fusion of high-frequency sub-band coefficients. The improved PAPCNN model was based on the automatic setting of the parameters, and the optimal method was configured for the time decay factor αe. The experimental results show that, in comparison with the five mainstream fusion algorithms, the new algorithm has significantly improved the visual effect over the comparison algorithm,enhanced the ability to characterize important information in images, and further improved the ability to protect the detailed information;the new algorithm has achieved at least four firsts in six objective indexes.
文摘Medical image fusion has been developed as an efficient assistive technology in various clinical applications such as medical diagnosis and treatment planning.Aiming at the problem of insufficient protection of image contour and detail information by traditional image fusion methods,a new multimodal medical image fusion method is proposed.This method first uses non-subsampled shearlet transform to decompose the source image to obtain high and low frequency subband coefficients,then uses the latent low rank representation algorithm to fuse the low frequency subband coefficients,and applies the improved PAPCNN algorithm to fuse the high frequency subband coefficients.Finally,based on the automatic setting of parameters,the optimization method configuration of the time decay factorαe is carried out.The experimental results show that the proposed method solves the problems of difficult parameter setting and insufficient detail protection ability in traditional PCNN algorithm fusion images,and at the same time,it has achieved great improvement in visual quality and objective evaluation indicators.
基金supported by the Applied Research Center of Artificial Intelligence,Wuhan College(Grant Number X2020113)the Wuhan College Research Project(Grant Number KYZ202009).
文摘User representation learning is crucial for capturing different user preferences,but it is also critical challenging because user intentions are latent and dispersed in complex and different patterns of user-generated data,and thus cannot be measured directly.Text-based data models can learn user representations by mining latent semantics,which is beneficial to enhancing the semantic function of user representations.However,these technologies only extract common features in historical records and cannot represent changes in user intentions.However,sequential feature can express the user’s interests and intentions that change time by time.But the sequential recommendation results based on the user representation of the item lack the interpretability of preference factors.To address these issues,we propose in this paper a novel model with Dual-Layer User Representation,named DLUR,where the user’s intention is learned based on two different layer representations.Specifically,the latent semantic layer adds an interactive layer based on Transformer to extract keywords and key sentences in the text and serve as a basis for interpretation.The sequence layer uses the Transformer model to encode the user’s preference intention to clarify changes in the user’s intention.Therefore,this dual-layer user mode is more comprehensive than a single text mode or sequence mode and can effectually improve the performance of recommendations.Our extensive experiments on five benchmark datasets demonstrate DLUR’s performance over state-of-the-art recommendation models.In addition,DLUR’s ability to explain recommendation results is also demonstrated through some specific cases.
基金National Natural Foundation of China(No.41971279)Fundamental Research Funds of the Central Universities(No.B200202012)。
文摘Low-Rank and Sparse Representation(LRSR)method has gained popularity in Hyperspectral Image(HSI)processing.However,existing LRSR models rarely exploited spectral-spatial classification of HSI.In this paper,we proposed a novel Low-Rank and Sparse Representation with Adaptive Neighborhood Regularization(LRSR-ANR)method for HSI classification.In the proposed method,we first represent the hyperspectral data via LRSR since it combines both sparsity and low-rankness to maintain global and local data structures simultaneously.The LRSR is optimized by using a mixed Gauss-Seidel and Jacobian Alternating Direction Method of Multipliers(M-ADMM),which converges faster than ADMM.Then to incorporate the spatial information,an ANR scheme is designed by combining Euclidean and Cosine distance metrics to reduce the mixed pixels within a neighborhood.Lastly,the predicted labels are determined by jointly considering the homogeneous pixels in the classification rule of the minimum reconstruction error.Experimental results based on three popular hyperspectral images demonstrate that the proposed method outperforms other related methods in terms of classification accuracy and generalization performance.
基金This research is funded by the National Natural Science Foundation of China(61771154).
文摘The conventional sparse representation-based image classification usually codes the samples independently,which will ignore the correlation information existed in the data.Hence,if we can explore the correlation information hidden in the data,the classification result will be improved significantly.To this end,in this paper,a novel weighted supervised spare coding method is proposed to address the image classification problem.The proposed method firstly explores the structural information sufficiently hidden in the data based on the low rank representation.And then,it introduced the extracted structural information to a novel weighted sparse representation model to code the samples in a supervised way.Experimental results show that the proposed method is superiority to many conventional image classification methods.
文摘针对潜在低秩表示学习的投影矩阵不能解释提取特征重要程度和保持数据的局部几何结构的问题,提出了一种基于双邻域和特征选择的潜在低秩稀疏投影算法(LLRSP:Latent Low-Rank And Sparse Projection)。该算法首先融合低秩约束和正交重构保持数据的主要能量,然后对投影矩阵施加行稀疏约束进行特征选择,使特征更加紧凑和具有可解释性。此外引入l_(2,1)范数对误差分量进行正则化使模型对噪声更具健壮性。最后在低维数据和低秩表示系数矩阵上施加邻域保持正则化以保留数据的局部几何结构。公开数据集上的大量实验结果表明,所提方法与其他先进算法相比具有更好的性能。
基金the China National Key Research and Development Program(Grant No.2022YFC3801300)the Young Scientists Fund of the National Natural Science Foundation of China(Grant No.52208113)+1 种基金the Key Program of National Natural Science Foundation of China(Grant No.52130803)the Hang Lung Center for Real Estate,Tsinghua University.The authors also express special thanks to the Command Center of Beijing Daxing International Airport for their long-term and strong support to this research.
文摘Indoor environment quality(IEQ)is one of the most concerned building performances during the operation stage.The non-uniform spatial distribution of various IEQ parameters in large-scale public buildings has been demonstrated to be an essential factor affecting occupant comfort and building energy consumption.Currently,IEQ sensors have been widely employed in buildings to monitor thermal,visual,acoustic and air quality.However,there is a lack of effective methods for exploring the typical spatial distribution of indoor environmental quality parameters,which is crucial for assessing and controlling non-uniform indoor environments.In this study,a novel clustering method for extracting IEQ spatial distribution patterns is proposed.Firstly,representation vectors reflecting IEQ distributions in the concerned space are generated based on the low-rank sparse representation.Secondly,a multi-step clustering method,which addressed the problems of the“curse of dimensionality”,is designed to obtain typical IEQ distribution patterns of the entire indoor space.The proposed method was applied to the analysis of indoor thermal environment in Beijing Daxing international airport terminal.As a result,four typical temperature spatial distribution patterns of the terminal were extracted from a four-month monitoring,which had been validated for their good representativeness.These typical patterns revealed typical environmental issues in the terminal,such as long-term localized overheating and temperature increases due to a sudden influx of people.The extracted typical IEQ spatial distribution patterns could assist building operators in effectively assessing the uneven distribution of IEQ space under current environmental conditions,facilitating targeted environmental improvements,optimization of thermal comfort levels,and application of energy-saving measures.
基金Supported by the National Natural Science Foundation of China(Grant No.11971149,11871381)Natural Science Foundation of Henan Province for Youth(Grant No.202300410146)。
文摘The task of dividing corrupted-data into their respective subspaces can be well illustrated,both theoretically and numerically,by recovering low-rank and sparse-column components of a given matrix.Generally,it can be characterized as a matrix and a 2,1-norm involved convex minimization problem.However,solving the resulting problem is full of challenges due to the non-smoothness of the objective function.One of the earliest solvers is an 3-block alternating direction method of multipliers(ADMM)which updates each variable in a Gauss-Seidel manner.In this paper,we present three variants of ADMM for the 3-block separable minimization problem.More preciously,whenever one variable is derived,the resulting problems can be regarded as a convex minimization with 2 blocks,and can be solved immediately using the standard ADMM.If the inner iteration loops only once,the iterative scheme reduces to the ADMM with updates in a Gauss-Seidel manner.If the solution from the inner iteration is assumed to be exact,the convergence can be deduced easily in the literature.The performance comparisons with a couple of recently designed solvers illustrate that the proposed methods are effective and competitive.
文摘The rapid expansion of online content and big data has precipitated an urgent need for efficient summarization techniques to swiftly comprehend vast textual documents without compromising their original integrity.Current approaches in Extractive Text Summarization(ETS)leverage the modeling of inter-sentence relationships,a task of paramount importance in producing coherent summaries.This study introduces an innovative model that integrates Graph Attention Networks(GATs)with Transformer-based Bidirectional Encoder Representa-tions from Transformers(BERT)and Latent Dirichlet Allocation(LDA),further enhanced by Term Frequency-Inverse Document Frequency(TF-IDF)values,to improve sentence selection by capturing comprehensive topical information.Our approach constructs a graph with nodes representing sentences,words,and topics,thereby elevating the interconnectivity and enabling a more refined understanding of text structures.This model is stretched to Multi-Document Summarization(MDS)from Single-Document Summarization,offering significant improvements over existing models such as THGS-GMM and Topic-GraphSum,as demonstrated by empirical evaluations on benchmark news datasets like Cable News Network(CNN)/Daily Mail(DM)and Multi-News.The results consistently demonstrate superior performance,showcasing the model’s robustness in handling complex summarization tasks across single and multi-document contexts.This research not only advances the integration of BERT and LDA within a GATs but also emphasizes our model’s capacity to effectively manage global information and adapt to diverse summarization challenges.
基金Project supported by the National Natural Science Foundation of China(Nos.U1836220 and 61672267)the Qing Lan Talent Program of Jiangsu Province,Chinathe Jiangsu Province Key Research and Development Plan(Industry Foresight and Key Core Technology)(No.BE2020036)。
文摘Extracting discriminative speaker-specific representations from speech signals and transforming them into fixed length vectors are key steps in speaker identification and verification systems.In this study,we propose a latent discriminative representation learning method for speaker recognition.We mean that the learned representations in this study are not only discriminative but also relevant.Specifically,we introduce an additional speaker embedded lookup table to explore the relevance between different utterances from the same speaker.Moreover,a reconstruction constraint intended to learn a linear mapping matrix is introduced to make representation discriminative.Experimental results demonstrate that the proposed method outperforms state-of-the-art methods based on the Apollo dataset used in the Fearless Steps Challenge in INTERSPEECH2019 and the TIMIT dataset.