•   按检索    检索词:    高级检索
     排序:相关度 OA 时间 被引次数 点击次数 下载次数 共有10000条符合的查询结果,以下是第261-280项 搜索用时 135 毫秒
[首页] « 上一页 [9] [10] [11] [12] [13] 14 [15] [16] [17] [18] [19] 下一页 » 末  页»
261.
  
超级计算机是“国之重器”,我国在“十四五”期间建设后E级国产超算,支撑关系国计民生的重大计算应用。操作系统作为超算核心系统软件之一,其开销将影响超算整机的运行性能,因此操作系统测评成为新一代国产超算技术路线的重要研究课题之一。openEuler在搭载了鲲鹏处理器的系统上有良好的性… …   相似文献
262.
  
随着云存储服务的快速发展,越来越多的数据拥有者愿意将数据存储到云服务器中,从而减小自己在本地的存储负担。然而,一旦数据拥有者上传数据至云服务器,本地将不保存数据,数据拥有者将失去对数据的直接控制权。为了保证保存在云服务器上远程数据的完整性,数据完整性检验是必不可少的。它可以使得数… …   相似文献
263.
  
鉴于边缘AI的高性能与低功耗需求,基于 RISC-V 指令集架构,针对边缘设备数字信号处理的实际问题,设计了一种边缘AI的专用指令集处理器,在有限的硬件开销下,提升了边缘AI的执行效率,降低了边缘AI的能量消耗,能够满足边缘AI应用中进行高效大语言模型(LLM) 推理计算的需求。… …   相似文献
264.
  
遥感图像的空间分辨率高,不同类型对象的尺度差异大、类别不平衡,是精准语义分割任务所面临的主要挑战。为了提高遥感图像语义分割的准确性,提出了一种改进U-Net的多尺度特征融合遥感图像语义分割网络(Multi-scale Feature Fusion Network,MFFNet)。… …   相似文献
265.
  
知识图谱补全旨在预测给定三元组中缺失的实体和关系,以增强知识图谱的完整性和质量。现有的知识图谱补全方法通常只考虑三元组自身的结构信息或者是实体单一的附加信息(如实体的文本描述或拓扑结构信息),而忽略了融合多种附加信息来增强实体的特征信息,从而导致现有方法补全缺失实体时性能不佳。针… …   相似文献
266.
ObjectiveSingle-modality medical imaging is often insufficient for providing a comprehensive review of lesion characteristics, including structure, metabolism, and other critical details. Medical images can generally be categorized into anatomical medical imaging and functional medical imaging. Anatomical medical imaging offers rich information on the structure of the body, but it lacks insight into metabolic processes. In contrast, functional medical imaging is the opposite. In clinical applications, doctors use medical imaging from multiple modalities to diagnose diseases, localize lesions, and plan surgeries. However, simultaneously observing multimodal medical images is not intuitive and may not fully capture all the relevant features of the lesion. Therefore, multimodal medical image fusion is commonly employed in practice to integrate and enhance the information from different imaging techniques. How to fully retain the unique features of each modality while effectively integrating the shared features between modalities is a common challenge in medical image fusion. The information interaction of shared modal features in currently used two-branch image coding methods is often underdeveloped, and the process is somewhat inadequate. This condition limits the establishment of feature correlations between multimodal images. A multiscale medical image fusion network is designed to address these issues. This network is based on progressive feature extraction, frequency domain information supplementation, and image reconstruction by Swin Transformer and convolutional neural network(CNN).MethodFirst, a multiscale feature extraction module guided by gradient information was designed, which can be integrated into a three-branch feature extraction architecture. The left and right branches are responsible for extracting the unique features from each modality of the medical images, while the middle branch extracts the shared features between modalities. The extraction architecture comprises several multiscale feature extraction modules, each based on gradient information guidance. These submodule can simultaneously integrate features from all scale levels. The extraction architecture fully considers the information interaction between modalities and can progressively extract the common and unique features across different modalities. In addition, this extraction architecture effectively integrates multiscale features from multimodal medical images. A progressive fusion module that integrates cross-attention mechanisms was designed to fully utilize the frequency domain information and guide the fusion process at the modal level. This fusion module enhances the interaction of spatial domain information between different modalities and leverages high- and low-frequency positional information from the frequency domain, guiding the model for more targeted multimodal fusion. Finally, a Swin-CNN reconstruction module was designed to determine the relationship between global and local area features of medical images. The reconstruction module uses Swin Transformer to capture global information, such as the overall structure and shape of the image, while simultaneously employing CNN to extract regional features, such as local texture details. The reconstruction module can effectively improve the quality of fused images by integrating the global and local feature information of medical images simultaneously.ResultThe datasets used for the experiments include the MRI-SPECT and MRI-PET fusion datasets from the whole brain database at Harvard Medical School and the GFP-PC fusion dataset from the John Innes Center, respectively. Considering the visual effect of the fused images, the proposed fusion model effectively preserves the structural and functional features of different medical image modalities and improves the quality of the fused images. The advantages of the fused images generated by this model are as follows: 1) The fused image has richer texture details and sharper features such as edges and contours. These images effectively preserve the information-rich regions of each modal image. 2) The fused image also effectively preserves the visual features in all original medical images, which ensures no bias toward preserving information from only one modality of the medical image. 3) The fused image is rendered effectively, with no artifacts affecting the visual effect. In addition, in terms of comparison of quantitative indicators, the model achieves optimization for all eight image fusion evaluation metrics in MRI-SPECT and MRI-PET fusion tasks. Compared to the model with the second-best performance, the mutual information (MI) and discrete cosine transform feature mutual information (FMIdct) are drastically improved. MI demonstrated an improvement of 4.42% and 17.30%, respectively, and FMIdct showed improvements of 5.17% and 11%, respectively. In the GFP-PC fusion task, six optimal and two sub-optimal results are achieved. Compared to the model with the second-best performance, MI and visual information fidelity (VIF) are substantially improved by 16.43% and 16.87%, respectively. Ablation experiments were also conducted for the network structure and loss function of the model to effectively analyze the experimental results and evaluate the effectiveness of each part of the model in this paper. Experimental results show that all model components and the loss function enhance the image fusion effect.ConclusionThe proposed fusion model leverages the common and unique features of different medical image modalities and progressively integrates multiscale information using a three-branch architecture. The model also utilizes a progressive fusion module that incorporates cross-attention to fuse high- and low-frequency features in a highly targeted manner. Furthermore, the model focuses on the global and local attribute information of medical images in the reconstruction process, effectively enhancing the quality of multimodal medical image fusion. The proposed model in this paper performs well in three medical image fusion tasks with good generalization capability. This model can provide multimodal medical fusion images with clear contour structures and rich texture details, aiding doctors in clinical diagnosis and improving diagnostic efficiency and accuracy. Future studies will investigate the constraints or effects of downstream medical semantic segmentation and other tasks on image fusion. The network architecture will also be optimized for specific tasks, ensuring a close integration between tasks such as semantic segmentation and image fusion. This research aims to improve the quality of fused images while enhancing the performance of downstream tasks, thereby expanding the application possibilities of multimodal medical image fusion.… …   相似文献
《中国图象图形学报》2025,30(5):1510-1527
267.
ObjectiveUltrasound imaging plays a crucial role in medical diagnosis due to its convenience, non-invasive nature, and cost-effectiveness, m… …   相似文献
《中国图象图形学报》2025,30(5):1303-1317
268.
ObjectiveRegong art, originating from the Longwu River valley in the Tibetan region of Huangnan, Qinghai Province, has flourished in this ar… …   相似文献
《中国图象图形学报》2025,30(5):1377-1388
269.
ObjectiveColorectal cancer, a high-incidence and extremely harmful disease, represents a serious threat to human health. Statistics show tha… …   相似文献
《中国图象图形学报》2025,30(5):1479-1496
270.
ObjectiveCeladon is not only a dazzling pearl among the cultural treasures of the Chinese nation but also a cultural messenger in cultural exchanges between China and other countries. It has rich historical and cultural connotations and demonstrates excellent artistic value. Its elegant shape and moist glaze make it an outstanding representative of traditional Chinese craft aesthetics. The production of celadon embodies the wisdom and creativity of ancient craftsmen and is an important carrier for the inheritance of excellent traditional Chinese culture. In the context of cultural digitization, constructing a cross-modal knowledge graph of celadon is one of the key technologies for promoting the protection and inheritance of celadon culture. In this process, matching the same entities across different modalities, which involves aligning the different modal features of equivalent entities, is crucial. However, the inherent structural differences between cross-modal data present challenges for alignment tasks. Traditional methods that rely on manually annotated data can ensure the accuracy of alignment to some extent, but they have problems such as low efficiency and high cost. In addition, coarse-grained annotated data can hardly meet the requirements for fine-grained concepts and for entity recognition when constructing a cross-modal knowledge graph. At present, the vision-language pretraining (VLP) model can effectively capture cross-modal semantic associations by learning rich cross-modal representations from large-scale unmarked image-text pair data. The strong cross-modal understanding ability of the VLP model can provide precise semantic associations and fine-grained entity recognition for aligning entities of different modalities in graph construction. Here, a cross-modal entity alignment method based on the VLP model, which can map multiple features of images, is proposed to maximize the degree of matching between celadon images and text.MethodThe cross-modal entity alignment method proposed in this study, which maps multiple features of images, is initialized with the publicly available VLP model for both the image and the text encoders, and the parameters of the encoders remain unchanged during the training process. The method mainly consists of four parts. First, on the basis of the visual characteristics of celadon images, local features in terms of contour, texture, and color are extracted. Then, a gated multifusion unit is introduced to adaptively assign weights to the image features, and the extracted multiple local image features are used to generate reliable fused features. Furthermore, a multilayer fully connected mapper is designed to learn the mapping of the fused features to an appropriate intermediate representation space by using multiple layers of nonlinear transformations, guiding the text encoder to generate text features that match the image features more closely. Finally, the model is trained and optimized via the information noise contrastive estimation loss function, that is, by optimizing the similarity of positive sample pairs and the difference in negative sample pairs through calculating the cosine similarity between cross-modality features, thereby establishing the connection between image features and text features.ResultThe proposed method was compared with four of the latest benchmark methods in an experimental comparison, namely, contrastive VLP in Chinese (CN-CLIP), context optimization (CoOp), conditional context optimization (CoCoOp), and mapping pictures to words (Pic2Word). The quantitative evaluation metrics are the recall rates, including R@1, R@5, R@10, and the mean recall (MR). The experiments were conducted using the ChinaWare dataset, so all methods were trained on this dataset. A data table comparing each method’s performance on recall rate metrics was provided. In terms of the MR metric, the proposed method outperformed zero-shot CN-CLIPViT-B/16 by 3.2% in the text-to-image alignment task and by 7.5% in the image-to-text task. CoOp focuses on text features; it also outperforms CoOp by 11.4% and 12.1%, respectively. Moreover, CoCoOp considers image features on the basis of CoOp, and the proposed method outperforms CoCoOp by 8.4% and 9.5%, respectively. Pic2Word also focuses on original image features and does not fully utilize other local image features to improve model performance, and the proposed method outperforms Pic2Word by 5.8% and 5.6%, respectively.ConclusionThe cross-modal entity alignment method proposed in this study can fully explore the effective intermediate representation of image features to reconstruct text features without changing the parameters of the VLP model, thereby improving the cross-modal recognition accuracy of the details of celadon. The experimental results show that this method is superior to several state-of-the-art methods and has improved the performance of alignment. Ultimately, a celadon cross-modal knowledge graph with 8 949 nodes and 18 211 relationships was successfully constructed by applying technologies such as ontology modeling, data mining, and the cross-modal entity alignment method proposed in this study.… …   相似文献
《中国图象图形学报》2025,30(5):1318-1333
271.
随机块模型可以拟合各种网络的生成,挖掘网络的隐含结构与潜在联系,在社团检测中具有明显的优势.广义随机块模型GSB是基于链接社团的思想发现广义社团的,但其仅适用于有向无属性网络.针对无向属性网络,对网络拓扑信息建模的同时对节点属性进行建模,提出一种度修正的属性网络广义随机块模型DC… …   相似文献
王笑  戴芳  郭文艳  王军锋 《软件学报》2025,36(5):2308-2320
272.
  
针对半监督视频目标分割(VOS)领域中基于记忆的方法存在由于目标交互造成的物体遮挡以及背景中类似对象或噪声的干扰等问题,提出一种基于时空解耦和区域鲁棒性增强的半监督VOS方法。首先,构建一个结构化Transformer架构去除所有像素共有的特征信息,突出每个像素之间的差异,深入挖… …   相似文献
273.
  
为了利用环论优化算法(RTEA)高效求解多维背包问题(MKP),在分析已有修复优化算子——基于物品整体资源消耗伪效用比的修复优化算子RO1和基于物品各维度资源消耗价值密度的修复优化算子RO3不足的基础上,结合互补策略提出一种新的修复优化算子——加权修复优化算子RO4。随后,引入继… …   相似文献
274.
  
针对智能反射面(RIS)辅助通信系统中信道估计精度低的问题,提出一种基于信道去噪网络(CDN)的信道估计方案,将信道估计问题建模为信道噪声消除的问题。首先使用传统算法对接收到的导频信号进行初步预估计,随后将该预估计信号输入信道估计网络以学习噪声特征并进行去噪处理,从而恢复出精确的… …   相似文献
275.
  
图像文字描述技术可以帮助计算机更好地理解图像内容,实现跨模态交互。针对图像中文描述任务中存在的图像多粒度特征提取不全面以及图文关联性理解不充分等问题,提出一种基于多级视觉与图文动态交互的图像中文描述方法。首先,在编码器端提取多级视觉特征,通过图像局部特征提取器的辅助引导模块获取多… …   相似文献
276.
  
语音到语音翻译(S2ST)是智能语音领域中新兴的研究方向,旨在将一种语言的语音准确翻译成另一种语言的语音。随着人们对跨语言交流需求的增加,S2ST受到广泛的关注,相关研究也不断涌现。传统的级联模型在S2ST过程中存在诸多问题,如错误传播、推理延迟和无法翻译无文字系统的语言等,因此… …   相似文献
277.
  
利用传统模型对糖尿病肾病(DN)高风险患者的视网膜疾病进行早期诊断时,由于糖尿病患者的视网膜图像数据少且类别不平衡,诊断精度不高。因此,提出一种基于知识蒸馏双分支结构的视网膜病变辅助诊断方法,以提高对少数类别的识别能力。该方法首先使用在大型医学数据集上训练的教师网络指导学生网络学… …   相似文献
278.
  
多模态人脸识别技术能充分利用人脸特征或其他生物特征提高识别的鲁棒性和安全性,具有广泛的实际应用价值。由于目前的多模态人脸识别研究存在模态差距和模态信息难以高效融合等问题,因此根据多种信息模态和应用目的对现有的多模态人脸识别方法进行分类综述,以梳理研究中存在的问题,并探讨未来的发展… …   相似文献
279.
  
工业缺陷检测在保障产品质量、提高企业竞争力方面具有极其重要的作用。传统的缺陷检测方法依赖人工检查,成本高且效率低下,难以满足大规模的质量检验需求。近年来,基于视觉的工业缺陷检测技术取得了显著进步,已成为产品外观质量检验的一种高效解决方案。但在许多实际工业场景中,获取大量带有标签的… …   相似文献
280.
  
现有的基于双向长短时记忆(BiLSTM)网络的命名实体识别(NER)模型难以全面理解文本的整体语义以及捕捉复杂的实体关系。因此,提出一种基于全域信息融合和多维关系感知的NER模型。首先,通过BERT(Bidirectional Encoder Representations fr… …   相似文献
[首页] « 上一页 [9] [10] [11] [12] [13] 14 [15] [16] [17] [18] [19] 下一页 » 末  页»