-
Multi-objective math problem generation using large language model through an adaptive multi-level retrieval augmentation framework Inform. Fusion (IF 14.7) Pub Date : 2025-02-25 Jianwen Sun, Wangzi Shi, Xiaoxuan Shen, Shengyingjie Liu, Luona Wei, Qian Wan
Math problems are an important knowledge carrier and evaluation means in personalized teaching. Their high cost of manual compilation promotes the research of math problem generation. Many previous studies have focused on the generation of math word problems, which are difficult to meet the real teaching needs due to the single task-objective orientation and small differences in generation results
-
Enhancing trust in automated 3D point cloud data interpretation through explainable counterfactuals Inform. Fusion (IF 14.7) Pub Date : 2025-02-25 Andreas Holzinger, Niko Lukač, Dzemail Rozajac, Emile Johnston, Veljka Kocic, Bernhard Hoerl, Christoph Gollob, Arne Nothdurft, Karl Stampfer, Stefan Schweng, Javier Del Ser
This paper introduces a novel framework for augmenting explainability in the interpretation of point cloud data by fusing expert knowledge with counterfactual reasoning. Given the complexity and voluminous nature of point cloud datasets, derived predominantly from LiDAR and 3D scanning technologies, achieving interpretability remains a significant challenge, particularly in smart cities, smart agriculture
-
Adaptive structural-guided multi-level representation learning with graph contrastive for incomplete multi-view clustering Inform. Fusion (IF 14.7) Pub Date : 2025-02-24 Haiyue Wang, Wensheng Zhang, Quan Wang, Xiaoke Ma
Incomplete multi-view clustering (IMC) is a pivotal task within the area of machine learning, encompassing several unresolved challenges, such as representation of objects, relations of various views, discriminative of features, and data restoration. To address these challenges, we propose a novel Adaptive Structural-guided Multi-level representation Learning with Graph Contrastive algorithm for IMC
-
From screens to scenes: A survey of embodied AI in healthcare Inform. Fusion (IF 14.7) Pub Date : 2025-02-21 Yihao Liu, Xu Cao, Tingting Chen, Yankai Jiang, Junjie You, Minghua Wu, Xiaosong Wang, Mengling Feng, Yaochu Jin, Jintai Chen
Healthcare systems worldwide face persistent challenges in efficiency, accessibility, and personalization. Modern artificial intelligence (AI) has shown promise in addressing these issues through precise predictive modeling; however, its impact remains constrained by limited integration into clinical workflows. Powered by modern AI technologies such as multimodal large language models and world models
-
Deep multi-view clustering: A comprehensive survey of the contemporary techniques Inform. Fusion (IF 14.7) Pub Date : 2025-02-20 Anal Roy Chowdhury, Avisek Gupta, Swagatam Das
Data can be represented by multiple sets of features, where each semantically coherent set of features is called a view. For example, an image can be represented by multiple sets of features that measure textures, shapes, edge features, etc. Collecting multiple views of data is generally easier than annotating it with the help of experts. Thus, the unsupervised exploration of data in consultation with
-
A review of medical text analysis: Theory and practice Inform. Fusion (IF 14.7) Pub Date : 2025-02-19 Yani Chen, Chunwu Zhang, Ruibin Bai, Tengfang Sun, Weiping Ding, Ruili Wang
Medical data analysis has emerged as an important driving force for smart healthcare with applications ranging from disease analysis to triage, diagnosis, and treatment. Text data plays a crucial role in providing contexts and details that other data types cannot capture alone, making its analysis an indispensable resource in medical research. Natural language processing, a key technology for analyzing
-
SSEFusion: Salient semantic enhancement for multimodal medical image fusion with Mamba and dynamic spiking neural networks Inform. Fusion (IF 14.7) Pub Date : 2025-02-19 Shiqiang Liu, Weisheng Li, Dan He, Guofen Wang, Yuping Huang
Multimodal medical image fusion technology enhances medical representations and plays a vital role in clinical diagnosis. However, fusing medical images remains a challenge due to the stochastic nature of lesions and the complex structures of organs. Although many fusion methods have been proposed recently, most struggle to effectively establish global context dependency while preserving salient semantic
-
Enhancing cross-domain generalization by fusing language-guided feature remapping Inform. Fusion (IF 14.7) Pub Date : 2025-02-19 Ziteng Qiao, Dianxi Shi, Songchang Jin, Yanyan Shi, Luoxi Jing, Chunping Qiu
Domain generalization refers to training a model with annotated source domain data and making it generalize to various unseen target domains. It has been extensively studied in classification, but remains challenging in object detection. Existing domain generalization object detection methods mainly rely on generative or adversarial data augmentation, which increases the complexity of training. Recently
-
MSF-Net: Multi-stage fusion network for emotion recognition from multimodal signals in scalable healthcare Inform. Fusion (IF 14.7) Pub Date : 2025-02-19 Md. Milon Islam, Fakhri Karray, Ghulam Muhammad
Automatic emotion recognition has attracted significant interest in healthcare, thanks to remarkable developments made recently in smart and innovative technologies. A real-time emotion recognition system allows for continuous monitoring, comprehension, and enhancement of the physical entity’s capacities, along with continuing advice for enhancing quality of life and well-being in the context of personalized
-
MapFusion: A novel BEV feature fusion network for multi-modal map construction Inform. Fusion (IF 14.7) Pub Date : 2025-02-18 Xiaoshuai Hao, Yunfeng Diao, Mengchuan Wei, Yifan Yang, Peng Hao, Rong Yin, Hui Zhang, Weiming Li, Shu Zhao, Yu Liu
Map construction task plays a vital role in providing precise and comprehensive static environmental information essential for autonomous driving systems. Primary sensors include cameras and LiDAR, with configurations varying between camera-only, LiDAR-only, or camera-LiDAR fusion, based on cost-performance considerations. While fusion-based methods typically perform best, existing approaches often
-
From patches to WSIs: A systematic review of deep Multiple Instance Learning in computational pathology Inform. Fusion (IF 14.7) Pub Date : 2025-02-18 Yuchen Zhang, Zeyu Gao, Kai He, Chen Li, Rui Mao
Clinical decision support systems for pathology, particularly those utilizing computational pathology (CPATH) for whole slide image (WSI) analysis, face significant challenges due to the need for high-quality annotated datasets. Given the vast amount of information contained in WSIs, creating such datasets is often prohibitively expensive and time-consuming. Multiple Instance Learning (MIL) has emerged
-
DGFD: A dual-graph convolutional network for image fusion and low-light object detection Inform. Fusion (IF 14.7) Pub Date : 2025-02-18 Xiaoxuan Chen, Shuwen Xu, Shaohai Hu, Xiaole Ma
Traditional convolutional operations primarily concentrate on local feature extraction, which can result in the loss of global features. However, current fusion methods for extracting global features exhibit high time complexity and have difficulties in capturing long-range dependencies. In this paper, a dual-graph convolutional neural network is constructed to perform cross-modal graph inference based
-
A comprehensive survey of visible and infrared imaging in complex environments: Principle, degradation and enhancement Inform. Fusion (IF 14.7) Pub Date : 2025-02-17 Yuanbo Li, Ping Zhou, Gongbo Zhou, Haozhe Wang, Yunqi Lu, Yuxing Peng
Images captured in extreme environments, including deep-earth, deep-sea, and deep-space exploration sites, often suffer from significant degradation due to complex visual factors, which adversely impact visual quality and complicate perceptual tasks. This survey systematically synthesizes recent advancements in visual perception and understanding within these challenging contexts. It focuses on the
-
Weighted-digraph-guided multi-kernelized learning for outlier explanation Inform. Fusion (IF 14.7) Pub Date : 2025-02-17 Lili Guan, Lei Duan, Xinye Wang, Haiying Wang, Rui Lin
Outlier explanation methods based on outlying subspace mining have been widely used in various applications due to their effectiveness and explainability. These existing methods aim to find an outlying subspace of the original space (a set of features) that can clearly distinguish a query outlier from all inliers. However, when the query outlier in the original space are linearly inseparable from inliers
-
STA-Net: Spatial–temporal alignment network for hybrid EEG-fNIRS decoding Inform. Fusion (IF 14.7) Pub Date : 2025-02-15 Mutian Liu, Banghua Yang, Lin Meng, Yonghuai Zhang, Shouwei Gao, Peng Zan, Xinxing Xia
Hybrid brain–computer interfaces (BCI) have garnered attention for the capacity to transcend the constraints of single-modality BCI. It is essential to develop innovative fusion methodologies to exploit the high temporal resolution of electroencephalography (EEG) and the high spatial resolution of functional near-infrared spectroscopy (fNIRS). We propose an end-to-end Spatial–Temporal Alignment Network
-
NeuralOOD: Improving out-of-distribution generalization performance with brain-machine fusion learning framework Inform. Fusion (IF 14.7) Pub Date : 2025-02-14 Shuangchen Zhao, Changde Du, Jingze Li, Hui Li, Huiguang He
Deep Neural Networks (DNNs) have demonstrated exceptional recognition capabilities in traditional computer vision (CV) tasks. However, existing CV models often suffer a significant decrease in accuracy when confronted with out-of-distribution (OOD) data. In contrast to these DNN models, human can maintain a consistently low error rate when facing OOD scenes, partly attributed to the rich prior cognitive
-
Mutual-support generalized category discovery Inform. Fusion (IF 14.7) Pub Date : 2025-02-14 Yu Duan, Zhanxuan Hu, Rong Wang, Zhensheng Sun, Feiping Nie, Xuelong Li
This work focuses on the problem of Generalized Category Discovery (GCD), a more realistic and challenging semi-supervised learning setting where unlabeled data may belong to either previously known or unseen categories. Recent advancements have demonstrated the efficacy of both pseudo-label-based parametric classification methods and representation-based non-parametric classification methods in tackling
-
CCSUMSP: A cross-subject Chinese speech decoding framework with unified topology and multi-modal semantic pre-training Inform. Fusion (IF 14.7) Pub Date : 2025-02-14 Shuai Huang, Yongxiong Wang, Huan Luo
Decoding speech from brain signals has been a long-standing challenge in neuroscience and brain–computer interface research. While significant progress has been made in English speech decoding, cross-subject Chinese speech decoding remains understudied, despite its potential applications and unique linguistic characteristics. Chinese, with its logographic writing system and tonal nature, presents unique
-
Multi-scale convolutional attention frequency-enhanced transformer network for medical image segmentation Inform. Fusion (IF 14.7) Pub Date : 2025-02-12 Shun Yan, Benquan Yang, Aihua Chen, Xiaoming Zhao, Shiqing Zhang
Automatic segmentation of medical images plays a crucial role in assisting doctors with diagnosis and treatment planning. Among them, multi-scale vision transformer has become a powerful tool for medical image segmentation. However, due to its overly aggressive self-attention design leads to issues such as insufficient local feature extraction and lack of detailed feature information. To address these
-
A multiview-slice feature fusion network for early diagnosis of Alzheimer’s disease with structural MRI images Inform. Fusion (IF 14.7) Pub Date : 2025-02-12 Hesheng Huang, Witold Pedrycz, Kaoru Hirota, Fei Yan
Alzheimer’s disease (AD) is a prevalent neurodegenerative disorder with high incidence and significant mortality among the elderly worldwide. Nevertheless, early and accurate diagnosis and treatment of the disease could delay its progression to evolve into more severe phases. Traditional methods, which are largely binary in classification, often struggle with the complexity of multi-classification
-
Learning to estimate probabilistic attitude via matrix Fisher distribution with inertial sensor Inform. Fusion (IF 14.7) Pub Date : 2025-02-12 Yuqiang Jin, Wen-An Zhang, Ling Shi
We propose a probabilistic framework for attitude estimation using the matrix Fisher distribution on the special orthogonal group SO(3). To be specific, a deep neural network is first designed to estimate the attitude probability distributions over SO(3), i.e., the unconstrained parameters of matrix Fisher distribution. The network takes common inertial measurements as input and overcomes the challenge
-
Anomaly detection for microservice system via augmented multimodal data and hybrid graph representations Inform. Fusion (IF 14.7) Pub Date : 2025-02-11 Peipeng Wang, Xiuguo Zhang, Zhiying Cao
Accurate anomaly detection is essential for ensuring the reliability of microservice systems. Current approaches typically analyze system anomaly patterns using single-modal data (i.e., traces, metrics, and logs) while neglecting the class imbalance between normal and abnormal samples, which can easily lead to misjudgment. This paper propose AMulSys, a graph-based anomaly detection approach, which
-
Eficient image denoising using deep learning: A brief survey Inform. Fusion (IF 14.7) Pub Date : 2025-02-11 Bo Jiang, Jinxing Li, Yao Lu, Qing Cai, Huaibo Song, Guangming Lu
Image denoising is a vital computer vision task that aims to remove noise from images. Deep learning techniques have made remarkable progress in this field in recent years. This survey provides a comprehensive overview of efficient deep learning-based image denoising methods. Unlike previous reviews, it focuses exclusively on models based on efficient deep learning and examines the latest developments
-
Learning and grounding visual multimodal adaptive graph for visual navigation Inform. Fusion (IF 14.7) Pub Date : 2025-02-11 Kang Zhou, Jianping Wang, Weitao Xu, Linqi Song, Zaiqiao Ye, Chi Guo, Cong Li
Visual navigation requires the agent reasonably perceives the environment and effectively navigates to the given target. In this task, we present a Multimodal Adaptive Graph (MAG) for learning and grounding the visual clues based on the object relationships. MAG consists of key navigation elements: object relative position relationships, previous navigation actions, past training experience, and target
-
DHHNN: A Dynamic Hypergraph Hyperbolic Neural Network based on variational autoencoder for multimodal data integration and node classification Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Zhangyu Mei, Xiao Bi, Dianguo Li, Wen Xia, Fan Yang, Hao Wu
In recent years, the integration of hyperbolic geometry with Graph Neural Networks (GNNs) has garnered significant attention due to its effectiveness in capturing complex hierarchical structures, particularly within real-world graphs and scale-free networks. Although hyperbolic neural networks have shown strong performance across various domains, most existing models rely on static graph structures
-
MulGad: Multi-granularity contrastive learning for multivariate time series anomaly detection Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Bo-Wen Xiao, Hong-Jie Xing, Chun-Guo Li
Since the normal patterns of time series change dynamically over time, unsupervised time series anomaly detection methods have to face the overfitting problem. Although some approaches based on contrastive learning try to solve this problem, these methods ignore the complex intrinsic correlations within the given time series and may face the representation collapse problem. As the result, the normal
-
A review on full-, zero-, and partial-knowledge based predictive models for industrial applications Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Stefano Zampini, Guido Parodi, Luca Oneto, Andrea Coraddu, Davide Anguita
In contemporary industrial applications, predictive models have been pivotal in bolstering production efficiency, product quality, scalability, and cost-effectiveness while promoting sustainability. These predictive models can be constructed solely based on domain-specific knowledge, exclusively on observational data, or by amalgamating both approaches. They are commonly referred to as Full-, Zero-
-
SaliencyI2PLoc: Saliency-guided image–point cloud localization using contrastive learning Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Yuhao Li, Jianping Li, Zhen Dong, Yuan Wang, Bisheng Yang
Image to point cloud global localization is crucial for robot navigation in GNSS-denied environments and has become increasingly important for multi-robot map fusion and urban asset management. The modality gap between images and point clouds poses significant challenges for cross-modality fusion. Current cross-modality global localization solutions either require modality unification, which leads
-
Deep learning based infrared small object segmentation: Challenges and future directions Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Zhengeng Yang, Hongshan Yu, Jianjun Zhang, Qiang Tang, Ajmal Mian
Infrared sensing is a core method for supporting unmanned systems, such as autonomous vehicles and drones. Recently, infrared sensors have been widely deployed on mobile and stationary platforms for detection and classification of objects from long distances and in wide field of views. Given its success in the vision image analysis domain, deep learning has also been applied for object recognition
-
MammoVLM: A generative large vision–language model for mammography-related diagnostic assistance Inform. Fusion (IF 14.7) Pub Date : 2025-02-10 Zhenjie Cao, Zhuo Deng, Jie Ma, Jintao Hu, Lan Ma
Inspired by the recent success of large language models (LLMs) in the general domain, many large multimodal models, such as vision–language models, have been developed to tackle problems across modalities.
-
DeepFusionSent: A novel feature fusion approach for deep learning-enhanced sentiment classification Inform. Fusion (IF 14.7) Pub Date : 2025-02-08 Ankit Thakkar, Devshri Pandya
Sentiment Analysis (SA) is the process to identify, extract, and quantify subjective information from text data, such as opinions, attitudes, emotions, and sentiments. Traditional approaches to sentiment classification primarily relied on utilizing raw data, which may not fully capture the detail insights within the dataset. Here, we propose an approach that seeks to enhance sentiment classification
-
Multimodal fusion of spatial–temporal and frequency representations for enhanced ECG classification Inform. Fusion (IF 14.7) Pub Date : 2025-02-08 Chien-Liang Liu, Bin Xiao, Chiu-Hsun Hsieh
Electrocardiogram (ECG) classification is pivotal in diagnosing and monitoring cardiovascular diseases (CVDs). However, existing methods predominantly rely on hand-crafted features or specific signal representations, often leading to incomplete and suboptimal analysis of ECG data. This paper addresses these limitations by identifying a significant research gap in the comprehensive utilization of ECG
-
Multi-source data driven decision-support model for product ranking with consumer psychology behavior Inform. Fusion (IF 14.7) Pub Date : 2025-02-07 Peng Wu, Yutong Xie, Ligang Zhou, Muhammet Deveci, Luis Martínez
In the context of information overload, leveraging online reviews for product ranking can enhance consumers’ decision-making efficiency and facilitate informed purchasing decisions. However, previous studies have mainly focused on online review data from a single source and do not capture consumer psychology behavior for scientific and reasonable product ranking. Therefore, this study proposes a multi-source
-
InsGNN: Interpretable spatio-temporal graph neural networks via information bottleneck Inform. Fusion (IF 14.7) Pub Date : 2025-02-06 Hui Fang, Haishuai Wang, Yang Gao, Yonggang Zhang, Jiajun Bu, Bo Han, Hui Lin
Spatio-temporal graph neural networks (STGNNs) have garnered considerable attention for their promising performance across various applications. While existing models have demonstrated superior performance in exploring the interpretability of graph neural networks (GNNs), the interpretability of STGNNs is constrained by their complex spatio-temporal correlations. In this paper, we introduce a novel
-
Vision-Language Models in medical image analysis: From simple fusion to general large models Inform. Fusion (IF 14.7) Pub Date : 2025-02-06 Xiang Li, Like Li, Yuchen Jiang, Hao Wang, Xinyu Qiao, Ting Feng, Hao Luo, Yong Zhao
Vision-Language Model (VLM) is a kind of multi-modality deep learning model that aims to fuse visual information with language information to enhance the understanding and analysis of visual content. VLM was originally used to integrate multi-modality information and improve task accuracy. Then, VLM was further developed in combination with zero-shot and few-shot learning to solve the problem of insufficient
-
Next-generation coupled structure-human sensing technology: Enhanced pedestrian-bridge interaction analysis using data fusion and machine learning Inform. Fusion (IF 14.7) Pub Date : 2025-02-05 Sahar Hassani, Samir Mustapha, Jianchun Li, Mohsen Mousavi, Ulrike Dackermann
The consequences of crowd behavior in high-density pedestrian flows, especially in response to exacerbating incidents, can result in tragic outcomes such as trampling and crushing, making the active monitoring of crowd motion crucial, to provide timely danger warnings and implement preventive measures. This paper proposes a novel approach for crowd behavior monitoring and prediction of bridge loads
-
Advances in DeepFake detection algorithms: Exploring fusion techniques in single and multi-modal approach Inform. Fusion (IF 14.7) Pub Date : 2025-02-05 Ashish Kumar, Divya Singh, Rachna Jain, Deepak Kumar Jain, Chenquan Gan, Xudong Zhao
In recent years, generative artificial intelligence has gained momentum and created extremely realistic synthetic multimedia content that can spread misinformation and mislead society. Deepfake detection is a technique consisting of frameworks, algorithms and approaches to predict manipulated contents namely, image, audio and video. To this end, we have analyzed and explored various deepfake detection
-
Modality-perceptive harmonization network for visible-infrared person re-identification Inform. Fusion (IF 14.7) Pub Date : 2025-02-05 Xutao Zuo, Jinjia Peng, Tianhang Cheng, Huibing Wang
Visible-infrared person re-identification (VI-ReID) remains a challenging task due to the inconsistencies in data distribution and semantic inconsistency between heterogeneous modalities. Some visible-infrared person re-identification methods that leverage auxiliary modalities have achieved significant progress. However, these methods merely apply pixel-level augmentation to the original images and
-
RK-VQA: Rational knowledge-aware fusion-in-decoder for knowledge-based visual question answering Inform. Fusion (IF 14.7) Pub Date : 2025-02-05 Weipeng Chen, Xu Huang, Zifeng Liu, Jin Liu, Lan Yo
Knowledge-based Visual Question Answering (KB-VQA) expands traditional VQA by utilizing world knowledge from external sources when the image alone is insufficient to infer a correct answer. Existing methods face challenges due to low recall rates, limiting the ability to gather essential information for accurate answers. While increasing the amount of retrieved knowledge entries can enhance recall
-
Distilling implicit multimodal knowledge into large language models for zero-resource dialogue generation Inform. Fusion (IF 14.7) Pub Date : 2025-02-04 Bo Zhang, Hui Ma, Jian Ding, Jian Wang, Bo Xu, Hongfei Lin
Integrating multimodal knowledge into large language models (LLMs) represents a significant advancement in dialogue generation capabilities. However, the effective incorporation of such knowledge in zero-resource scenarios remains a substantial challenge due to the scarcity of diverse, high-quality dialogue datasets. To address this, we propose the Visual Implicit Knowledge Distillation Framework (VIKDF)
-
Wi-Fi fine time measurement–Principles, applications, and future trends: A survey Inform. Fusion (IF 14.7) Pub Date : 2025-02-03 Yuan Wu, Mengyi He, Wei Li, Izzy Yi Jian, Yue Yu, Liang Chen, Ruizhi Chen
IEEE 802.11–2016 proposed the Wi-Fi Fine Time Measurement (FTM) protocol, aiming at providing meter or sub-meter level ranging function between smart terminals and Wi-Fi access points (APs). Compared with other indoor positioning technologies for instance, Bluetooth, acoustic, visible light, Ultra-wideband, etc., Wi-Fi has been characterized by low cost, no deployment, and potentially high positioning
-
CSLP: A novel pansharpening method based on compressed sensing and L-PNN Inform. Fusion (IF 14.7) Pub Date : 2025-02-03 Yingxia Chen, Zhenglong Wan, Zeqiang Chen, Mingming Wei
To address spectral distortion and the loss of spatial detail information caused by full-resolution pansharpening, this study proposes an unsupervised method combining compressed sensing (CS) and a deeper attention-based network architecture (L-PNN), namely CSLP. First, in the compressed sensing module, we apply sparse theory for image compression and reconstruction to reduce detail loss and enhance
-
A homogeneous multimodality sentence representation for relation extraction Inform. Fusion (IF 14.7) Pub Date : 2025-02-03 Kai Wang, Yanping Chen, WeiZhe Yang, Yongbin Qin
Deep neural networks enable a sentence to be transformed into different multimodalities such as a token sequence representation (a one-dimensional semantic representation) or a semantic plane (a two-dimensional semantic representation). Sequence representation has the advantage of learning sequential dependencies of a sentence. Semantic plane is built by organizing all spans of a sentence, which is
-
EHR-based prediction modelling meets multimodal deep learning: A systematic review of structured and textual data fusion methods Inform. Fusion (IF 14.7) Pub Date : 2025-02-01 Ariel Soares Teles, Ivan Rodrigues de Moura, Francisco Silva, Angus Roberts, Daniel Stahl
Electronic Health Records (EHRs) have transformed healthcare by digitally consolidating patient medical history, encompassing structured data (e.g., demographic data, lab results), and unstructured textual data (e.g., clinical notes). These data hold significant potential for predictive modelling, and recent studies have dedicated efforts to leverage the different modalities in a cohesive and effective
-
A maximum satisfaction-based feedback mechanism for non-cooperative behavior management with agreeableness personality traits detection in group decision making Inform. Fusion (IF 14.7) Pub Date : 2025-02-01 Yujia Liu, Yuwei Song, Jian Wu, Changyong Liang, Francisco Javier Cabrerizo
Non-cooperative behaviors will lead to consensus failure in group decision making problems. As a result, managing non-cooperative behavior is a significant challenge in group consensus reaching processes, which involves two main research questions:(1) How to define non-cooperative behavior? (2) How to design an appropriate model to manage non-cooperative behavior? Existing studies often overlook the
-
A comprehensive taxonomy of machine consciousness Inform. Fusion (IF 14.7) Pub Date : 2025-01-31 Ruilin Qin, Changle Zhou, Mengjie He
Machine consciousness (MC) is the ultimate challenge to artificial intelligence. Although great progress has been made in artificial intelligence and robotics, consciousness is still an enigma and machines are far from having it. To clarify the concepts of consciousness and the research directions of machine consciousness, in this review, a comprehensive taxonomy for machine consciousness is proposed
-
Deep spatial–spectral fusion transformer for remote sensing pansharpening Inform. Fusion (IF 14.7) Pub Date : 2025-01-31 Mengting Ma, Yizhen Jiang, Mengjiao Zhao, Xiaowen Ma, Wei Zhang, Siyang Song
Pansharpening is the task which reconstructs spatial–spectral properties during the fusion of high-resolution panchromatic (PAN) with low-resolution multi-spectral (LR-MS) images, to generate a high-resolution multi-spectral (HR-MS) image. Recent approaches typically model spatial and spectral properties and fuse them using end-to-end deep learning networks, which fail to take their crucial task-related
-
Explainable artificial intelligence and advanced feature selection methods for predicting gas concentration in longwall mining Inform. Fusion (IF 14.7) Pub Date : 2025-01-31 Haoqian Chang, Xiangqian Wang, Alexandra I. Cristea, Xiangrui Meng, Zuxiang Hu, Ziqi Pan
Accurate prediction of gas concentrations at longwall mining faces is critical for safety production, yet current methods still face challenges in interpretability and reliability. This study aims to enhance prediction accuracy and model interpretability by employing advanced feature selection techniques. We integrate Shapley Additive Explanations (SHAP) into feature selection process to identify and
-
Hybrid spatial–temporal graph neural network for traffic forecasting Inform. Fusion (IF 14.7) Pub Date : 2025-01-30 Peng Wang, Longxi Feng, Yijie Zhu, Haopeng Wu
Accurate traffic forecasting is the foundation of the intelligent transportation system (ITS). Among existing methods, researchers utilize deep neural networks to capture spatial–temporal correlations. While state-of-the-art methods have shown remarkable progress, they still have some limitations: (i) Most existing models struggle to make accurate prediction performance over longer time horizons due
-
Application of transfer learning for biomedical signals: A comprehensive review of the last decade (2014–2024) Inform. Fusion (IF 14.7) Pub Date : 2025-01-30 Mahboobeh Jafari, Xiaohui Tao, Prabal Barua, Ru-San Tan, U.Rajendra Acharya
Precise and timely disease diagnosis is essential for making effective treatment decisions and halting disease progression. Biomedical signals offer the potential for non-invasive diagnosis of diverse conditions, enhancing the ability to predict clinical outcomes and plan treatments more effectively. These signals have garnered significant attention, particularly in +conjunction with artificial intelligence
-
Sparse multi-label feature selection via pseudo-label learning and dynamic graph constraints Inform. Fusion (IF 14.7) Pub Date : 2025-01-30 Yao Zhang, Jun Tang, Ziqiang Cao, Han Chen
In multi-label feature selection (MLFS), pseudo-label learning techniques are often employed to mitigate the issue that the binary nature of ground-truth labels is incompatible with linear mappings. Existing studies have demonstrated that manifold learning and distance discrepancy are powerful facilities to well constrain pseudo-labels. However, there are still two weaknesses for the existing manifold
-
Integrating personalized and contextual information in fine-grained emotion recognition in text: A multi-source fusion approach with explainability Inform. Fusion (IF 14.7) Pub Date : 2025-01-30 Anh Ngo, Jan Kocoń
Emotion recognition in textual data is a rapidly evolving field with diverse applications. While the state-of-the-art (SOTA) models based on pre-trained large language models (LLMs) have demonstrated significant achievements, the existing approaches often overlook fine-grained emotional nuances within individual sentences and the influence of contextual information. Additionally, despite the growing
-
LGT: Luminance-guided transformer-based multi-feature fusion network for underwater image enhancement Inform. Fusion (IF 14.7) Pub Date : 2025-01-29 Jiashuo Shang, Ying Li, Hu Xing, Jingyi Yuan
When light propagates through water, absorption and scattering effects lead to uneven brightness distribution and significant color deviation in underwater images. Existing methods struggle to compensate for light attenuation at each wavelength while simultaneously balancing luminance variations across different regions. To address these challenges, we propose the Luminance-Guided Transformer (LGT)
-
Multi-agent reinforcement learning with weak ties Inform. Fusion (IF 14.7) Pub Date : 2025-01-29 Huan Wang, Xu Zhou, Yu Kang, Jian Xue, Chenguang Yang, Xiaofeng Liu
Existing multi-agent reinforcement learning (MARL) algorithms focus primarily on maximizing global game gains or encouraging cooperation between agents, often overlooking the weak ties between them. In multi-agent environments, the quality of the information exchanged is crucial for optimal policy learning. To this end, we propose a novel MARL framework that integrates weak-tie theory with graph modeling
-
Counterfactual explanations for remaining useful life estimation within a Bayesian framework Inform. Fusion (IF 14.7) Pub Date : 2025-01-28 Jilles Andringa, Marcia L. Baptista, Bruno F. Santos
Machine learning has contributed to the advancement of maintenance in many industries, including aviation. In recent years, many neural network models have been proposed to address the problems of failure identification and estimating the remaining useful life (RUL). Nevertheless, the black-box nature of neural networks often limits their transparency and interpretability. Interpretability (or explainability)
-
Multimodal Document Analytics for Banking Process Automation Inform. Fusion (IF 14.7) Pub Date : 2025-01-27 Christopher Gerling, Stefan Lessmann
Traditional banks are increasingly challenged by FinTechs, particularly in leveraging advanced technologies to enhance operational efficiency. Our study addresses this by focusing on improving the efficiency of document-intensive business processes in banking. We review the landscape of business documents in the customer banking segment, which often includes text, layout, and visuals, indicating that
-
SenCounter: Towards category-agnostic action counting in open sensor sequences Inform. Fusion (IF 14.7) Pub Date : 2025-01-27 Shuangshuang Cao, Yanwen Wu, Yin Tang, Di Ge, Yanmei Ma, Cong Xiao
Repetition counting of multiple actions in sensor-based data is a critical task for human-centric applications like health monitoring and exercise training. Existing sensor-based repetition counting (SRC) methods are limited to single-action scenarios and predefined categories, which do not scale well in real-world scenarios. To address this, we introduce the Open Sensor Sequences Counting (OSSC) task
-
Unfolding coupled convolutional sparse representation for multi-focus image fusion Inform. Fusion (IF 14.7) Pub Date : 2025-01-25 Kecheng Zheng, Juan Cheng, Yu Liu
Multi-focus image fusion (MFIF) aims to generate an all-in-focus image from multiple partially focused images of the same scene captured with different focal settings. In this paper, we present a coupled convolutional sparse representation (CCSR) model for MFIF. Instead of being solved by an iterative thresholding algorithm, the proposed CCSR model is unfolded into a learnable neural network (termed
-
Explainable multi-frequency and multi-region fusion model for affective brain-computer interfaces Inform. Fusion (IF 14.7) Pub Date : 2025-01-24 Tao Wang, Rui Mao, Shuang Liu, Erik Cambria, Dong Ming
An affective brain-computer interface (aBCI) has demonstrated great potential in the field of emotion recognition. However, existing aBCI models encounter significant challenges in explainability and the effective fusion of multi-frequency and multi-region features, which greatly limits their practical applicability. To address these issues, this paper proposes an explainable multi-frequency and multi-region
-
Hallucinations of large multimodal models: Problem and countermeasures Inform. Fusion (IF 14.7) Pub Date : 2025-01-23 Shiliang Sun, Zhilin Lin, Xuhan Wu
The integration of multimodal capabilities into large models has unlocked unprecedented potential for tasks that involve understanding and generating diverse data modalities, including text, images, and audio. However, despite these advancements, such systems often suffer from hallucinations, that is, inaccurate, irrelevant, or entirely fabricated contents, which raise significant concerns about their