arXiv Analytics

Sign in

arXiv:2505.05763 [cs.LG]AbstractReferencesReviewsResources

BMMDetect: A Multimodal Deep Learning Framework for Comprehensive Biomedical Misconduct Detection

Yize Zhou, Jie Zhang, Meijie Wang, Lun Yu

Published 2025-05-09Version 1

Academic misconduct detection in biomedical research remains challenging due to algorithmic narrowness in existing methods and fragmented analytical pipelines. We present BMMDetect, a multimodal deep learning framework that integrates journal metadata (SJR, institutional data), semantic embeddings (PubMedBERT), and GPT-4o-mined textual attributes (methodological statistics, data anomalies) for holistic manuscript evaluation. Key innovations include: (1) multimodal fusion of domain-specific features to reduce detection bias; (2) quantitative evaluation of feature importance, identifying journal authority metrics (e.g., SJR-index) and textual anomalies (e.g., statistical outliers) as dominant predictors; and (3) the BioMCD dataset, a large-scale benchmark with 13,160 retracted articles and 53,411 controls. BMMDetect achieves 74.33% AUC, outperforming single-modality baselines by 8.6%, and demonstrates transferability across biomedical subfields. This work advances scalable, interpretable tools for safeguarding research integrity.

Related articles: Most relevant | Search more
arXiv:1803.02099 [cs.LG] (Published 2018-03-06)
A Hybrid Method for Traffic Flow Forecasting Using Multimodal Deep Learning
arXiv:2202.06218 [cs.LG] (Published 2022-02-13)
Emotion Based Hate Speech Detection using Multimodal Learning
arXiv:2302.01676 [cs.LG] (Published 2023-02-03)
Show me your NFT and I tell you how it will perform: Multimodal representation learning for NFT selling price prediction