Autonomous vehicles rely heavily on accurate and efficient scene segmentation for safe navigation and efficient operations.Traditional Bird’s Eye View(BEV)methods on semantic scene segmentation,which leverage multimo...Autonomous vehicles rely heavily on accurate and efficient scene segmentation for safe navigation and efficient operations.Traditional Bird’s Eye View(BEV)methods on semantic scene segmentation,which leverage multimodal sensor fusion,often struggle with noisy data and demand high-performance GPUs,leading to sensor misalignment and performance degradation.This paper introduces an Enhanced Channel Attention BEV(ECABEV),a novel approach designed to address the challenges under insufficient GPU memory conditions.ECABEV integrates camera and radar data through a de-noise enhanced channel attention mechanism,which utilizes global average and max pooling to effectively filter out noise while preserving discriminative features.Furthermore,an improved fusion approach is proposed to efficiently merge categorical data across modalities.To reduce computational overhead,a bilinear interpolation layer normalizationmethod is devised to ensure spatial feature fidelity.Moreover,a scalable crossentropy loss function is further designed to handle the imbalanced classes with less computational efficiency sacrifice.Extensive experiments on the nuScenes dataset demonstrate that ECABEV achieves state-of-the-art performance with an IoU of 39.961,using a lightweight ViT-B/14 backbone and lower resolution(224×224).Our approach highlights its cost-effectiveness and practical applicability,even on low-end devices.The code is publicly available at:https://github.com/YYF-CQU/ECABEV.git.展开更多
Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-onl...Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-only approaches.To address this issue,this paper proposes a framework named“a lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge”.This framework innovatively designs a lightweight vision-only student model based on Res Net,which leverages a dual distillation mechanism to learn from a powerful teacher model that integrates temporal information from both image and light detection and ranging(LiDAR)modalities.Specifically,we distill efficient multi-modal feature extraction and spatial fusion capabilities from the BEVFusion model,and distill advanced temporal information fusion and spatiotemporal attention mechanisms from the BEVFormer model.This dual distillation strategy enables the student model to achieve perception performance close to that of multi-modal models without relying on Li DAR.Experimental results on the nu Scenes dataset demonstrate that the proposed model significantly outperforms classical vision-only algorithms,achieves comparable performance to current state-of-the-art vision-only methods on the nu Scenes detection leaderboard in terms of both mean average precision(mAP)and the nu Scenes detection score(NDS)metrics,and exhibits notable advantages in inference computational efficiency.Although the proposed dual-teacher paradigm incurs higher offline training costs compared to single-model approaches,it yields a streamlined and highly efficient student model suitable for resource-constrained real-time deployment.This provides an effective pathway toward low-cost,high-performance autonomous driving perception systems.展开更多
Background: Circulating tumor cells (CTCs) are often undetected through the immunomagnetic epithelial cell adhesion molecule (EpCAM)-based CellSearch~ System in breast and colorectal cancer (CRC) patients treat...Background: Circulating tumor cells (CTCs) are often undetected through the immunomagnetic epithelial cell adhesion molecule (EpCAM)-based CellSearch~ System in breast and colorectal cancer (CRC) patients treated with bevacizumab (BEV), where low CTC numbers have been reported even in patients with evidence of progression of disease. To date, the reasons for this discrepancy have not been clarified. This study was carried out to investigate the molecular and phenotypic changes in CRC cells after chronic exposure to BEV in vitro. Methods: The human CRC cell line WiDr was exposed to a clinically relevant dose of BEV for 3 months in vitro. The expression of epithelial and mesenchymal markers and EpCAM isoforms was determined by western blotting and immunofluorescence. To evaluate the impact of EpCAM variant isoforms expression on CTC enumeration by CellSearch, untreated and treated colon cancer cells were spiked into 7.5 mL of blood from a healthy donor and enumerated by CellSearch. Results: Chronic exposure of CRC cell line to BEV induced decreased expression of EpCAM 40 kDa isoform and increased expression EpCAM 42 kDa isoform, together with a decreased expression of cytokeratins (CK), while no evidence of epithelial to mesenchymal transition (EMT) in treated cells was observed. The recovery rate of cells through CellSearch was gradually reduced in course of treatment with BEV, being 84% , 70% and 40% at l, 2 and 3 months, respectively. Conclusions: We hypothesize that BEV may prevent CellSearch from capturing CTCs through altering EpCAM isoforms.展开更多
基金funded by the National Natural Science Foundation of China,grant number 62262045the Fundamental Research Funds for the Central Universities,grant number 2023CDJYGRH-YB11the Open Funding of SUGON Industrial Control and Security Center,grant number CUIT-SICSC-2025-03.
文摘Autonomous vehicles rely heavily on accurate and efficient scene segmentation for safe navigation and efficient operations.Traditional Bird’s Eye View(BEV)methods on semantic scene segmentation,which leverage multimodal sensor fusion,often struggle with noisy data and demand high-performance GPUs,leading to sensor misalignment and performance degradation.This paper introduces an Enhanced Channel Attention BEV(ECABEV),a novel approach designed to address the challenges under insufficient GPU memory conditions.ECABEV integrates camera and radar data through a de-noise enhanced channel attention mechanism,which utilizes global average and max pooling to effectively filter out noise while preserving discriminative features.Furthermore,an improved fusion approach is proposed to efficiently merge categorical data across modalities.To reduce computational overhead,a bilinear interpolation layer normalizationmethod is devised to ensure spatial feature fidelity.Moreover,a scalable crossentropy loss function is further designed to handle the imbalanced classes with less computational efficiency sacrifice.Extensive experiments on the nuScenes dataset demonstrate that ECABEV achieves state-of-the-art performance with an IoU of 39.961,using a lightweight ViT-B/14 backbone and lower resolution(224×224).Our approach highlights its cost-effectiveness and practical applicability,even on low-end devices.The code is publicly available at:https://github.com/YYF-CQU/ECABEV.git.
基金supported by the National Natural Science Foundation of China(42476084,62203456,42276199)the Stable Support Project of National Key Laboratory(WDZC 20245250302)the National Key R&D Program of China(2024YFC2813502,2024YFC2813302)。
文摘Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-only approaches.To address this issue,this paper proposes a framework named“a lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge”.This framework innovatively designs a lightweight vision-only student model based on Res Net,which leverages a dual distillation mechanism to learn from a powerful teacher model that integrates temporal information from both image and light detection and ranging(LiDAR)modalities.Specifically,we distill efficient multi-modal feature extraction and spatial fusion capabilities from the BEVFusion model,and distill advanced temporal information fusion and spatiotemporal attention mechanisms from the BEVFormer model.This dual distillation strategy enables the student model to achieve perception performance close to that of multi-modal models without relying on Li DAR.Experimental results on the nu Scenes dataset demonstrate that the proposed model significantly outperforms classical vision-only algorithms,achieves comparable performance to current state-of-the-art vision-only methods on the nu Scenes detection leaderboard in terms of both mean average precision(mAP)and the nu Scenes detection score(NDS)metrics,and exhibits notable advantages in inference computational efficiency.Although the proposed dual-teacher paradigm incurs higher offline training costs compared to single-model approaches,it yields a streamlined and highly efficient student model suitable for resource-constrained real-time deployment.This provides an effective pathway toward low-cost,high-performance autonomous driving perception systems.
文摘Background: Circulating tumor cells (CTCs) are often undetected through the immunomagnetic epithelial cell adhesion molecule (EpCAM)-based CellSearch~ System in breast and colorectal cancer (CRC) patients treated with bevacizumab (BEV), where low CTC numbers have been reported even in patients with evidence of progression of disease. To date, the reasons for this discrepancy have not been clarified. This study was carried out to investigate the molecular and phenotypic changes in CRC cells after chronic exposure to BEV in vitro. Methods: The human CRC cell line WiDr was exposed to a clinically relevant dose of BEV for 3 months in vitro. The expression of epithelial and mesenchymal markers and EpCAM isoforms was determined by western blotting and immunofluorescence. To evaluate the impact of EpCAM variant isoforms expression on CTC enumeration by CellSearch, untreated and treated colon cancer cells were spiked into 7.5 mL of blood from a healthy donor and enumerated by CellSearch. Results: Chronic exposure of CRC cell line to BEV induced decreased expression of EpCAM 40 kDa isoform and increased expression EpCAM 42 kDa isoform, together with a decreased expression of cytokeratins (CK), while no evidence of epithelial to mesenchymal transition (EMT) in treated cells was observed. The recovery rate of cells through CellSearch was gradually reduced in course of treatment with BEV, being 84% , 70% and 40% at l, 2 and 3 months, respectively. Conclusions: We hypothesize that BEV may prevent CellSearch from capturing CTCs through altering EpCAM isoforms.