In radar automatic target recognition(RATR),the high-resolution range profile(HRRP)has garnered considerable attention owing to its minimal computational demands.However,radar HRRP target recognition still faces numer...In radar automatic target recognition(RATR),the high-resolution range profile(HRRP)has garnered considerable attention owing to its minimal computational demands.However,radar HRRP target recognition still faces numerous challenges,primarily due to substantial variations in the amplitude and distribution of HRRP scattering points because of slight azimuthal changes.To alleviate the effect of aspect sensitivity,a novel multi-frame attention network(MFA-Net)comprising a range deformable convolution module(RDCM),multi-frame attention module(MFAM),and global-local Transformer module(GLTM)is proposed.The RDCM is designed to adaptively learn the distance of scattering center migration.Subsequently,the MFAM extracts consistent features across different frames to alleviate the influence of power fluctuation.Finally,the GLTM allocates attention between global and local fea-tures.The feasibility and effectiveness of the proposed method are validated through simulation and experimental datasets,and the recognition rate is enhanced by more than 3%compared to the state-of-the-art methods.展开更多
In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In additi...In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In addition,methods for deploying multiple cameras for motion capture of users(e.g.,performers)are widely used in computer graphics.As the need to minimize and optimize the number of cameras grows to reduce costs,various technologies and research approaches focused on Optimal Camera Placement(OCP)are continually being proposed.However,as most existing studies assume homogeneous camera setups,there is a growing demand for studies on heterogeneous camera setups.For instance,technical demands keep emerging in scenarios with minimal camera configurations,especially regarding cost factors,the physical placement of cameras given the spatial structure,and image capture strategies for heterogeneous cameras,such as high-resolution RGB cameras and depth cameras.In this study,we propose a pre-visualization and simulation method for the optimal placement of heterogeneous cameras in XR environments,accounting for both the specifications of heterogeneous cameras(e.g.,field of view)and the physical configuration(e.g.,wall configuration)in real-world spaces.The proposed method performs a visibility analysis of cameras by considering each camera’s field-of-view volume,resolution,and unique characteristics,along with physicalspace constraints.This approach enables the optimal position and rotation of each camera to be recommended,along with the minimum number of cameras required.In the results of our study conducted in heterogeneous camera combinations,the proposed method achieved 81.7%~82.7%coverage of the target visual information using only 2~3 cameras.In contrast,single(or homogeneous)-typed cameras were required to use 11 cameras for 81.6%coverage.Accordingly,we found that camera deployment resources can be reduced with the proposed approaches.展开更多
To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of ...To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of view and operates at both visible and near-infrared wavelengths.Using the principles of light field imaging,the proposed design enables 3D reconstruction of optical surfaces,thus enabling vertical surface height measurements with enhanced accuracy.Using Zemax-based simulations,we evaluate the system’s modulation transfer function,its optical aberrations,and its tolerance to shape variations through Zernike coefficient adjustments.The results demonstrate that this camera can achieve the required spatial resolution while also maintaining high imaging quality and thus offers a promising solution for advanced optical surface defect inspection.展开更多
Vascular abnormalities are closely associated with the pathogenesis and progression of numerous diseases, such as thrombosis, tumors, and diabetes. Blood flow velocity serves as a critical biomarker for evaluating per...Vascular abnormalities are closely associated with the pathogenesis and progression of numerous diseases, such as thrombosis, tumors, and diabetes. Blood flow velocity serves as a critical biomarker for evaluating perfusion status. Quantitative detection of full-field blood flow variations in lesion areas holds significant scientific and clinical value for pathological studies,diagnosis, and intraoperative monitoring of related diseases. While laser speckle contrast imaging(LSCI) enables full-field blood flow visualization, its reliance on frame-based sensors necessitates handling massive data volumes, leading to inherent trade-offs among spatiotemporal resolution, real-time performance, and quantitative capabilities. Leveraging the asynchronous dynamic sensing, high temporal sampling rate, and low data redundancy of event cameras, this study proposes a quantitative blood flow imaging method termed laser speckle event imaging(LSEI). Experiments using off-the-shelf event cameras demonstrate that LSEI achieves real-time blood flow imaging with minimal computational overhead compared to frame-based LSCI. Furthermore,we investigate the relationship between event data streams and flow velocity through spatial-temporal autocorrelation analysis,enabling quantitative measurements without compromising temporal or spatial resolution. In in vivo imaging experiments of mouse ear blood flow, LSEI exhibits superior imaging details and real-time performance over conventional methods. The proposed approach holds promise as an efficient tool for diagnosis, therapeutic evaluation, and research on vascular-related diseases.展开更多
LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as ...LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as odometry estimation and 3D reconstruction.Fusing the information from these two sensors can significantly increase the robustness and accuracy of these perception tasks.The extrinsic calibration between cameras and LiDAR is a fundamental prerequisite for multimodal systems.Recently,extensive studies have been conducted on the calibration of extrinsic parameters.Although several calibration methods facilitate sensor fusion,a comprehensive summary for researchers and,especially,non-expert users is lacking.Thus,we present an overview of extrinsic calibration and discuss diverse calibration methods from the perspective of calibration system design.Based on the calibration information sources,this study classifies these methods as target-based or targetless.For each type of calibration method,further classification was performed according to the diverse types of features or constraints used in the calibration process,and their detailed implementations and key characteristics were introduced.Thereafter,calibration-accuracy evaluation methods are presented.Finally,we comprehensively compare the advantages and disadvantages of each calibration method and suggest directions for practical applications and future research.展开更多
In this paper,a novel multi-frame track-before-detect algorithm is proposed,which is based on root label clustering to reduce the high computational complexity arising by observation area expansion and clutter/noise d...In this paper,a novel multi-frame track-before-detect algorithm is proposed,which is based on root label clustering to reduce the high computational complexity arising by observation area expansion and clutter/noise density increase.A criterion of track extrapolation is used to construct state transition set,root label is marked by state transition set to obtain the distribution information of multiple targets in measurement space,then measurement plots of multi-frame are divided into several clusters,and finally multi-frame track-before-detect algorithm is implemented in each cluster.The computational complexity can be reduced by employing the proposed algorithm.Simulation results show that the proposed algorithm can accurately detect multiple targets in close proximity and reduce the number of false tracks.展开更多
Multi-frame coding is supported by the emerging H.264. It is important for the enhancement of both coding efficiency and error robustness. In this paper, error resilient schemes for H.264 based on multi-frame were inv...Multi-frame coding is supported by the emerging H.264. It is important for the enhancement of both coding efficiency and error robustness. In this paper, error resilient schemes for H.264 based on multi-frame were investigated. Error robust H.264 video transmission schemes were introduced for the applications with and without a feedback channel. The experimental results demonstrate the effectiveness of the proposed schemes.展开更多
Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution...Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution sparse compound-eye camera(CEC)based on dual-end collaborative optimization is proposed,which provides a cost-effective way to break through the trade-off among the field of view,resolution,and imaging speed.In the optical end,a sparse CEC based on liquid lenses is designed,which can realize large-field-of-view imaging in real time,and fast zooming within 5 ms.In the computational end,a disturbed degradation model driven super-resolution network(DDMDSR-Net)is proposed to deal with complex image degradation issues in actual imaging situations,achieving high-robustness and high-fidelity resolution enhancement.Based on the proposed dual-end collaborative optimization framework,the angular resolution of the CEC can be enhanced from 71.6"to 26.0",which provides a solution to realize high-resolution imaging for array camera dispensing with high optical hardware complexity and data transmission bandwidth.Experiments verify the advantages of the CEC based on dual-end collaborative optimization in high-fidelity reconstruction of real scene images,kilometer-level long-distance detection,and dynamic imaging and precise recognition of targets of interest.展开更多
Observatories typically deploy all-sky cameras for monitoring cloud cover and weather conditions.However,many of these cameras lack scientific-grade sensors,r.esulting in limited photometric precision,which makes calc...Observatories typically deploy all-sky cameras for monitoring cloud cover and weather conditions.However,many of these cameras lack scientific-grade sensors,r.esulting in limited photometric precision,which makes calculating the sky area visibility distribution via extinction measurement challenging.To address this issue,we propose the Photometry-Free Sky Area Visibility Estimation(PFSAVE)method.This method uses the standard magnitude of the faintest star observed within a given sky area to estimate visibility.By employing a pertransformation refitting optimization strategy,we achieve a high-precision coordinate transformation model with an accuracy of 0.42 pixels.Using the results of HEALPix segmentation is also introduced to achieve high spatial resolution.Comprehensive analysis based on real allsky images demonstrates that our method exhibits higher accuracy than the extinction-based method.Our method supports both manual and robotic dynamic scheduling,especially under partially cloudy conditions.展开更多
This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,an...This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,and a CMOS sensor.In view of the significant contrast between face and background in thermal infra⁃red images,this paper explores a suitable accuracy-latency tradeoff for thermal face detection and proposes a tiny,lightweight detector named YOLO-Fastest-IR.Four YOLO-Fastest-IR models(IR0 to IR3)with different scales are designed based on YOLO-Fastest.To train and evaluate these lightweight models,a multi-user low-resolution thermal face database(RGBT-MLTF)was collected,and the four networks were trained.Experiments demon⁃strate that the lightweight convolutional neural network performs well in thermal infrared face detection tasks.The proposed algorithm outperforms existing face detection methods in both positioning accuracy and speed,making it more suitable for deployment on mobile platforms or embedded devices.After obtaining the region of interest(ROI)in the infrared(IR)image,the RGB camera is guided by the thermal infrared face detection results to achieve fine positioning of the RGB face.Experimental results show that YOLO-Fastest-IR achieves a frame rate of 92.9 FPS on a Raspberry Pi 4B and successfully detects 97.4%of faces in the RGBT-MLTF test set.Ultimate⁃ly,an infrared temperature measurement system with low cost,strong robustness,and high real-time perfor⁃mance was integrated,achieving a temperature measurement accuracy of 0.3℃.展开更多
Blackmagic Camera for iOS 3.2版本可直接从i Phone推流到You Tube、Vimeo和Twitch,无需借助其他硬件或第三方软件,只要选择平台、输入推流密钥,就能以专业质量直播。本次更新还加入了对SRT推流的支持,可以把高质量、低延迟的视频发送...Blackmagic Camera for iOS 3.2版本可直接从i Phone推流到You Tube、Vimeo和Twitch,无需借助其他硬件或第三方软件,只要选择平台、输入推流密钥,就能以专业质量直播。本次更新还加入了对SRT推流的支持,可以把高质量、低延迟的视频发送到Blackmagic Streaming Decoder或其他支持SRT的广电制作系统,非常适合远程制作、新闻采集、活动实况报道。另外,现在用外部存储时会有更清晰的通知信息。当硬盘在录制、待机或上传素材过程中被断开时,警告会立即出现。展开更多
Camera technology advancement and deployment continue to play a vital role in modern life and production,amongst other capabilities,generating relevant visual information.The challenge of optimizing the use and deploy...Camera technology advancement and deployment continue to play a vital role in modern life and production,amongst other capabilities,generating relevant visual information.The challenge of optimizing the use and deployment of camera networks in various applications(e.g.,surveillance,traffic monitoring,and public safety to name but just a few)has attracted considerable attention from both academia and industries.Camera planning is the first step in addressing this challenge.The surveillance objectives and scenes of a camera network dictate the modelling and optimization algorithms for camera planning.However,existing reviews have primarily focused on models or optimization algorithms,with insufficient attention given to surveillance scenes.This review aims to bridge this gap by 1)Classifying surveillance scenes into the urban environment and rural outdoor environment and comparing the surveillance requirements and challenges;2)Summarizing the details of camera coverage optimization in the relevant literature from the perspective of deployment scenes;and 3)Proposing a new surveillance scene-Solar Insecticidal Lamps with the Internet of Things—as a case study to analyze the surveillance requirement and challenges in agricultural outdoor environment.Finally,we state the technical outlook on the physical safety of outdoor electronic devices in agriculture settings and provide insights to draw more attention and effort into this area.展开更多
This paper designs and implements a single-camera 360°panoramic imaging system based on motor-driven fisheye rotation.The system utilizes a stepper motor for precise angular control,enabling the camera to rotate ...This paper designs and implements a single-camera 360°panoramic imaging system based on motor-driven fisheye rotation.The system utilizes a stepper motor for precise angular control,enabling the camera to rotate around its optical center to capture multi-view images,thereby avoiding the parallax and geometric mismatch problems inherent in traditional multi-camera configurations.To address the strong distortion characteristics of fisheye images,an equidistant projection model is adopted for distortion correction.On this basis,a brightness normalization method combining global linear brightness correction and local illumination compensation is proposed to enhance stitching consistency.By establishing a geometry model constrained by camera rotation and integrating cylindrical projection with cosine-weighted blending,the system achieves high-precision panoramic stitching and seamless visual transitions.展开更多
Photomechanics is a crucial branch of solid mechanics.The localization of point targets constitutes a fundamental problem in optical experimental mechanics,with extensive applications in various missions of unmanned a...Photomechanics is a crucial branch of solid mechanics.The localization of point targets constitutes a fundamental problem in optical experimental mechanics,with extensive applications in various missions of unmanned aerial vehicles.Localizing moving targets is crucial for analyzing their motion characteristics and dynamic properties.Reconstructing the trajectories of points from asynchronous cameras is a significant challenge.It encompasses two coupled sub-problems:Trajectory reconstruction and camera synchronization.Present methods typically address only one of these sub-problems individually.This paper proposes a 3D trajectory reconstruction method for point targets based on asynchronous cameras,simultaneously solving both sub-problems.Firstly,we extend the trajectory intersection method to asynchronous cameras to resolve the limitation of traditional triangulation that requires camera synchronization.Secondly,we develop models for camera temporal information and target motion,based on imaging mechanisms and target dynamics characteristics.The parameters are optimized simultaneously to achieve trajectory reconstruction without accurate time parameters.Thirdly,we optimize the camera rotations alongside the camera time information and target motion parameters,using tighter and more continuous constraints on moving points.The reconstruction accuracy is significantly improved,especially when the camera rotations are inaccurate.Finally,the simulated and real-world experimental results demonstrate the feasibility and accuracy of the proposed method.The real-world results indicate that the proposed algorithm achieved a localization error of 112.95 m at an observation distance range of 15-20 km.展开更多
In various imaging applications such as autonomous vehicles and drones,autofocus lenses are indispensable for capturing clear images.However,conventional camera calibration methods typically rely either on processing ...In various imaging applications such as autonomous vehicles and drones,autofocus lenses are indispensable for capturing clear images.However,conventional camera calibration methods typically rely either on processing multiple images at a fixed focal length or on detecting multi-plane markers in a single image and then applying multi-image calibration models.This paper proposes a flexible and accurate calibration approach that extracts subpixel saddle points from a single image containing three non-coplanar calibration boards.To compute accurate homography matrices for the three boards,outliers are removed by eliminating chessboard points that deviated from the fitted grid lines according to their row and column positions.Initial estimates of the intrinsic parameters and the poses of the three planar chessboards are obtained using the three homography matrices in combination with Zhang’s calibration method.During parameter refinement,a multi-objective optimization function is constructed,incorporating three error terms:(1)Reprojection error of the inlier grid points;(2)Mechanism-driven error derived from the relationship between homography matrices and camera parameters;(3)Cross-planar linearity constraint error,which preserves the pre-imaging collinearity of any five points across different planes after projection.For weight selection in the optimization process,confidence intervals of the detected grid points are analyzed by horizontally rotating the reprojection lines to reduce bias introduced by line slope.The optimal weights are determined by minimizing the number of points whose confidence intervals does not intersect the reprojected lines.When multiple candidates yield similar reprojection performance,the parameter set with the smallest reprojection error is selected as the final result.This method efficiently estimates both intrinsic and extrinsic camera parameters.Simulations and real-world experiments validate the high precision and effectiveness of the proposed approach.Our technique is straightforward,practical,and holds significant theoretical and practical value for rapid and reliable camera calibration.展开更多
It is important to understand the development of joints and fractures in rock masses to ensure drilling stability and blasting effectiveness.Traditional manual observation techniques for identifying and extracting fra...It is important to understand the development of joints and fractures in rock masses to ensure drilling stability and blasting effectiveness.Traditional manual observation techniques for identifying and extracting fracture characteristics have been proven to be inefficient and prone to subjective interpretation.Moreover,conventional image processing algorithms and classical deep learning models often encounter difficulties in accurately identifying fracture areas,resulting in unclear contours.This study proposes an intelligent method for detecting internal fractures in mine rock masses to address these challenges.The proposed approach captures a nodal fracture map within the targeted blast area and integrates channel and spatial attention mechanisms into the ResUnet(RU)model.The channel attention mechanism dynamically recalibrates the importance of each feature channel,and the spatial attention mechanism enhances feature representation in key areas while minimizing background noise,thus improving segmentation accuracy.A dynamic serpentine convolution module is also introduced that adaptively adjusts the shape and orientation of the convolution kernel based on the local structure of the input feature map.Furthermore,this method enables the automatic extraction and quantification of borehole nodal fracture information by fitting sinusoidal curves to the boundaries of the fracture contours using the least squares method.In comparison to other advanced deep learning models,our enhanced RU demonstrates superior performance across evaluation metrics,including accuracy,pixel accuracy(PA),and intersection over union(IoU).Unlike traditional manual extraction methods,our intelligent detection approach provides considerable time and cost savings,with an average error rate of approximately 4%.This approach has the potential to greatly improve the efficiency of geological surveys of borehole fractures.展开更多
This study presents a drone-based aerial imaging method for automated rice seedling detection and counting in paddy fields.Utilizing a drone equipped with a high-resolution camera,images are captured 14 days postsowin...This study presents a drone-based aerial imaging method for automated rice seedling detection and counting in paddy fields.Utilizing a drone equipped with a high-resolution camera,images are captured 14 days postsowing at a consistent altitude of six meters,employing autonomous flight for uniform data acquisition.The approach effectively addresses the distinct growth patterns of both single and clustered rice seedlings at this early stage.The methodology follows a two-step process:first,the GoogleNet deep learning network identifies the location and center points of rice plants.Then,the U-Net deep learning network performs classification and counting of individual plants and clusters.This combination of deep learning models achieved a 90%accuracy rate in classifying and counting both single and clustered seedlings.To validate the method’s effectiveness,results were compared against traditional manual counting conducted by agricultural experts.The comparison revealed minimal discrepancies,with a variance of only 2–4 clumps per square meter,confirming the reliability of the proposed method.This automated approach offers significant benefits by providing an efficient,accurate,and scalable solution for monitoring seedling growth.It enables farmers to optimize fertilizer and pesticide application,improve resource allocation,and enhance overall crop management,ultimately contributing to increased agricultural productivity.展开更多
基金The National Natural Science Foundation of China(No.62388102)the Natural Science Foundation of Shandong Province(No.ZR2021MF134).
文摘In radar automatic target recognition(RATR),the high-resolution range profile(HRRP)has garnered considerable attention owing to its minimal computational demands.However,radar HRRP target recognition still faces numerous challenges,primarily due to substantial variations in the amplitude and distribution of HRRP scattering points because of slight azimuthal changes.To alleviate the effect of aspect sensitivity,a novel multi-frame attention network(MFA-Net)comprising a range deformable convolution module(RDCM),multi-frame attention module(MFAM),and global-local Transformer module(GLTM)is proposed.The RDCM is designed to adaptively learn the distance of scattering center migration.Subsequently,the MFAM extracts consistent features across different frames to alleviate the influence of power fluctuation.Finally,the GLTM allocates attention between global and local fea-tures.The feasibility and effectiveness of the proposed method are validated through simulation and experimental datasets,and the recognition rate is enhanced by more than 3%compared to the state-of-the-art methods.
基金supported by the 2024 Research Fund of University of Ulsan.
文摘In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In addition,methods for deploying multiple cameras for motion capture of users(e.g.,performers)are widely used in computer graphics.As the need to minimize and optimize the number of cameras grows to reduce costs,various technologies and research approaches focused on Optimal Camera Placement(OCP)are continually being proposed.However,as most existing studies assume homogeneous camera setups,there is a growing demand for studies on heterogeneous camera setups.For instance,technical demands keep emerging in scenarios with minimal camera configurations,especially regarding cost factors,the physical placement of cameras given the spatial structure,and image capture strategies for heterogeneous cameras,such as high-resolution RGB cameras and depth cameras.In this study,we propose a pre-visualization and simulation method for the optimal placement of heterogeneous cameras in XR environments,accounting for both the specifications of heterogeneous cameras(e.g.,field of view)and the physical configuration(e.g.,wall configuration)in real-world spaces.The proposed method performs a visibility analysis of cameras by considering each camera’s field-of-view volume,resolution,and unique characteristics,along with physicalspace constraints.This approach enables the optimal position and rotation of each camera to be recommended,along with the minimum number of cameras required.In the results of our study conducted in heterogeneous camera combinations,the proposed method achieved 81.7%~82.7%coverage of the target visual information using only 2~3 cameras.In contrast,single(or homogeneous)-typed cameras were required to use 11 cameras for 81.6%coverage.Accordingly,we found that camera deployment resources can be reduced with the proposed approaches.
基金supported by the Jilin Science and Technology Development Plan (20240101029JJ) for the following study:synchronized high-speed detection of surface shape and defects in the grinding stage of complex surfaces (KLMSZZ202305)for the high-precision wide dynamic large aperture optical inspection system for fine astronomical observation by the National Major Research Instrument Development Project (62127901)+2 种基金for ultrasmooth manufacturing technology of large diameter complex curved surface by the National Key R&D Program(2022YFB3403405)for research on the key technology of rapid synchronous detection of surface shape and subsurface defects in the grinding stage of large diameter complex surfaces by the International Cooperation Project(2025010157)The Key Laboratory of Optical System Advanced Manufacturing Technology,Chinese Academy of Sciences (2022KLOMT02-04) also supported this study
文摘To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of view and operates at both visible and near-infrared wavelengths.Using the principles of light field imaging,the proposed design enables 3D reconstruction of optical surfaces,thus enabling vertical surface height measurements with enhanced accuracy.Using Zemax-based simulations,we evaluate the system’s modulation transfer function,its optical aberrations,and its tolerance to shape variations through Zernike coefficient adjustments.The results demonstrate that this camera can achieve the required spatial resolution while also maintaining high imaging quality and thus offers a promising solution for advanced optical surface defect inspection.
基金supported by the National Natural Science Foundation of China (Grant No.12572210)the Scientific Instrument Developing Project of Shenzhen University (Grant Nos.2023YQ011,2024YQ001)the Shenzhen Science and Technology Innovation Commission Project—Stable Support (General Project)(Grant No.20231120175055001)。
文摘Vascular abnormalities are closely associated with the pathogenesis and progression of numerous diseases, such as thrombosis, tumors, and diabetes. Blood flow velocity serves as a critical biomarker for evaluating perfusion status. Quantitative detection of full-field blood flow variations in lesion areas holds significant scientific and clinical value for pathological studies,diagnosis, and intraoperative monitoring of related diseases. While laser speckle contrast imaging(LSCI) enables full-field blood flow visualization, its reliance on frame-based sensors necessitates handling massive data volumes, leading to inherent trade-offs among spatiotemporal resolution, real-time performance, and quantitative capabilities. Leveraging the asynchronous dynamic sensing, high temporal sampling rate, and low data redundancy of event cameras, this study proposes a quantitative blood flow imaging method termed laser speckle event imaging(LSEI). Experiments using off-the-shelf event cameras demonstrate that LSEI achieves real-time blood flow imaging with minimal computational overhead compared to frame-based LSCI. Furthermore,we investigate the relationship between event data streams and flow velocity through spatial-temporal autocorrelation analysis,enabling quantitative measurements without compromising temporal or spatial resolution. In in vivo imaging experiments of mouse ear blood flow, LSEI exhibits superior imaging details and real-time performance over conventional methods. The proposed approach holds promise as an efficient tool for diagnosis, therapeutic evaluation, and research on vascular-related diseases.
基金Supported by Beijing Natural Science Foundation(Grant No.L241012)the National Natural Science Foundation of China(Grant No.62572468).
文摘LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as odometry estimation and 3D reconstruction.Fusing the information from these two sensors can significantly increase the robustness and accuracy of these perception tasks.The extrinsic calibration between cameras and LiDAR is a fundamental prerequisite for multimodal systems.Recently,extensive studies have been conducted on the calibration of extrinsic parameters.Although several calibration methods facilitate sensor fusion,a comprehensive summary for researchers and,especially,non-expert users is lacking.Thus,we present an overview of extrinsic calibration and discuss diverse calibration methods from the perspective of calibration system design.Based on the calibration information sources,this study classifies these methods as target-based or targetless.For each type of calibration method,further classification was performed according to the diverse types of features or constraints used in the calibration process,and their detailed implementations and key characteristics were introduced.Thereafter,calibration-accuracy evaluation methods are presented.Finally,we comprehensively compare the advantages and disadvantages of each calibration method and suggest directions for practical applications and future research.
基金supported by the Innovation Project of Science and Technology Commission of the Central Military Commission,China(No.19-HXXX-01-ZD-006-XXX-XX)。
文摘In this paper,a novel multi-frame track-before-detect algorithm is proposed,which is based on root label clustering to reduce the high computational complexity arising by observation area expansion and clutter/noise density increase.A criterion of track extrapolation is used to construct state transition set,root label is marked by state transition set to obtain the distribution information of multiple targets in measurement space,then measurement plots of multi-frame are divided into several clusters,and finally multi-frame track-before-detect algorithm is implemented in each cluster.The computational complexity can be reduced by employing the proposed algorithm.Simulation results show that the proposed algorithm can accurately detect multiple targets in close proximity and reduce the number of false tracks.
文摘Multi-frame coding is supported by the emerging H.264. It is important for the enhancement of both coding efficiency and error robustness. In this paper, error resilient schemes for H.264 based on multi-frame were investigated. Error robust H.264 video transmission schemes were introduced for the applications with and without a feedback channel. The experimental results demonstrate the effectiveness of the proposed schemes.
基金financial supports from National Natural Science Foundation of China(Grant Nos.U23A20368 and 62175006)Academic Excellence Foundation of BUAA for PhD Students.
文摘Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution sparse compound-eye camera(CEC)based on dual-end collaborative optimization is proposed,which provides a cost-effective way to break through the trade-off among the field of view,resolution,and imaging speed.In the optical end,a sparse CEC based on liquid lenses is designed,which can realize large-field-of-view imaging in real time,and fast zooming within 5 ms.In the computational end,a disturbed degradation model driven super-resolution network(DDMDSR-Net)is proposed to deal with complex image degradation issues in actual imaging situations,achieving high-robustness and high-fidelity resolution enhancement.Based on the proposed dual-end collaborative optimization framework,the angular resolution of the CEC can be enhanced from 71.6"to 26.0",which provides a solution to realize high-resolution imaging for array camera dispensing with high optical hardware complexity and data transmission bandwidth.Experiments verify the advantages of the CEC based on dual-end collaborative optimization in high-fidelity reconstruction of real scene images,kilometer-level long-distance detection,and dynamic imaging and precise recognition of targets of interest.
基金supported by Natural Science Foundation of Jilin Province(20210101468JC)Chinese Academy of Sciences and Local Government Cooperation Project(2023SYHZ0027,23SH04)National Natural Science Foundation of China(12273063&12203078)。
文摘Observatories typically deploy all-sky cameras for monitoring cloud cover and weather conditions.However,many of these cameras lack scientific-grade sensors,r.esulting in limited photometric precision,which makes calculating the sky area visibility distribution via extinction measurement challenging.To address this issue,we propose the Photometry-Free Sky Area Visibility Estimation(PFSAVE)method.This method uses the standard magnitude of the faintest star observed within a given sky area to estimate visibility.By employing a pertransformation refitting optimization strategy,we achieve a high-precision coordinate transformation model with an accuracy of 0.42 pixels.Using the results of HEALPix segmentation is also introduced to achieve high spatial resolution.Comprehensive analysis based on real allsky images demonstrates that our method exhibits higher accuracy than the extinction-based method.Our method supports both manual and robotic dynamic scheduling,especially under partially cloudy conditions.
基金Supported by the Fundamental Research Funds for the Central Universities(2024300443)the Natural Science Foundation of Jiangsu Province(BK20241224).
文摘This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,and a CMOS sensor.In view of the significant contrast between face and background in thermal infra⁃red images,this paper explores a suitable accuracy-latency tradeoff for thermal face detection and proposes a tiny,lightweight detector named YOLO-Fastest-IR.Four YOLO-Fastest-IR models(IR0 to IR3)with different scales are designed based on YOLO-Fastest.To train and evaluate these lightweight models,a multi-user low-resolution thermal face database(RGBT-MLTF)was collected,and the four networks were trained.Experiments demon⁃strate that the lightweight convolutional neural network performs well in thermal infrared face detection tasks.The proposed algorithm outperforms existing face detection methods in both positioning accuracy and speed,making it more suitable for deployment on mobile platforms or embedded devices.After obtaining the region of interest(ROI)in the infrared(IR)image,the RGB camera is guided by the thermal infrared face detection results to achieve fine positioning of the RGB face.Experimental results show that YOLO-Fastest-IR achieves a frame rate of 92.9 FPS on a Raspberry Pi 4B and successfully detects 97.4%of faces in the RGBT-MLTF test set.Ultimate⁃ly,an infrared temperature measurement system with low cost,strong robustness,and high real-time perfor⁃mance was integrated,achieving a temperature measurement accuracy of 0.3℃.
文摘Blackmagic Camera for iOS 3.2版本可直接从i Phone推流到You Tube、Vimeo和Twitch,无需借助其他硬件或第三方软件,只要选择平台、输入推流密钥,就能以专业质量直播。本次更新还加入了对SRT推流的支持,可以把高质量、低延迟的视频发送到Blackmagic Streaming Decoder或其他支持SRT的广电制作系统,非常适合远程制作、新闻采集、活动实况报道。另外,现在用外部存储时会有更清晰的通知信息。当硬盘在录制、待机或上传素材过程中被断开时,警告会立即出现。
基金supported by the National Natural Science Foundation of China(62072248).
文摘Camera technology advancement and deployment continue to play a vital role in modern life and production,amongst other capabilities,generating relevant visual information.The challenge of optimizing the use and deployment of camera networks in various applications(e.g.,surveillance,traffic monitoring,and public safety to name but just a few)has attracted considerable attention from both academia and industries.Camera planning is the first step in addressing this challenge.The surveillance objectives and scenes of a camera network dictate the modelling and optimization algorithms for camera planning.However,existing reviews have primarily focused on models or optimization algorithms,with insufficient attention given to surveillance scenes.This review aims to bridge this gap by 1)Classifying surveillance scenes into the urban environment and rural outdoor environment and comparing the surveillance requirements and challenges;2)Summarizing the details of camera coverage optimization in the relevant literature from the perspective of deployment scenes;and 3)Proposing a new surveillance scene-Solar Insecticidal Lamps with the Internet of Things—as a case study to analyze the surveillance requirement and challenges in agricultural outdoor environment.Finally,we state the technical outlook on the physical safety of outdoor electronic devices in agriculture settings and provide insights to draw more attention and effort into this area.
基金Graduate Innovation Ability Training Program of the Hebei Provincial Department of Education,2025(Project No.:CXZZSS2025095)。
文摘This paper designs and implements a single-camera 360°panoramic imaging system based on motor-driven fisheye rotation.The system utilizes a stepper motor for precise angular control,enabling the camera to rotate around its optical center to capture multi-view images,thereby avoiding the parallax and geometric mismatch problems inherent in traditional multi-camera configurations.To address the strong distortion characteristics of fisheye images,an equidistant projection model is adopted for distortion correction.On this basis,a brightness normalization method combining global linear brightness correction and local illumination compensation is proposed to enhance stitching consistency.By establishing a geometry model constrained by camera rotation and integrating cylindrical projection with cosine-weighted blending,the system achieves high-precision panoramic stitching and seamless visual transitions.
基金supported by the Hunan Provin〓〓cial Natural Science Foundation for Excellent Young Scholars(Grant No.2023JJ20045)the National Natural Science Foundation of China(Grant No.12372189)。
文摘Photomechanics is a crucial branch of solid mechanics.The localization of point targets constitutes a fundamental problem in optical experimental mechanics,with extensive applications in various missions of unmanned aerial vehicles.Localizing moving targets is crucial for analyzing their motion characteristics and dynamic properties.Reconstructing the trajectories of points from asynchronous cameras is a significant challenge.It encompasses two coupled sub-problems:Trajectory reconstruction and camera synchronization.Present methods typically address only one of these sub-problems individually.This paper proposes a 3D trajectory reconstruction method for point targets based on asynchronous cameras,simultaneously solving both sub-problems.Firstly,we extend the trajectory intersection method to asynchronous cameras to resolve the limitation of traditional triangulation that requires camera synchronization.Secondly,we develop models for camera temporal information and target motion,based on imaging mechanisms and target dynamics characteristics.The parameters are optimized simultaneously to achieve trajectory reconstruction without accurate time parameters.Thirdly,we optimize the camera rotations alongside the camera time information and target motion parameters,using tighter and more continuous constraints on moving points.The reconstruction accuracy is significantly improved,especially when the camera rotations are inaccurate.Finally,the simulated and real-world experimental results demonstrate the feasibility and accuracy of the proposed method.The real-world results indicate that the proposed algorithm achieved a localization error of 112.95 m at an observation distance range of 15-20 km.
基金supported by the Research on the Reform of Curriculum Assessment Methods for College Mathematics Platform Courses(No.53111104016)。
文摘In various imaging applications such as autonomous vehicles and drones,autofocus lenses are indispensable for capturing clear images.However,conventional camera calibration methods typically rely either on processing multiple images at a fixed focal length or on detecting multi-plane markers in a single image and then applying multi-image calibration models.This paper proposes a flexible and accurate calibration approach that extracts subpixel saddle points from a single image containing three non-coplanar calibration boards.To compute accurate homography matrices for the three boards,outliers are removed by eliminating chessboard points that deviated from the fitted grid lines according to their row and column positions.Initial estimates of the intrinsic parameters and the poses of the three planar chessboards are obtained using the three homography matrices in combination with Zhang’s calibration method.During parameter refinement,a multi-objective optimization function is constructed,incorporating three error terms:(1)Reprojection error of the inlier grid points;(2)Mechanism-driven error derived from the relationship between homography matrices and camera parameters;(3)Cross-planar linearity constraint error,which preserves the pre-imaging collinearity of any five points across different planes after projection.For weight selection in the optimization process,confidence intervals of the detected grid points are analyzed by horizontally rotating the reprojection lines to reduce bias introduced by line slope.The optimal weights are determined by minimizing the number of points whose confidence intervals does not intersect the reprojected lines.When multiple candidates yield similar reprojection performance,the parameter set with the smallest reprojection error is selected as the final result.This method efficiently estimates both intrinsic and extrinsic camera parameters.Simulations and real-world experiments validate the high precision and effectiveness of the proposed approach.Our technique is straightforward,practical,and holds significant theoretical and practical value for rapid and reliable camera calibration.
基金supported by the National Natural Science Foundation of China(No.52474172).
文摘It is important to understand the development of joints and fractures in rock masses to ensure drilling stability and blasting effectiveness.Traditional manual observation techniques for identifying and extracting fracture characteristics have been proven to be inefficient and prone to subjective interpretation.Moreover,conventional image processing algorithms and classical deep learning models often encounter difficulties in accurately identifying fracture areas,resulting in unclear contours.This study proposes an intelligent method for detecting internal fractures in mine rock masses to address these challenges.The proposed approach captures a nodal fracture map within the targeted blast area and integrates channel and spatial attention mechanisms into the ResUnet(RU)model.The channel attention mechanism dynamically recalibrates the importance of each feature channel,and the spatial attention mechanism enhances feature representation in key areas while minimizing background noise,thus improving segmentation accuracy.A dynamic serpentine convolution module is also introduced that adaptively adjusts the shape and orientation of the convolution kernel based on the local structure of the input feature map.Furthermore,this method enables the automatic extraction and quantification of borehole nodal fracture information by fitting sinusoidal curves to the boundaries of the fracture contours using the least squares method.In comparison to other advanced deep learning models,our enhanced RU demonstrates superior performance across evaluation metrics,including accuracy,pixel accuracy(PA),and intersection over union(IoU).Unlike traditional manual extraction methods,our intelligent detection approach provides considerable time and cost savings,with an average error rate of approximately 4%.This approach has the potential to greatly improve the efficiency of geological surveys of borehole fractures.
基金funded by the Ministry of Education and Training Project(code number:B2023-TCT-08).
文摘This study presents a drone-based aerial imaging method for automated rice seedling detection and counting in paddy fields.Utilizing a drone equipped with a high-resolution camera,images are captured 14 days postsowing at a consistent altitude of six meters,employing autonomous flight for uniform data acquisition.The approach effectively addresses the distinct growth patterns of both single and clustered rice seedlings at this early stage.The methodology follows a two-step process:first,the GoogleNet deep learning network identifies the location and center points of rice plants.Then,the U-Net deep learning network performs classification and counting of individual plants and clusters.This combination of deep learning models achieved a 90%accuracy rate in classifying and counting both single and clustered seedlings.To validate the method’s effectiveness,results were compared against traditional manual counting conducted by agricultural experts.The comparison revealed minimal discrepancies,with a variance of only 2–4 clumps per square meter,confirming the reliability of the proposed method.This automated approach offers significant benefits by providing an efficient,accurate,and scalable solution for monitoring seedling growth.It enables farmers to optimize fertilizer and pesticide application,improve resource allocation,and enhance overall crop management,ultimately contributing to increased agricultural productivity.