In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In additi...In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In addition,methods for deploying multiple cameras for motion capture of users(e.g.,performers)are widely used in computer graphics.As the need to minimize and optimize the number of cameras grows to reduce costs,various technologies and research approaches focused on Optimal Camera Placement(OCP)are continually being proposed.However,as most existing studies assume homogeneous camera setups,there is a growing demand for studies on heterogeneous camera setups.For instance,technical demands keep emerging in scenarios with minimal camera configurations,especially regarding cost factors,the physical placement of cameras given the spatial structure,and image capture strategies for heterogeneous cameras,such as high-resolution RGB cameras and depth cameras.In this study,we propose a pre-visualization and simulation method for the optimal placement of heterogeneous cameras in XR environments,accounting for both the specifications of heterogeneous cameras(e.g.,field of view)and the physical configuration(e.g.,wall configuration)in real-world spaces.The proposed method performs a visibility analysis of cameras by considering each camera’s field-of-view volume,resolution,and unique characteristics,along with physicalspace constraints.This approach enables the optimal position and rotation of each camera to be recommended,along with the minimum number of cameras required.In the results of our study conducted in heterogeneous camera combinations,the proposed method achieved 81.7%~82.7%coverage of the target visual information using only 2~3 cameras.In contrast,single(or homogeneous)-typed cameras were required to use 11 cameras for 81.6%coverage.Accordingly,we found that camera deployment resources can be reduced with the proposed approaches.展开更多
To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of ...To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of view and operates at both visible and near-infrared wavelengths.Using the principles of light field imaging,the proposed design enables 3D reconstruction of optical surfaces,thus enabling vertical surface height measurements with enhanced accuracy.Using Zemax-based simulations,we evaluate the system’s modulation transfer function,its optical aberrations,and its tolerance to shape variations through Zernike coefficient adjustments.The results demonstrate that this camera can achieve the required spatial resolution while also maintaining high imaging quality and thus offers a promising solution for advanced optical surface defect inspection.展开更多
Background:Studies have shown that heart rate variability(HRV)is a predictor of the prognosis of cardiovascular diseases.Contact heartbeat monitoring equipment is widely used,especially in hospitals,and benefits from ...Background:Studies have shown that heart rate variability(HRV)is a predictor of the prognosis of cardiovascular diseases.Contact heartbeat monitoring equipment is widely used,especially in hospitals,and benefits from the rapidity and accuracy of the detection of physiological health indicators.However,long-term contact with equipment has many adverse effects.The purpose of this study was to improve the accuracy of HRV detection via noncontact equipment,thus enabling HRV to be assessed in various scenarios.Methods:A novel deep learning approach was proposed for measuring heartbeats through camera videos.First,we performed facial segmentation and divided the face into 16 grid cells with different light balance scores.After the trend is filtered by the Hamming window,a transformer-based neural network is used to further filter the signal.Finally,heart rate(HR)and HRV are estimated.Results:We used 1 million synthetic data points for pretraining and a public dataset in combination with a dataset that we constructed for task training.The final results were obtained on a test dataset that we constructed.The accuracy for HR with a low light balance score(0.867-0.983)was greater than that with a high score(0.667-0.750).Our method had higher accuracy in estimating HR than traditional filtering methods(0.167-0.417)and state-of-the-art neural network filtering methods(0.783-0.917)did.The root mean square error of the HRV from the time domain was the lowest,and the correlation index score was the highest for the HRV from the frequency domain estimated by our method compared with those estimated by two neural networks.Conclusions:Light balance,large sample training,and two-stage training can improve the accuracy of HRV estimation.展开更多
LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as ...LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as odometry estimation and 3D reconstruction.Fusing the information from these two sensors can significantly increase the robustness and accuracy of these perception tasks.The extrinsic calibration between cameras and LiDAR is a fundamental prerequisite for multimodal systems.Recently,extensive studies have been conducted on the calibration of extrinsic parameters.Although several calibration methods facilitate sensor fusion,a comprehensive summary for researchers and,especially,non-expert users is lacking.Thus,we present an overview of extrinsic calibration and discuss diverse calibration methods from the perspective of calibration system design.Based on the calibration information sources,this study classifies these methods as target-based or targetless.For each type of calibration method,further classification was performed according to the diverse types of features or constraints used in the calibration process,and their detailed implementations and key characteristics were introduced.Thereafter,calibration-accuracy evaluation methods are presented.Finally,we comprehensively compare the advantages and disadvantages of each calibration method and suggest directions for practical applications and future research.展开更多
介绍一种应用于USB video camera中的自动对焦系统。由USB video camera获取的视频图像经计算机进行FFT运算或微分运算,得到其频谱幅值数据或微分幅值数据,计算机根据所得数据判断USB video camera中的镜头是否处于离焦位置并控制电机...介绍一种应用于USB video camera中的自动对焦系统。由USB video camera获取的视频图像经计算机进行FFT运算或微分运算,得到其频谱幅值数据或微分幅值数据,计算机根据所得数据判断USB video camera中的镜头是否处于离焦位置并控制电机将镜头移到对焦位置。文章还进一步讨论了提高自动对焦准确度的措施。实验结果表明该自动对焦系统能很好地实现USB video camera的自动对焦,该系统将使具有USB接口的video camera使用更简单方便。展开更多
Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution...Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution sparse compound-eye camera(CEC)based on dual-end collaborative optimization is proposed,which provides a cost-effective way to break through the trade-off among the field of view,resolution,and imaging speed.In the optical end,a sparse CEC based on liquid lenses is designed,which can realize large-field-of-view imaging in real time,and fast zooming within 5 ms.In the computational end,a disturbed degradation model driven super-resolution network(DDMDSR-Net)is proposed to deal with complex image degradation issues in actual imaging situations,achieving high-robustness and high-fidelity resolution enhancement.Based on the proposed dual-end collaborative optimization framework,the angular resolution of the CEC can be enhanced from 71.6"to 26.0",which provides a solution to realize high-resolution imaging for array camera dispensing with high optical hardware complexity and data transmission bandwidth.Experiments verify the advantages of the CEC based on dual-end collaborative optimization in high-fidelity reconstruction of real scene images,kilometer-level long-distance detection,and dynamic imaging and precise recognition of targets of interest.展开更多
基金supported by the 2024 Research Fund of University of Ulsan.
文摘In recent years,three-dimensional reconstruction technologies that employ multiple cameras have continued to evolve significantly,enabling remote collaboration among users in extended Reality(XR)environments.In addition,methods for deploying multiple cameras for motion capture of users(e.g.,performers)are widely used in computer graphics.As the need to minimize and optimize the number of cameras grows to reduce costs,various technologies and research approaches focused on Optimal Camera Placement(OCP)are continually being proposed.However,as most existing studies assume homogeneous camera setups,there is a growing demand for studies on heterogeneous camera setups.For instance,technical demands keep emerging in scenarios with minimal camera configurations,especially regarding cost factors,the physical placement of cameras given the spatial structure,and image capture strategies for heterogeneous cameras,such as high-resolution RGB cameras and depth cameras.In this study,we propose a pre-visualization and simulation method for the optimal placement of heterogeneous cameras in XR environments,accounting for both the specifications of heterogeneous cameras(e.g.,field of view)and the physical configuration(e.g.,wall configuration)in real-world spaces.The proposed method performs a visibility analysis of cameras by considering each camera’s field-of-view volume,resolution,and unique characteristics,along with physicalspace constraints.This approach enables the optimal position and rotation of each camera to be recommended,along with the minimum number of cameras required.In the results of our study conducted in heterogeneous camera combinations,the proposed method achieved 81.7%~82.7%coverage of the target visual information using only 2~3 cameras.In contrast,single(or homogeneous)-typed cameras were required to use 11 cameras for 81.6%coverage.Accordingly,we found that camera deployment resources can be reduced with the proposed approaches.
基金supported by the Jilin Science and Technology Development Plan (20240101029JJ) for the following study:synchronized high-speed detection of surface shape and defects in the grinding stage of complex surfaces (KLMSZZ202305)for the high-precision wide dynamic large aperture optical inspection system for fine astronomical observation by the National Major Research Instrument Development Project (62127901)+2 种基金for ultrasmooth manufacturing technology of large diameter complex curved surface by the National Key R&D Program(2022YFB3403405)for research on the key technology of rapid synchronous detection of surface shape and subsurface defects in the grinding stage of large diameter complex surfaces by the International Cooperation Project(2025010157)The Key Laboratory of Optical System Advanced Manufacturing Technology,Chinese Academy of Sciences (2022KLOMT02-04) also supported this study
文摘To address the challenges of high-precision optical surface defect detection,we propose a novel design for a wide-field and broadband light field camera in this work.The proposed system can achieve a 50°field of view and operates at both visible and near-infrared wavelengths.Using the principles of light field imaging,the proposed design enables 3D reconstruction of optical surfaces,thus enabling vertical surface height measurements with enhanced accuracy.Using Zemax-based simulations,we evaluate the system’s modulation transfer function,its optical aberrations,and its tolerance to shape variations through Zernike coefficient adjustments.The results demonstrate that this camera can achieve the required spatial resolution while also maintaining high imaging quality and thus offers a promising solution for advanced optical surface defect inspection.
基金National Natural Science Foundation of China,Grant/Award Number:72204169Department of Science and Technology of Sichuan Province,Grant/Award Number:2021YFS0393。
文摘Background:Studies have shown that heart rate variability(HRV)is a predictor of the prognosis of cardiovascular diseases.Contact heartbeat monitoring equipment is widely used,especially in hospitals,and benefits from the rapidity and accuracy of the detection of physiological health indicators.However,long-term contact with equipment has many adverse effects.The purpose of this study was to improve the accuracy of HRV detection via noncontact equipment,thus enabling HRV to be assessed in various scenarios.Methods:A novel deep learning approach was proposed for measuring heartbeats through camera videos.First,we performed facial segmentation and divided the face into 16 grid cells with different light balance scores.After the trend is filtered by the Hamming window,a transformer-based neural network is used to further filter the signal.Finally,heart rate(HR)and HRV are estimated.Results:We used 1 million synthetic data points for pretraining and a public dataset in combination with a dataset that we constructed for task training.The final results were obtained on a test dataset that we constructed.The accuracy for HR with a low light balance score(0.867-0.983)was greater than that with a high score(0.667-0.750).Our method had higher accuracy in estimating HR than traditional filtering methods(0.167-0.417)and state-of-the-art neural network filtering methods(0.783-0.917)did.The root mean square error of the HRV from the time domain was the lowest,and the correlation index score was the highest for the HRV from the frequency domain estimated by our method compared with those estimated by two neural networks.Conclusions:Light balance,large sample training,and two-stage training can improve the accuracy of HRV estimation.
基金Supported by Beijing Natural Science Foundation(Grant No.L241012)the National Natural Science Foundation of China(Grant No.62572468).
文摘LiDAR and camera are two of the most common sensors used in the fields of robot perception,autonomous driving,augmented reality,and virtual reality,where these sensors are widely used to perform various tasks such as odometry estimation and 3D reconstruction.Fusing the information from these two sensors can significantly increase the robustness and accuracy of these perception tasks.The extrinsic calibration between cameras and LiDAR is a fundamental prerequisite for multimodal systems.Recently,extensive studies have been conducted on the calibration of extrinsic parameters.Although several calibration methods facilitate sensor fusion,a comprehensive summary for researchers and,especially,non-expert users is lacking.Thus,we present an overview of extrinsic calibration and discuss diverse calibration methods from the perspective of calibration system design.Based on the calibration information sources,this study classifies these methods as target-based or targetless.For each type of calibration method,further classification was performed according to the diverse types of features or constraints used in the calibration process,and their detailed implementations and key characteristics were introduced.Thereafter,calibration-accuracy evaluation methods are presented.Finally,we comprehensively compare the advantages and disadvantages of each calibration method and suggest directions for practical applications and future research.
文摘介绍一种应用于USB video camera中的自动对焦系统。由USB video camera获取的视频图像经计算机进行FFT运算或微分运算,得到其频谱幅值数据或微分幅值数据,计算机根据所得数据判断USB video camera中的镜头是否处于离焦位置并控制电机将镜头移到对焦位置。文章还进一步讨论了提高自动对焦准确度的措施。实验结果表明该自动对焦系统能很好地实现USB video camera的自动对焦,该系统将使具有USB接口的video camera使用更简单方便。
基金financial supports from National Natural Science Foundation of China(Grant Nos.U23A20368 and 62175006)Academic Excellence Foundation of BUAA for PhD Students.
文摘Due to the limitations of spatial bandwidth product and data transmission bandwidth,the field of view,resolution,and imaging speed constrain each other in an optical imaging system.Here,a fast-zoom and high-resolution sparse compound-eye camera(CEC)based on dual-end collaborative optimization is proposed,which provides a cost-effective way to break through the trade-off among the field of view,resolution,and imaging speed.In the optical end,a sparse CEC based on liquid lenses is designed,which can realize large-field-of-view imaging in real time,and fast zooming within 5 ms.In the computational end,a disturbed degradation model driven super-resolution network(DDMDSR-Net)is proposed to deal with complex image degradation issues in actual imaging situations,achieving high-robustness and high-fidelity resolution enhancement.Based on the proposed dual-end collaborative optimization framework,the angular resolution of the CEC can be enhanced from 71.6"to 26.0",which provides a solution to realize high-resolution imaging for array camera dispensing with high optical hardware complexity and data transmission bandwidth.Experiments verify the advantages of the CEC based on dual-end collaborative optimization in high-fidelity reconstruction of real scene images,kilometer-level long-distance detection,and dynamic imaging and precise recognition of targets of interest.