This paper proposes a novel mapping scheme for bit-interleaved coded modulation with iterative decoding(BICM-ID).The symbol mapping is composed of two QPSK with different radiuses and phases,called cross equalization-...This paper proposes a novel mapping scheme for bit-interleaved coded modulation with iterative decoding(BICM-ID).The symbol mapping is composed of two QPSK with different radiuses and phases,called cross equalization-8PSK-quasi-semi set partitioning(CE-8PSK-Quasi-SSP).Providing the same average power,the proposed scheme can increase the minimum squared Euclidean distance(MSED)and then improve the receiving performance of BICM-ID compared with conventional symbol mapping schemes.Simultaneously,a modified iteration decoding algorithm is proposed in this paper.In the process of iteration decoding,different proportion of the extrinsic information to the systematic observations results in distinct decoding performance.At high SNR(4~9dB),the observation information plays a more important role than the extrinsic information.Simulation results show that the proportion set at 1.2 is more suitable for the novel mapping in BICM-ID.When the BER is 10^(-4),more than 0.9dB coding gain over Rayleigh channels can be achieved for the improved mapping and decoding scheme.展开更多
This contribution proposes a new combination symbol mapper/8-ary constellation, which is a joint optimization of an 8-ary signal constellation and its symbol mapping operation, to improve the performance of Bit Interl...This contribution proposes a new combination symbol mapper/8-ary constellation, which is a joint optimization of an 8-ary signal constellation and its symbol mapping operation, to improve the performance of Bit Interleaved Coded Modulation with Iterative Decoding (BICM-ID). The basic idea was to use the so called (1,7) constellation (which is a capacitive efficient constellation) instead of the conventional 8-PSK constellation and to choose the most suitable mapping for it. A comparative study between the combinations most suitable mapping/(1,7) constellation and SSP mapping/conventional 8-PSK constellation has been carried out. Simulation results showed that the 1st combination significantly outperforms the 2nd combination and with only 4 iterations, it gives better performance than the 2nd combination with 8 iterations. A gain of 4 dB is given by iteration 4 of the 1st combination compared to iteration 8 of the 2nd combination at a BER level equal to 10-5, and it (iteration 4 of the 1st combination) can attain a BER equal to 10-7 for, only, a SNR = 5.6 dB.展开更多
The process of generating descriptive captions for images has witnessed significant advancements in last years,owing to the progress in deep learning techniques.Despite significant advancements,the task of thoroughly ...The process of generating descriptive captions for images has witnessed significant advancements in last years,owing to the progress in deep learning techniques.Despite significant advancements,the task of thoroughly grasping image content and producing coherent,contextually relevant captions continues to pose a substantial challenge.In this paper,we introduce a novel multimodal method for image captioning by integrating three powerful deep learning architectures:YOLOv8(You Only Look Once)for robust object detection,EfficientNetB7 for efficient feature extraction,and Transformers for effective sequence modeling.Our proposed model combines the strengths of YOLOv8 in detecting objects,the superior feature representation capabilities of EfficientNetB7,and the contextual understanding and sequential generation abilities of Transformers.We conduct extensive experiments on standard benchmark datasets to evaluate the effectiveness of our approach,demonstrating its ability to generate informative and semantically rich captions for diverse images.The experimental results showcase the synergistic benefits of integrating YOLOv8,EfficientNetB7,and Transformers in advancing the state-of-the-art in image captioning tasks.The proposed multimodal approach has yielded impressive outcomes,generating informative and semantically rich captions for a diverse range of images.By combining the strengths of YOLOv8,EfficientNetB7,and Transformers,the model has achieved state-of-the-art results in image captioning tasks.The significance of this approach lies in its ability to address the challenging task of generating coherent and contextually relevant captions while achieving a comprehensive understanding of image content.The integration of three powerful deep learning architectures demonstrates the synergistic benefits of multimodal fusion in advancing the state-of-the-art in image captioning.Furthermore,this approach has a profound impact on the field,opening up new avenues for research in multimodal deep learning and paving the way for more sophisticated and context-aware image captioning systems.These systems have the potential to make significant contributions to various fields,encompassing human-computer interaction,computer vision and natural language processing.展开更多
基金Supported by the Key Project of Chinese Ministry of Education(No.106042)the Scientific Research Foundation for the Returned Overseas Chinese Scholars,State Education Ministry(2007[24])
文摘This paper proposes a novel mapping scheme for bit-interleaved coded modulation with iterative decoding(BICM-ID).The symbol mapping is composed of two QPSK with different radiuses and phases,called cross equalization-8PSK-quasi-semi set partitioning(CE-8PSK-Quasi-SSP).Providing the same average power,the proposed scheme can increase the minimum squared Euclidean distance(MSED)and then improve the receiving performance of BICM-ID compared with conventional symbol mapping schemes.Simultaneously,a modified iteration decoding algorithm is proposed in this paper.In the process of iteration decoding,different proportion of the extrinsic information to the systematic observations results in distinct decoding performance.At high SNR(4~9dB),the observation information plays a more important role than the extrinsic information.Simulation results show that the proportion set at 1.2 is more suitable for the novel mapping in BICM-ID.When the BER is 10^(-4),more than 0.9dB coding gain over Rayleigh channels can be achieved for the improved mapping and decoding scheme.
文摘This contribution proposes a new combination symbol mapper/8-ary constellation, which is a joint optimization of an 8-ary signal constellation and its symbol mapping operation, to improve the performance of Bit Interleaved Coded Modulation with Iterative Decoding (BICM-ID). The basic idea was to use the so called (1,7) constellation (which is a capacitive efficient constellation) instead of the conventional 8-PSK constellation and to choose the most suitable mapping for it. A comparative study between the combinations most suitable mapping/(1,7) constellation and SSP mapping/conventional 8-PSK constellation has been carried out. Simulation results showed that the 1st combination significantly outperforms the 2nd combination and with only 4 iterations, it gives better performance than the 2nd combination with 8 iterations. A gain of 4 dB is given by iteration 4 of the 1st combination compared to iteration 8 of the 2nd combination at a BER level equal to 10-5, and it (iteration 4 of the 1st combination) can attain a BER equal to 10-7 for, only, a SNR = 5.6 dB.
基金funded by Researchers Supporting Project number(RSPD2024R698),King Saud University,Riyadh,Saudi Arabia.
文摘The process of generating descriptive captions for images has witnessed significant advancements in last years,owing to the progress in deep learning techniques.Despite significant advancements,the task of thoroughly grasping image content and producing coherent,contextually relevant captions continues to pose a substantial challenge.In this paper,we introduce a novel multimodal method for image captioning by integrating three powerful deep learning architectures:YOLOv8(You Only Look Once)for robust object detection,EfficientNetB7 for efficient feature extraction,and Transformers for effective sequence modeling.Our proposed model combines the strengths of YOLOv8 in detecting objects,the superior feature representation capabilities of EfficientNetB7,and the contextual understanding and sequential generation abilities of Transformers.We conduct extensive experiments on standard benchmark datasets to evaluate the effectiveness of our approach,demonstrating its ability to generate informative and semantically rich captions for diverse images.The experimental results showcase the synergistic benefits of integrating YOLOv8,EfficientNetB7,and Transformers in advancing the state-of-the-art in image captioning tasks.The proposed multimodal approach has yielded impressive outcomes,generating informative and semantically rich captions for a diverse range of images.By combining the strengths of YOLOv8,EfficientNetB7,and Transformers,the model has achieved state-of-the-art results in image captioning tasks.The significance of this approach lies in its ability to address the challenging task of generating coherent and contextually relevant captions while achieving a comprehensive understanding of image content.The integration of three powerful deep learning architectures demonstrates the synergistic benefits of multimodal fusion in advancing the state-of-the-art in image captioning.Furthermore,this approach has a profound impact on the field,opening up new avenues for research in multimodal deep learning and paving the way for more sophisticated and context-aware image captioning systems.These systems have the potential to make significant contributions to various fields,encompassing human-computer interaction,computer vision and natural language processing.