Query:
学者姓名:郑明魁
Refining:
Year
Type
Indexed by
Source
Complex
Former Name
Co-
Language
Clean All
Abstract :
针对当前视频预测算法在生成视频帧时细节模糊、精度较低的问题,提出了一种基于边缘增强和多尺度时空重组的视频预测方法.首先通过频域分离技术,将视频帧划分为高频信息和低频信息,并对二者分别进行针对性处理.其次,设计了高频边缘增强模块,专注于高频边缘特征的学习与优化.同时,引入多尺度时空重组模块,针对低频结构信息,深入挖掘其时空依赖性.最终将高低频特征进行充分融合,用以生成高质量的预测视频帧.实验结果表明,与现有先进算法相比,该方法在预测性能上实现了提升,充分验证了其有效性.
Keyword :
多尺度时空重组 多尺度时空重组 视频预测 视频预测 边缘增强 边缘增强 频域分离 频域分离
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 吴孔贤 , 郑明魁 . 基于边缘增强和多尺度时空重组的视频预测方法 [J]. | 网络安全与数据治理 , 2025 , 44 (3) : 22-26 . |
MLA | 吴孔贤 等. "基于边缘增强和多尺度时空重组的视频预测方法" . | 网络安全与数据治理 44 . 3 (2025) : 22-26 . |
APA | 吴孔贤 , 郑明魁 . 基于边缘增强和多尺度时空重组的视频预测方法 . | 网络安全与数据治理 , 2025 , 44 (3) , 22-26 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
针对传统的激光SLAM算法在室外动态场景下定位精度低和缺少语义信息等问题,本文设计了一种基于语义信息融合的激光SLAM改进算法,并在公开数据集KITTI上进行测试实验,为提升整体位姿估计精度和建图精度提供有益参考.
Keyword :
LeGO-LOAM LeGO-LOAM 深度学习 深度学习 激光SLAM 激光SLAM 语义分割 语义分割 语义约束 语义约束
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 王占宝 , 郑明魁 . 融合语义信息的激光SLAM研究 [J]. | 广播电视网络 , 2024 , 31 (5) : 28-30 . |
MLA | 王占宝 等. "融合语义信息的激光SLAM研究" . | 广播电视网络 31 . 5 (2024) : 28-30 . |
APA | 王占宝 , 郑明魁 . 融合语义信息的激光SLAM研究 . | 广播电视网络 , 2024 , 31 (5) , 28-30 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Sampling is a crucial concern for outdoor light detection and ranging (LiDAR) point cloud registration due to the large amounts of point cloud. Numerous algorithms have been devised to tackle this issue by selecting key points. However, these approaches often necessitate extensive computations, giving rise to challenges related to computational time and complexity. This letter proposes a multi-domain uniform sampling method (MDU-sampling) for large-scale outdoor LiDAR point cloud registration. The feature extraction based on deep learning aggregates information from the neighbourhood, so there is redundancy between adjacent features. The sampling method in this paper is carried out in the spatial and feature domains. First, uniform sampling is executed in the spatial domain, maintaining local point cloud uniformity. This is believed to preserve more potential point correspondences and is beneficial for subsequent neighbourhood information aggregation and feature sampling. Subsequently, a secondary sampling in the feature domain is performed to reduce redundancy among the features of neighbouring points. Notably, only points on the same ring in LiDAR data are considered as neighbouring points, eliminating the need for additional neighbouring point search and thereby speeding up processing rates. Experimental results demonstrate that the approach enhances accuracy and robustness compared with benchmarks. The feature extraction based on deep learning aggregates information from the neighbourhood, so there is redundancy between adjacent features. The sampling method in this paper is carried out in the spatial and feature domains, reducing the computational resources for registration. The proposed method preserves more effective information compared to other algorithms. Points are only considered on the same ring in LiDAR data as neighbouring points, eliminating the need for additional neighbouring point search. This makes it efficient and suitable for large-scale outdoor LiDAR point cloud registration. image
Keyword :
artificial intelligence artificial intelligence robot vision robot vision signal processing signal processing SLAM (robots) SLAM (robots)
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Ou, Wengjun , Zheng, Mingkui , Zheng, Haifeng . MDU-sampling: Multi-domain uniform sampling method for large-scale outdoor LiDAR point cloud registration [J]. | ELECTRONICS LETTERS , 2024 , 60 (5) . |
MLA | Ou, Wengjun 等. "MDU-sampling: Multi-domain uniform sampling method for large-scale outdoor LiDAR point cloud registration" . | ELECTRONICS LETTERS 60 . 5 (2024) . |
APA | Ou, Wengjun , Zheng, Mingkui , Zheng, Haifeng . MDU-sampling: Multi-domain uniform sampling method for large-scale outdoor LiDAR point cloud registration . | ELECTRONICS LETTERS , 2024 , 60 (5) . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
无线电频谱监测海量数据存储和分析是无线电监管工作的重要组成部分.频谱数据具有时间相关性以及不同频点间的相关冗余,对此本文设计了一种基于类小波变换的无线电频谱监测数据无损压缩方法.该方法首先基于时间相关性将一维频谱信号转换成二维矩阵;转换成二维矩阵后数据在水平方向以及垂直方向都存在冗余,算法采用卷积神经网络来代替传统小波中的预测和更新模块,并引入了自适应压缩块来处理不同维度的特征,从而获得更紧凑的频谱数据表示.研究进一步设计了一种基于上下文的深度熵模型,利用类小波变换不同子带系数获得熵编码参数,以此估计累积概率,从而实现频谱数据的压缩.实验结果表明,与已有的Deflate等传统频谱监测数据无损压缩方法相比,本文算法有进一步的性能提升,与典型的JPEG2000、PNG、JPEG-LS等二维图像无损压缩方法相比,本文所提出的方法的压缩效果也提高了20%以上.
Keyword :
卷积神经网络 卷积神经网络 无损压缩 无损压缩 熵编码 熵编码 类小波变换 类小波变换 频谱监测数据 频谱监测数据
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 张承琰 , 郑明魁 , 刘会明 et al. 一种基于类小波变换的无线电频谱监测数据无损压缩方法 [J]. | 电子测量与仪器学报 , 2024 , 38 (7) : 152-158 . |
MLA | 张承琰 et al. "一种基于类小波变换的无线电频谱监测数据无损压缩方法" . | 电子测量与仪器学报 38 . 7 (2024) : 152-158 . |
APA | 张承琰 , 郑明魁 , 刘会明 , 易天儒 , 李少良 , 陈祖儿 . 一种基于类小波变换的无线电频谱监测数据无损压缩方法 . | 电子测量与仪器学报 , 2024 , 38 (7) , 152-158 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
The comprehension of 3D semantic scenes holds paramount significance in autonomous driving and robotics technology. Nevertheless, the simultaneous achievement of real-time processing and high precision in complex, expansive outdoor environments poses a formidable challenge. In response to this challenge, we propose a novel occupancy network named RTONet, which is built on a teacher-student model. To enhance the ability of the network to recognize various objects, the decoder incorporates dilated convolution layers with different receptive fields and utilizes a multi-path structure. Furthermore, we develop an automatic frame selection algorithm to augment the guidance capability of the teacher network. The proposed method outperforms the existing grid-based approaches in semantic completion (mIoU), and achieves the state-of-the-art performance in terms of real-time inference speed while exhibiting competitive performance in scene completion (IoU) on the SemanticKITTI benchmark.
Keyword :
Decoding Decoding deep learning for visual perception deep learning for visual perception Feature extraction Feature extraction Laser radar Laser radar LiDAR LiDAR mapping mapping occupancy grid occupancy grid Point cloud compression Point cloud compression Real-time systems Real-time systems Semantics Semantics Semantic scene understanding Semantic scene understanding Three-dimensional displays Three-dimensional displays
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Lai, Quan , Zheng, Haifeng , Feng, Xinxin et al. RTONet: Real-Time Occupancy Network for Semantic Scene Completion [J]. | IEEE ROBOTICS AND AUTOMATION LETTERS , 2024 , 9 (10) : 8370-8377 . |
MLA | Lai, Quan et al. "RTONet: Real-Time Occupancy Network for Semantic Scene Completion" . | IEEE ROBOTICS AND AUTOMATION LETTERS 9 . 10 (2024) : 8370-8377 . |
APA | Lai, Quan , Zheng, Haifeng , Feng, Xinxin , Zheng, Mingkui , Chen, Huacong , Chen, Wenqiang . RTONet: Real-Time Occupancy Network for Semantic Scene Completion . | IEEE ROBOTICS AND AUTOMATION LETTERS , 2024 , 9 (10) , 8370-8377 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
For moving cameras, the video content changes significantly, which leads to inaccurate prediction in traditional inter prediction and results in limited compression efficiency. To solve these problems, first, we propose a camera pose-based background modeling (CP-BM) framework that uses the camera motion and the textures of reconstructed frames to model the background of the current frame. Compared with the reconstructed frames, the predicted background frame generated by CP-BM is more geometrically similar to the current frame in position and is more strongly correlated with it at the pixel level; thus, it can serve as a higher-quality reference for inter prediction, and the compression efficiency can be improved. Second, to compensate the motion of the background pixels, we construct a pixel-level motion vector field that can accurately describe various complex motions with only a small overhead. Our method is more general than other motion models because it has more degrees of freedom, and when the degrees of freedom are decreased, it encompasses other motion models as special cases. Third, we propose an optical flow-based depth estimation (OF-DE) method to synchronize the depth information at the codec, which is used to build the motion vector field. Finally, we integrate the overall scheme into the High Efficiency Video Coding (HEVC) and Versatile Video Coding (VVC) reference software HM-16.7 and VTM-10.0. Experimental results demonstrate that in HM-16.7, for in-vehicle video sequences, our solution has an average Bj & oslash;ntegaard delta bit rate (BD-rate) gain of 8.02% and reduces the encoding time by 20.9% due to the superiority of our scheme in motion estimation. Moreover, in VTM-10.0 with affine motion compensation (MC) turned off and turned on, our method has average BD-rate gains of 5.68% and 0.56%, respectively.
Keyword :
background modeling background modeling Bit rate Bit rate camera pose camera pose Cameras Cameras Computational modeling Computational modeling Encoding Encoding Estimation Estimation moving cameras moving cameras Predictive models Predictive models Video coding Video coding
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Fang, Zheng , Zheng, Mingkui , Chen, Pingping et al. Camera Pose-Based Background Modeling for Video Coding in Moving Cameras [J]. | IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY , 2024 , 34 (5) : 4054-4069 . |
MLA | Fang, Zheng et al. "Camera Pose-Based Background Modeling for Video Coding in Moving Cameras" . | IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 34 . 5 (2024) : 4054-4069 . |
APA | Fang, Zheng , Zheng, Mingkui , Chen, Pingping , Chen, Zhifeng , Oliver Wu, Dapeng . Camera Pose-Based Background Modeling for Video Coding in Moving Cameras . | IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY , 2024 , 34 (5) , 4054-4069 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Depth estimation from a single image is a fundamental problem in the field of computer vision. With the great success of deep learning techniques, various self-supervised monocular depth estimation methods using encoder-decoder architectures have emerged. However, most previous approaches regress the depth map directly using a single encoder-decoder structure, which may not obtain sufficient features in the image and results in a depth map with low accuracy and blurred details. To improve the accuracy of self-supervised monocular depth estimation, we propose a simple but very effective scheme for depth estimation using a dual encoder-decoder structure network. Specifically, we introduce a novel global feature extraction network (GFN) to extract global features from images. GFN includes PoolAttentionFormer and ResBlock, which work together to extract and fuse hierarchical global features into the depth estimation network (DEN). To further improve the accuracy, we design two feature fusion mechanisms, including global feature fusion and multiscale fusion. The experimental results of various dual encoder-decoder combination schemes tested on the KITTI dataset show that our proposed one is effective in improving the accuracy of self-supervised monocular depth estimation, which reached 89.6% (delta < 1.25).
Keyword :
Accuracy Accuracy Convolutional neural networks Convolutional neural networks Data mining Data mining Decoding Decoding dual encoder-decoder dual encoder-decoder Estimation Estimation Feature extraction Feature extraction Fuses Fuses global information global information monocular depth estimation monocular depth estimation self-supervised self-supervised Training Training
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Zheng, Mingkui , Luo, Lin , Zheng, Haifeng et al. A Dual Encoder-Decoder Network for Self-Supervised Monocular Depth Estimation [J]. | IEEE SENSORS JOURNAL , 2023 , 23 (17) : 19747-19756 . |
MLA | Zheng, Mingkui et al. "A Dual Encoder-Decoder Network for Self-Supervised Monocular Depth Estimation" . | IEEE SENSORS JOURNAL 23 . 17 (2023) : 19747-19756 . |
APA | Zheng, Mingkui , Luo, Lin , Zheng, Haifeng , Ye, Zhangfan , Su, Zhe . A Dual Encoder-Decoder Network for Self-Supervised Monocular Depth Estimation . | IEEE SENSORS JOURNAL , 2023 , 23 (17) , 19747-19756 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Video based point cloud compression (V-PCC) provides an efficient solution for compressing dynamic point clouds, but the projection of V-PCC from 3D to 2D destroys the correlation of 3D inter-frame motion and reduces the performance of inter-frame coding. To solve this problem, we proposes an adaptive segmentation based multi-mode inter-frame coding method for video point cloud to improve V-PCC, and designs a new dynamic point cloud inter-frame encoding framework. Firstly, in order to achieve more accurate block prediction, a block matching method based on adaptive regional segmentation is proposed to find the best matching block; Secondly, in order to further improve the performance of inter coding, a multi-mode inter-frame coding method based on joint attribute rate distortion optimization (RDO) is proposed to increase the prediction accuracy and reduce the bit rate consumption. Experimental results show that the improved algorithm proposed in this paper achieves -22.57% Bjontegaard delta bit rate (BD-BR) gain compared with V-PCC. The algorithm is especially suitable for dynamic point cloud scenes with little change between frames, such as video surveillance and video conference. © 2023 Science Press. All rights reserved.
Keyword :
Electric distortion Electric distortion Image coding Image coding Image compression Image compression Security systems Security systems Signal distortion Signal distortion Video signal processing Video signal processing
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Chen, Jian , Liao, Yan-Jun , Wang, Kuo et al. An Adaptive Segmentation Based Multi-mode Inter-frame Coding Method for Video Point Cloud [J]. | Acta Automatica Sinica , 2023 , 49 (8) : 1707-1722 . |
MLA | Chen, Jian et al. "An Adaptive Segmentation Based Multi-mode Inter-frame Coding Method for Video Point Cloud" . | Acta Automatica Sinica 49 . 8 (2023) : 1707-1722 . |
APA | Chen, Jian , Liao, Yan-Jun , Wang, Kuo , Zheng, Ming-Kui , Su, Li-Chao . An Adaptive Segmentation Based Multi-mode Inter-frame Coding Method for Video Point Cloud . | Acta Automatica Sinica , 2023 , 49 (8) , 1707-1722 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
高效视频编码标准(HEVC/H.265)是目前国际市场上广泛采用的视频编码标准.基于上下文自适应二进制算术编码(CABAC)作为HEVC熵编码的核心编码方式,通过建立更加精准的概率模型,提高了算术编码的压缩效率.此外,HEVC定义了更多种类的语法元素并建立更复杂的编码结构,进一步减少信息冗余,从而降低了码率.然而,语法元素作为CABAC的输入,其预处理过程的高复杂性,增加了硬件并行难度,导致熵编码硬件的吞吐率难以提高,成为HEVC编码器实现更高分辨率实时编码的瓶颈之一.为了进一步加快熵编码模块的速度,本文设计了一种基于现场可编程门阵列(FPGA)的高吞吐量CABAC熵编码架构.该架构提出的预头信息编码、上下文模型初始化和编码单元(CU)结构优化策略,可以加快语法元素的产生,以供自适应二进制算术编码器使用;通过高效的残差编码架构和部分上下文索引流水计算方案,在保持高吞吐量的同时,可以减少由复杂计算带来的路径延迟,提高工作频率.本设计使用90 nm标准单元库进行综合,共使用了2.099×104 个逻辑门数,工作频率可以达到200 MHz.本文对HEVC官方提供的视频序列进行仿真测试,统计了在不同量化参数(QP)下编码1个编码树单元(CTU)所需要的时间,实验统计数据表明,本文设计使得编码1个CTU的时间平均节省了38.2%.
Keyword :
CABAC CABAC FPGA FPGA 吞吐率 吞吐率 熵编码 熵编码 语法元素 语法元素
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 林志坚 , 黄萍 , 郑明魁 et al. 基于FPGA的HEVC熵编码语法元素硬件加速设计 [J]. | 华南理工大学学报(自然科学版) , 2023 , 51 (8) : 110-117 . |
MLA | 林志坚 et al. "基于FPGA的HEVC熵编码语法元素硬件加速设计" . | 华南理工大学学报(自然科学版) 51 . 8 (2023) : 110-117 . |
APA | 林志坚 , 黄萍 , 郑明魁 , 陈平平 . 基于FPGA的HEVC熵编码语法元素硬件加速设计 . | 华南理工大学学报(自然科学版) , 2023 , 51 (8) , 110-117 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
本发明提出一种基于激光雷达的点云实时采集压缩传输系统及方法,包括:实时采集激光雷达点云,对点云进行自适应编码和封装,实时传输,解封装和自适应解码,渲染可视化并保存本地。本系统具有时间复杂度低,实时性高的优点,根据带宽动态压缩后的数据在低带宽的情况下也可实现可靠低时延的传输,远程实时地观测并处理激光雷达采集的第一手3D点云数据。高带宽情况下该系统还可用于传输多路数据,符合车路协同、远程智能驾驶、机器人视觉等行业对远程采集传输点云数据并进行分析处理的低时延需求。
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 陈建 , 黄炜 , 陈锋 et al. 基于激光雷达的点云实时采集压缩传输系统及方法 : CN202111074168.3[P]. | 2021-09-14 00:00:00 . |
MLA | 陈建 et al. "基于激光雷达的点云实时采集压缩传输系统及方法" : CN202111074168.3. | 2021-09-14 00:00:00 . |
APA | 陈建 , 黄炜 , 陈锋 , 郑明魁 , 黄昕 . 基于激光雷达的点云实时采集压缩传输系统及方法 : CN202111074168.3. | 2021-09-14 00:00:00 . |
Export to | NoteExpress RIS BibTex |
Version :
Export
Results: |
Selected to |
Format: |