期刊文献+

回归——聚类联合框架下的手写文本行提取

Combination of regression and clustering for handwritten text line extraction
原文传递
导出
摘要 目的手写文本行提取是文档图像处理中的重要基础步骤,对于无约束手写文本图像,文本行都会有不同程度的倾斜、弯曲、交叉、粘连等问题。利用传统的几何分割或聚类的方法往往无法保证文本行边缘的精确分割。针对这些问题提出一种基于文本行回归-聚类联合框架的手写文本行提取方法。方法首先,采用各向异性高斯滤波器组对图像进行多尺度、多方向分析,利用拖尾效应检测脊形结构提取文本行主体区域,并对其骨架化得到文本行回归模型。然后,以连通域为基本图像单元建立超像素表示,为实现超像素的聚类,建立了像素-超像素-文本行关联层级随机场模型,利用能量函数优化的方法实现超像素的聚类与所属文本行标注。在此基础上,检测出所有的行间粘连字符块,采用基于回归线的k-means聚类算法由回归模型引导粘连字符像素聚类,实现粘连字符分割与所属文本行标注。最后,利用文本行标签开关实现了文本行像素的操控显示与定向提取,而不再需要几何分割。结果在HIT-MW脱机手写中文文档数据集上进行文本行提取测试,检测率DR为99.83%,识别准确率RA为99.92%。结论实验表明,提出的文本行回归-聚类联合分析框架相比于传统的分段投影分析、最小生成树聚类、Seam Carving等方法提高了文本行边缘的可控性与分割精度。在高效手写文本行提取的同时,最大程度地避免了相邻文本行的干扰,具有较高的准确率和鲁棒性。 Objective Handwritten text line extraction is fundamental in document image processing. The text lines may suffer from tilting curving crossing and adhesion because of unconstrained paper layout and free writing style. Traditional text line segmentation or clustering method cannot guarantee the classification accuracy of the pixels between text lines. In this study,a text line regression-clustering joint framework for handwritten text line extraction is proposed. Method First,the anisotropic Gaussian filter bank is used to filter the handwritten document image in multiple scales and directions. The main body area(MBA) of text line is first extracted by smearing,andthe text line regression model is then obtained by extracting the skeleton structure of the MBA. Second,the super-pixel representation is constructed with connected component as the basic image element. For super-pixel classification and clustering,an approach based on associative hierarchical random fields is presented. A higher-order energy model is established by constructing a hierarchical network of pixel-connected component text lines. On the basis of the model,an energy function is built whose minimization yields the text line labels of the connected components. With the achieved instance labels of connected components as basis,the sticky characters that share the same label are detected. Third,the pixels of the sticky characters are re-clustered with k-means algorithm under the constraint of the text line regression model. With the instance labels of text lines,the manipulation of the text lines can be achieved by label switch. Therefore,the geometric segmentation of the document image is no longer needed,and the bounding box can be used to extract text line directly. Result Experiments were performed on HIT-MW document level dataset. The proposed framework achieved an overall detection rate of 99. 83% and recognition accuracy of99. 92% which reach to the state-of-the-art performance for Chinese handwritten text line extraction. Conclusion E
作者 朱健菲 应自炉 陈鹏飞 Zhu Jianfei;Ying Zilu;Chen Pengfei(School of lnformation Engineering,Wuyi University,Jiangmen 529020,China)
出处 《中国图象图形学报》 CSCD 北大核心 2018年第8期1207-1217,共11页 Journal of Image and Graphics
基金 国家自然科学基金项目(61771347 61372193)~~
关键词 手写文本行提取 超像素 图像分割 回归 聚类 handwritten text line extraction superpixel image segmentation regression clustering
  • 相关文献

参考文献2

二级参考文献162

  • 1苏金玲,王朝晖.基于Graph Cut和超像素的自然场景显著对象分割方法[J].苏州大学学报(自然科学版),2012,28(2):27-33. 被引量:7
  • 2Ren X, Malik J. Learning a classification model for segmentation [ C]//Proceedings of the IEEE International Conference on Com- puter Vision. Washington DC, USA: IEEE, 2003: 10-17. [ DOI: 10. 1109/ICCV. 2003. 1238308 ]. 被引量:1
  • 3Achanta R, Shaji A, Smith K, et al. SLIC superpixels compared to state-of-the-art superpixel methods[ J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, 34 ( 11 ) : 2274-2282. [DOI: 10. 1109/TPAMI. 2012. 120]. 被引量:1
  • 4Xu C, Corso J J. Evaluation of super-voxel methods for early vid- eo processing[ C ]//Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Washington DC, USA: IEEE, 2012 : 1202-1209. [DOI : 10. 1109/CVPR. 2012. 6247802 ]. 被引量:1
  • 5Shi J, Malik ./. Normalized cuts and image segmentation [ J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000, 22(8): 888-905. [DOI: 10. 1109/34. 868688]. 被引量:1
  • 6Moore A P, Prince S, Warrell J, et al. Superpixel lattices[ C]// Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Washington DC, USA : IEEE, 2008 : 1-8. [ DOI: 10. 1109/CVPR. 2008. 4587471 ]. 被引量:1
  • 7Veksler O, Boykov Y, Mehrani P. Superpixels and supervoxels in an energy optimization framework [ M ]//Computer Vision-EC- CV 2010. Berlin Heidelberg: Springer, 2010: 211-224. [DOI: 10. 1007/978-3-642-15555-0_16 ]. 被引量:1
  • 8Achanta R, Shaji A, Smith K, et al. Slic superpixels[ R]. Lau- sanne, Vaud, Switzerland: Swiss federal Institute of Technology, 2010. 被引量:1
  • 9Liu M Y, Tuzel O, Ramalingam S, et al. Entropy rate superpixel segmentation[ C ]//Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Washington DC, USA : IEEE, 2011 : 209/-2104. [DOI: 10. 1109/CVPR.2011. 5995323]. 被引量:1
  • 10Zhang Y, Hartley R, Mashford J, et al. Superpixels via pseudo-boolean optimization [ C ]//Proceedings of IEEE International Conference on Computer Vision. Washington DC, USA: IEEE, 2011 : 1387-1394. [DOI : 10. 1109/ICCV. 2011. 6126393 ]. 被引量:1

共引文献119

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部