摘要
在互联网时代,各类新媒体平台出现使得信息数据得到广泛传播。为加强对新浪微博内容的监管和分析,应对舆情分析的需求,该文主要研究采用Python语言设计新浪微博爬虫程序,在对网络爬虫基本概念和原理研究的基础上,设计了具有配置、爬取、存储、分析这4个功能模块的应用程序,为媒体内容监管和数据分析提供了技术支持。
In the Internet era, the emergence of various new media platforms makes information and data widely disseminated. In order to strengthen the management of Sina Weibo information and meet the needs of public opinion analysis, this paper focuses on the design of Sina Weibo crawler program in Python language. Based on the research on the basic concept and principle of web crawler, an application program with four functional modules of configuration, crawling, storage and analysis is designed. It provides technical support for media content management and data analysis.
作者
孙握瑜
SUN Woyu(School of Information and Artificial Intelligence,Anhui Business College of Vocational Technology,Wuhu,Anhui Province,241000 China)
出处
《科技资讯》
2022年第12期34-37,共4页
Science & Technology Information
基金
安徽商贸职业技术学院“双高计划”建设项目(项目编号:2020sgxm08-4)
安徽商贸职业技术学院三平台两基地项目(项目编号:2019ZDX01)。