尚未解决:150页左右收到反爬机制暂停
Crawl_data.py
import requestsimport pandas as pdfrom pprint import pprintfrom lxml import etreeimport timeimport warningswarnings.filterwarnings("ignore")for i in range(1,2001):print("正在爬取第" + str(i) + "页的数据")url_pre = "https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%2595%25B0%25E6%258D%25AE,2,"url_end = ".html?"url = url_pre + str(i) + url_endheaders = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36'}web = requests.get(url, headers=headers)web.encoding = "gbk"dom = etree.HTML(web.text)# 1、岗位名称job_name = dom.xpath('//div[@class="dw_table"]/div[@class="el"]//p/span/a[@target="_blank"]/@title')# 2、公司名称company_name = dom.xpath('//div[@class="dw_table"]/div[@class="el"]/span[@class="t2"]/a[@target="_blank"]/@title')# 3、工作地点address = dom.xpath('//div[@class="dw_table"]/div[@class="el"]/span[@class="t3"]/text()')# 4、工资salary_mid = dom.xpath('//div[@class="dw_table"]/div[@class="el"]/span[@class="t4"]')salary = [i.text for i in salary_mid]# 5、发布日期release_time = dom.xpath('//div[@class="dw_table"]/div[@class="el"]/span[@class="t5"]/text()')# 6、获取二级网址urldeep_url = dom.xpath('//div[@class="dw_table"]/div[@class="el"]//p/span/a[@target="_blank"]/@href')RandomAll = []JobDescribe = []CompanyType = []CompanySize = []Industry = []for i in range(len(deep_url)):web_test = requests.get(deep_url[i], headers=headers)web_test.encoding = "gbk"dom_test = etree.HTML(web_test.text)# 7、爬取经验、学历信息,先合在一个字段里面,以后再做数据清洗。命名为random_allrandom_all = dom_test.xpath('//div[@class="tHeader tHjob"]//div[@class="cn"]/p[@class="msg ltype"]/text()')# 8、岗位描述性息job_describe = dom_test.xpath('//div[@class="tBorderTop_box"]//div[@class="bmsg job_msg inbox"]/p/text()')# 9、公司类型company_type = dom_test.xpath('//div[@class="tCompany_sidebar"]//div[@class="com_tag"]/p[1]/@title')# 10、公司规模(人数)company_size = dom_test.xpath('//div[@class="tCompany_sidebar"]//div[@class="com_tag"]/p[2]/@title')# 11、所属行业(公司)industry = dom_test.xpath('//div[@class="tCompany_sidebar"]//div[@class="com_tag"]/p[3]/@title')#print(industry,company_size,company_type,job_describe,random_all)# 将上述信息保存到各自的列表中RandomAll.append(random_all)JobDescribe.append(job_describe)CompanyType.append(company_type)CompanySize.append(company_size)Industry.append(industry)# 为了反爬,设置睡眠时间time.sleep(1)# 由于我们需要爬取很多页,为了防止最后一次性保存所有数据出现的错误,因此,我们每获取一夜的数据,就进行一次数据存取。df = pd.DataFrame()df["岗位名称"] = job_namedf["公司名称"] = company_namedf["工作地点"] = addressdf["工资"] = salarydf["发布日期"] = release_timedf["经验、学历"] = RandomAlldf["公司类型"] = CompanyTypedf["公司规模"] = CompanySizedf["所属行业"] = Industrydf["岗位描述"] = JobDescribe# 这里在写出过程中,有可能会写入失败,为了解决这个问题,我们使用异常处理。try:df.to_csv("job_info.csv", mode="a+", header=None, index=None, encoding="gbk")except:print("当页数据写入失败")time.sleep(1)print("数据爬取完毕,是不是很开心!!!")
Analysis_data.py
import reimport numpy as npimport pandas as pdimport jiebaimport openpyxlimport warningswarnings.filterwarnings("ignore")df = pd.read_csv(r"job_info.csv",engine="python",header=None,encoding='gbk')# 为数据框指定行索引df.index = range(len(df))# 为数据框指定列索引df.columns = ["岗位名","公司名","工作地点","工资","发布日期","经验与学历","公司类型","公司规模","行业","工作描述"]# 去重之前的记录数print("去重之前的记录数",df.shape)# 记录去重df.drop_duplicates(subset=["公司名","岗位名"],inplace=True)# 去重之后的记录数print("去重之后的记录数",df.shape)df["岗位名"].value_counts()df["岗位名"] = df["岗位名"].apply(lambda x:x.lower())#job_info.shapetarget_job = ['算法', '开发', '分析', '工程师', '数据', '运营', '运维']index = [df["岗位名"].str.count(i) for i in target_job]# 筛选和上面七个职位有关的岗位,本来700多,筛选之后3000多index = np.array(index).sum(axis=0) > 0job_info = df[index]job_info.shapejob_list = ['数据分析', "数据统计","数据专员",'数据挖掘', '算法','大数据','开发工程师', '运营', '软件工程', '前端开发','深度学习', 'ai', '数据库', '数据库', '数据产品','客服', 'java', '.net', 'andrio', '人工智能', 'c++','数据管理',"测试","运维"]job_list = np.array(job_list)def rename(x=None,job_list=job_list):index = [i in x for i in job_list]if sum(index) > 0:return job_list[index][0]else:return xjob_info["岗位名"] = job_info["岗位名"].apply(rename)job_info["岗位名"].value_counts() # 查看岗位名这一列下面有哪些不同的值,并计算每个值有多少个重复值# 数据统计、数据专员、数据分析统一归为数据分析job_info["岗位名"] = job_info["岗位名"].apply(lambda x:re.sub("数据专员","数据分析",x))job_info["岗位名"] = job_info["岗位名"].apply(lambda x:re.sub("数据统计","数据分析",x))job_info["工资"].str[-1].value_counts()job_info["工资"].str[-3].value_counts()index1 = job_info["工资"].str[-1].isin(["年","月"])index2 = job_info["工资"].str[-3].isin(["万","千"])job_info = job_info[index1 & index2]def get_money_max_min(x):try:if x[-3] == "万":z = [float(i)*10000 for i in re.findall("[0-9]+\.?[0-9]*",x)]elif x[-3] == "千":z = [float(i) * 1000 for i in re.findall("[0-9]+\.?[0-9]*", x)]if x[-1] == "年":z = [i/12 for i in z]return zexcept:return xsalary = job_info["工资"].apply(get_money_max_min)job_info["最低工资"] = salary.str[0]job_info["最高工资"] = salary.str[1]job_info["工资水平"] = job_info[["最低工资","最高工资"]].mean(axis=1)#job_info["工作地点"].value_counts()address_list = ['北京', '上海', '广州', '深圳', '杭州', '苏州', '长沙','武汉', '天津', '成都', '西安', '东莞', '合肥', '佛山','宁波', '南京', '重庆', '长春', '郑州', '常州', '福州','沈阳', '济南', '宁波', '厦门', '贵州', '珠海', '青岛','中山', '大连','昆山',"惠州","哈尔滨","昆明","南昌","无锡"]address_list = np.array(address_list)def rename(x=None,address_list=address_list):index = [i in x for i in address_list]if sum(index) > 0:return address_list[index][0]else:return xjob_info["工作地点"] = job_info["工作地点"].apply(rename)job_info.loc[job_info["公司类型"].apply(lambda x:len(x)<6),"公司类型"] = np.nanjob_info["公司类型"] = job_info["公司类型"].str[2:-2]# job_info["行业"].value_counts()job_info["行业"] = job_info["行业"].apply(lambda x:re.sub(",","/",x))job_info.loc[job_info["行业"].apply(lambda x:len(x)<6),"行业"] = np.nanjob_info["行业"] = job_info["行业"].str[2:-2].str.split("/").str[0]#学历字段job_info["学历"] = job_info["经验与学历"].apply(lambda x:re.findall("本科|大专|应届生|在校生|硕士",x))#re.findall返回的是在"经验与学历"这一列里边与"本科|大专|应届生|在校生|硕士"相匹配的全部字符串,返回形式为数组# 最好加r表示raw string,不加会把\n识别成换行,而是\和n# 把找到的学历插入到新建的“学历”中,这是每一行的数据是数组形式def func(x):if len(x) == 0:return np.nanelif len(x) == 1 or len(x) == 2:return x[0]else:return x[2]job_info["学历"] = job_info["学历"].apply(func)#这部分是将数组中的字符串提取并重新插入#实习经验字段job_info["经验"] = job_info["经验与学历"].apply(lambda x:re.findall("1年|2年|3-4年|5-7年|8-9年|10年以上",x))def func2(x):if len(x) == 0:return np.nanelse:return x[0]job_info["经验"] = job_info["经验"].apply(func2)with open(r"stopword.txt","r",encoding='UTF-8') as f:stopword = f.read()stopword = stopword.split()stopword = stopword + ["任职","职位"," "]job_info["工作描述"] = job_info["工作描述"].str[2:-2].apply(lambda x:x.lower()).apply(lambda x:"".join(x))\.apply(jieba.lcut).apply(lambda x:[i for i in x if i not in stopword])job_info.loc[job_info["工作描述"].apply(lambda x:len(x) < 6),"工作描述"] = np.nan#job_info["公司规模"].value_counts()def func(x):if x == "['少于50人']":return "<50"elif x == "['50-150人']":return "50-150"elif x == "['150-500人']":return '150-500'elif x == "['500-1000人']":return '500-1000'elif x == "['1000-5000人']":return '1000-5000'elif x == "['5000-10000人']":return '5000-10000'elif x == "['10000人以上']":return ">10000"else:return np.nanjob_info["公司规模"] = job_info["公司规模"].apply(func)feature = ["公司名","岗位名","工作地点","工资水平","发布日期","学历","经验","公司类型","公司规模","行业"]final_df = job_info[feature]final_df.to_excel(r"词云图.xlsx",encoding="gbk",index=None)job_info.to_excel(r"new_job_info1.xlsx",encoding="gbk",index=None)df = pd.read_excel(r"new_job_info1.xlsx",encoding="gbk")dfdef get_word_cloud(data=None, job_name=None):words = []describe = data['工作描述'][data['岗位名'] == job_name].str[1:-1]describe.dropna(inplace=True)[words.extend(i.split(',')) for i in describe]words = pd.Series(words)word_fre = words.value_counts()return word_frezz = ['数据分析', '算法', '大数据','开发工程师', '运营', '软件工程','运维', '数据库','java',"测试"]for i in zz:word_fre = get_word_cloud(data=df, job_name='{}'.format(i))word_fre = word_fre[1:].reset_index()[:100]word_fre["岗位名"] = pd.Series("{}".format(i),index=range(len(word_fre)))word_fre.to_csv(r"bb.csv", mode='a',index=False, header=None,encoding="gbk")
