利用selenium获取前程无忧招聘信息-2

时间:2020-05-23
本文章向大家介绍利用selenium获取前程无忧招聘信息-2,主要包括利用selenium获取前程无忧招聘信息-2使用实例、应用技巧、基本知识点总结和需要注意事项,具有一定的参考价值,需要的朋友可以参考一下。

目的:

前一篇文章写的也是爬取前程无忧,该篇文章对脚本进行了部分改动,增加了灵活性

1.利用隐式等待,废除time.sleep的使用,节约时间成本

2.添加了用户名,密码及job名称参数,更加灵活

3.增加了下一页判断,如果某个job有多页结果,可以查询到全部信息

代码:

#!/usr/bin/env python
#-*- coding:utf-8 -*-


"""
目的:

从前程无忧网站上提取指定工作的详细信息
"""


import time
from selenium import webdriver
import requests
from bs4 import BeautifulSoup
from getpass import getpass



def get_soup(url):
    headers = {
        "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0"
    }
    try:
        response = requests.get(url, headers=headers)
        if response.status_code == 200:
            # response.apparent_encoding = "utf-8"
            html = response.content
            soup = BeautifulSoup(html, 'html.parser')
    except:
        print("爬取失败")
    return soup


def get_content(soup):
    content = soup.find("div", class_="bmsg job_msg inbox").text
    # print(content)
    return content.strip()
    

def final_result(url):
    soup = get_soup(url)
    result = get_content(soup)
    return result


def next_page():
    print("\033[32m开始搜索下一页\033[0m")
    try:
       next_page = driver.find_element_by_xpath("/html/body/div[2]/div[4]/div[55]/div/div/div/ul/li[8]/a")
       next_page.click()
       return True
    except Exception as e:
        print("这是最后一页")
        return False

def get_job_info():
        # 获取内容
    all_list = driver.find_element_by_id("resultList").find_elements_by_class_name("el")

    for item in all_list[1:]:
        jobname = item.find_element_by_class_name("t1").find_element_by_tag_name("a").text
        jobhref = item.find_element_by_class_name("t1").find_element_by_tag_name("a").get_attribute("href") 
        jobskill = final_result(jobhref)
            
        company = item.find_element_by_class_name("t2").text
        site = item.find_element_by_class_name("t3").text
        money = item.find_element_by_class_name("t4").text
        date = item.find_element_by_class_name("t5").text 
        
        out_dict = {
            "职位": jobname,
            "职位要求": jobskill,
            "职位详细信息": jobhref,
            "招聘公司": company,
            "工作地点": site,
            "薪水": money,
            "发布日期": date
        }
        
        print(out_dict)




if __name__ == "__main__":
    
    user = input("user: ").strip()
    passwd = getpass()
    jobname = input("jobname: ")

    # 登陆
    driver = webdriver.Firefox()
    driver.get("https://login.51job.com/login.php?lang=c") 
    driver.implicitly_wait(20)   # 设置隐式等待, 这样无需到处充满着time.sleep

    driver.find_element_by_id("loginname").send_keys(user)
    driver.find_element_by_id("password").send_keys(passwd)
    driver.find_element_by_id("login_btn").click()
    print("登陆中,请稍等。。。")

    # 返回首页
    print("进入首页。。。")
    driver.find_element_by_xpath("/html/body/div[1]/div[4]/div/p/a[1]").click()
    # 添加多个地区
    driver.find_element_by_xpath("/html/body/div[3]/div/div[1]/div/div/p[2]/em").click()
    driver.find_element_by_xpath('//*[@id="work_position_click_center_right_list_category_000000_180200"]').click()
    driver.find_element_by_xpath('//*[@id="work_position_click_bottom_save"]').click()

    # 搜索job
    print("搜索职位。。")
    driver.find_element_by_id("kwdselectid").send_keys(jobname)
    driver.find_element_by_xpath("/html/body/div[3]/div/div[1]/div/button").click()
    
    get_job_info()
    
    while next_page():
        get_job_info()

原文地址:https://www.cnblogs.com/lmt921108/p/12944026.html