seo icon indicating copy to clipboard operation
seo copied to clipboard

脚本失效了

Open amd5 opened this issue 3 years ago • 1 comments

http://seo.chinaz.com/?host={domain_url} 正则匹配的地方,对方修改为了JS渲染

调用接口为 http://seo.chinaz.com/AjaxTool.ashx?action=baidupcrank&callback=jQuery111303238088157608141_1625537181092&host=www.aizhan.com&secretkey=AuzdRj7rz8ERZgDyBTrSvfNP7EyR8Q10khWE|G3afAf3|GGUU3u3Tw==&_=1625537181093

返回值为: jQuery111303238088157608141_1625537181092({"StateCode":1,"Message":"成功","Result":{"rank":7,"uv":0,"kw":114257,"uv_min":45043,"uv_max":71951},"Total":0,"TotalPages":0})

amd5 avatar Jul 06 '21 02:07 amd5

更新一份

import re
import os
import time
import argparse
import requests
from multiprocessing import Pool


def args():
    """
    命令行参数以及说明
    """
    parser = argparse.ArgumentParser()
    parser.add_argument('-r', '--read', dest='read', help='input domains file path')
    parse_args = parser.parse_args()

    # 参数为空 输出--help命令
    if parse_args.read is None:
        parser.print_help()
        os._exit(0)

    return parse_args.read


def seo(domain_url):
    """
    利用爱站接口查询权重信息
    """
    url = f'https://www.aizhan.com/cha/{domain_url}/'
    headers = {
        'Host': 'www.aizhan.com',
        'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:61.0) Gecko/20100101 Firefox/61.0',
        'Content-Type': 'application/x-www-form-urlencoded',
    }
    r = requests.get(url=url, headers=headers, timeout=6)
    html = r.text
    # 百度权重正则
    baidu_pattern = re.compile(r'aizhan.com/images/br/(.*?).png')
    baidu = baidu_pattern.findall(html)[0]
    # 站点标题正则
    site_name_rules = re.compile(r'id="webpage_title">(.*?)</div>')
    site_name = site_name_rules.findall(html)[0]
    print(str(domain_url).ljust(30), '\t', baidu, '\t', site_name)


def main():
    start = time.time()
    file_path = args()
    try:
        #  读取文件所有行
        with open(file_path, "r") as f:
            lines = ''.join(f.readlines()).split("\n")

        print('域名'.ljust(30), '权重\t 站点标题')
        #  设置一个容量为8的进程池
        pool_number = 10
        pool = Pool(pool_number)

        for line in lines:
            if 'http://' in line:
                line = line[7:]
            elif 'https://' in line:
                line = line[8:]
            if line:
                pool.apply_async(seo, (line,))

        pool.close()
        pool.join()

        end = time.time()
        print(f'\n耗时: {end - start:.4f} 秒')
    except Exception as e:
        print('文件读取异常,请检查文件路径是否正确!')
        print(e)


if __name__ == '__main__':
    main()

amd5 avatar Jul 06 '21 02:07 amd5