├── .gitattributes ├── 11-10评论结果 ├── 11-10 │ ├── lianjie.csv │ ├── pictureName.csv │ ├── pinglun.csv │ ├── 口碑评论.xlsx │ ├── 抓取评论详情结果.xlsx │ └── 评论详情.xlsx ├── 2019-11-10 │ ├── 2019-11-10-21_20_19.png │ ├── 2019-11-10-21_20_34.png │ ├── 2019-11-10-21_20_49.png │ ├── 2019-11-10-21_21_04.png │ ├── 2019-11-10-21_21_19.png │ ├── 2019-11-10-21_21_49.png │ ├── 2019-11-10-21_22_27.png │ ├── 2019-11-10-21_22_42.png │ ├── 2019-11-10-21_22_59.png │ ├── 2019-11-10-21_23_14.png │ ├── 2019-11-10-21_23_28.png │ ├── 2019-11-10-21_23_43.png │ ├── 2019-11-10-21_24_09.png │ ├── 2019-11-10-21_24_24.png │ ├── 2019-11-10-21_24_39.png │ ├── 2019-11-10-21_24_54.png │ ├── 2019-11-10-21_25_20.png │ ├── 2019-11-10-21_26_09.png │ ├── 2019-11-10-21_26_36.png │ ├── 2019-11-10-21_26_50.png │ ├── 2019-11-10-21_27_17.png │ ├── 2019-11-10-21_27_43.png │ ├── 2019-11-10-21_27_58.png │ ├── 2019-11-10-21_28_26.png │ ├── 2019-11-10-21_28_41.png │ ├── 2019-11-10-21_29_07.png │ ├── 2019-11-10-21_29_21.png │ ├── 2019-11-10-21_29_36.png │ ├── 2019-11-10-21_29_51.png │ ├── 2019-11-10-21_30_30.png │ ├── 2019-11-10-21_30_57.png │ ├── 2019-11-10-21_31_12.png │ ├── 2019-11-10-21_31_39.png │ ├── 2019-11-10-21_31_54.png │ ├── 2019-11-10-21_32_08.png │ ├── 2019-11-10-21_32_24.png │ ├── 2019-11-10-21_32_50.png │ ├── 2019-11-10-21_33_28.png │ ├── 2019-11-10-21_33_43.png │ ├── 2019-11-10-21_33_57.png │ ├── 2019-11-10-21_34_34.png │ ├── 2019-11-10-21_35_00.png │ ├── 2019-11-10-21_35_13.png │ ├── 2019-11-10-21_35_27.png │ ├── 2019-11-10-21_35_41.png │ ├── 2019-11-10-21_36_07.png │ ├── 2019-11-10-21_36_21.png │ ├── 2019-11-10-21_36_34.png │ ├── 2019-11-10-21_36_48.png │ ├── 2019-11-10-21_37_02.png │ ├── 2019-11-10-21_37_16.png │ ├── 2019-11-10-21_37_30.png │ ├── 2019-11-10-21_38_32.png │ ├── 2019-11-10-21_38_58.png │ ├── 2019-11-10-21_39_12.png │ ├── 2019-11-10-21_39_26.png │ ├── 2019-11-10-21_39_39.png │ ├── 2019-11-10-21_39_54.png │ ├── 2019-11-10-21_40_08.png │ ├── 2019-11-10-21_40_22.png │ ├── 2019-11-10-21_40_37.png │ ├── 2019-11-10-21_40_51.png │ ├── 2019-11-10-21_41_41.png │ ├── 2019-11-10-21_41_56.png │ ├── 2019-11-10-21_42_10.png │ ├── 2019-11-10-21_42_25.png │ ├── 2019-11-10-21_43_27.png │ ├── 2019-11-10-21_43_54.png │ ├── 2019-11-10-21_44_44.png │ ├── 2019-11-10-21_45_10.png │ ├── 2019-11-10-21_45_25.png │ ├── 2019-11-10-21_45_52.png │ ├── 2019-11-10-21_46_07.png │ ├── 2019-11-10-21_46_33.png │ ├── 2019-11-10-21_46_48.png │ ├── 2019-11-10-21_47_03.png │ ├── 2019-11-10-21_47_17.png │ ├── 2019-11-10-21_47_32.png │ ├── 2019-11-10-21_47_47.png │ ├── 2019-11-10-21_48_02.png │ ├── 2019-11-10-21_48_17.png │ ├── 2019-11-10-21_48_32.png │ ├── 2019-11-10-21_48_46.png │ ├── 2019-11-10-21_49_01.png │ ├── 2019-11-10-21_49_40.png │ ├── 2019-11-10-21_49_54.png │ ├── 2019-11-10-21_50_20.png │ ├── 2019-11-10-21_50_37.png │ ├── 2019-11-10-21_50_52.png │ ├── 2019-11-10-21_51_07.png │ ├── 2019-11-10-21_51_47.png │ ├── 2019-11-10-21_52_01.png │ ├── 2019-11-10-21_52_53.png │ ├── 2019-11-10-21_53_19.png │ ├── 2019-11-10-21_53_45.png │ ├── 2019-11-10-21_54_10.png │ ├── 2019-11-10-21_54_24.png │ ├── 2019-11-10-21_54_38.png │ ├── 2019-11-10-21_55_03.png │ ├── 2019-11-10-21_55_40.png │ ├── 2019-11-10-21_56_06.png │ ├── 2019-11-10-21_56_32.png │ ├── 2019-11-10-21_56_46.png │ ├── 2019-11-10-21_56_59.png │ ├── 2019-11-10-21_57_13.png │ ├── 2019-11-10-21_58_14.png │ ├── 2019-11-10-21_58_29.png │ └── 2019-11-10-21_58_56.png ├── Tableprocess.py ├── koubeipinlun.py ├── pictureName.csv ├── pictureandlink.csv ├── pinglun.csv ├── 口碑评论.xlsx ├── 抓取评论详情结果.xlsx └── 评论详情.xlsx ├── 5K壁纸爬虫.py ├── NBA ├── NBA Player.py └── 球员信息.csv ├── README.md ├── 多线程爬取表情包.py ├── 大学排名爬虫.py ├── 安居客二手房.py ├── 王者荣耀英雄信息爬取.py ├── 美桌网图片爬虫.py ├── 表情包爬虫.py ├── 豆瓣电影排名爬虫.py ├── 豆瓣电影爬虫(GUI界面) ├── DoubanMovie.py ├── Movie.py └── main.py └── 链家爬虫(含二手房、新房、租房、小区等爬虫) ├── .idea ├── inspectionProfiles │ └── profiles_settings.xml ├── misc.xml ├── modules.xml ├── workspace.xml └── 链家爬虫.iml ├── 20191030爬取链家上海小区信息 ├── 嘉定 │ ├── jiading_anting.csv │ ├── jiading_fengzhuang.csv │ ├── jiading_huating.csv │ ├── jiading_jiadinglaocheng.csv │ ├── jiading_jiadingxincheng.csv │ ├── jiading_jiangqiao.csv │ ├── jiading_juyuanxinqu.csv │ ├── jiading_malu.csv │ ├── jiading_nanxiang.csv │ ├── jiading_shangda.csv │ ├── jiading_waigang.csv │ ├── jiading_xinchenglu1.csv │ └── jiading_xuxing.csv ├── 奉贤 │ ├── fengxian_fengcheng.csv │ ├── fengxian_fengxianjinhui.csv │ ├── fengxian_haiwan.csv │ ├── fengxian_nanqiao.csv │ ├── fengxian_qingcun.csv │ ├── fengxian_situan.csv │ ├── fengxian_xidu.csv │ ├── fengxian_zhelin.csv │ └── fengxian_zhuanghang.csv ├── 宝山 │ ├── baoshan_dachangzhen.csv │ ├── baoshan_dahua.csv │ ├── baoshan_gongfu.csv │ ├── baoshan_gongkang.csv │ ├── baoshan_gucun.csv │ ├── baoshan_luodian.csv │ ├── baoshan_luojing.csv │ ├── baoshan_songbao.csv │ ├── baoshan_songnan.csv │ ├── baoshan_tonghe.csv │ ├── baoshan_yanghang.csv │ ├── baoshan_yuepu.csv │ └── baoshan_zhangmiao.csv ├── 小区信息 │ ├── 上海各区小区信息汇总.xlsx │ ├── 嘉定小区信息.xlsx │ ├── 奉贤小区信息.xlsx │ ├── 宝山小区信息.xlsx │ ├── 崇明小区信息.xlsx │ ├── 徐家汇小区信息.xlsx │ ├── 普陀小区信息.xlsx │ ├── 杨浦小区信息.xlsx │ ├── 松江小区信息.xlsx │ ├── 浦东小区信息.xlsx │ ├── 虹口小区信息.xlsx │ ├── 金山小区信息.xlsx │ ├── 长宁区小区信息.xlsx │ ├── 闵行小区信息.xlsx │ ├── 青浦小区信息.xlsx │ ├── 静安小区信息.xlsx │ └── 黄埔小区信息.xlsx ├── 崇明 │ ├── chongming_baozhen.csv │ ├── chongming_changxingdao21211.csv │ ├── chongming_chenjiazhen.csv │ ├── chongming_chongmingqita.csv │ ├── chongming_chongmingxincheng.csv │ └── chongming_hengshadao.csv ├── 徐家汇 │ ├── xuhui_caohejing.csv │ ├── xuhui_changqiao.csv │ ├── xuhui_hengshanlu.csv │ ├── xuhui_huadongligong.csv │ ├── xuhui_huajing.csv │ ├── xuhui_jianguoxilu.csv │ ├── xuhui_kangjian.csv │ ├── xuhui_longhua.csv │ ├── xuhui_shanghainanzhan.csv │ ├── xuhui_tianlin.csv │ ├── xuhui_wantiguan.csv │ ├── xuhui_xietulu.csv │ ├── xuhui_xuhuibinjiang.csv │ ├── xuhui_xujiahui.csv │ └── xuhui_zhiwuyuan.csv ├── 普陀 │ ├── putuo_caoyang.csv │ ├── putuo_changfeng1.csv │ ├── putuo_changshoulu.csv │ ├── putuo_changzheng.csv │ ├── putuo_ganquanyichuan.csv │ ├── putuo_guangxin.csv │ ├── putuo_taopu.csv │ ├── putuo_wanli.csv │ ├── putuo_wuning.csv │ ├── putuo_zhenguang.csv │ ├── putuo_zhenru.csv │ └── putuo_zhongyuanliangwancheng.csv ├── 杨浦 │ ├── yangpu_anshan.csv │ ├── yangpu_dongwaitan.csv │ ├── yangpu_gaojing.csv │ ├── yangpu_huangxinggongyuan.csv │ ├── yangpu_kongjianglu.csv │ ├── yangpu_wujiaochang.csv │ ├── yangpu_xinjiangwancheng.csv │ ├── yangpu_zhongyuan1.csv │ └── yangpu_zhoujiazuilu.csv ├── 松江 │ ├── songjiang_chedun.csv │ ├── songjiang_jiuting.csv │ ├── songjiang_maogang.csv │ ├── songjiang_shenminbieshu.csv │ ├── songjiang_sheshan.csv │ ├── songjiang_shihudang.csv │ ├── songjiang_sijing.csv │ ├── songjiang_songjiangdaxuecheng.csv │ ├── songjiang_songjianglaocheng.csv │ ├── songjiang_songjiangxincheng.csv │ ├── songjiang_xiaokunshan.csv │ ├── songjiang_xinbang.csv │ ├── songjiang_xinqiao.csv │ └── songjiang_yexie.csv ├── 浦东 │ ├── pudong_beicai.csv │ ├── pudong_biyun.csv │ ├── pudong_caolu.csv │ ├── pudong_chuansha.csv │ ├── pudong_datuanzhen.csv │ ├── pudong_gaodong.csv │ ├── pudong_gaohang.csv │ ├── pudong_geqing.csv │ ├── pudong_hangtou.csv │ ├── pudong_huamu.csv │ ├── pudong_huinan.csv │ ├── pudong_jinqiao.csv │ ├── pudong_jinyang.csv │ ├── pudong_kangqiao.csv │ ├── pudong_laogangzhen.csv │ ├── pudong_lianyang.csv │ ├── pudong_lingangxincheng.csv │ ├── pudong_lujiazui.csv │ ├── pudong_nanmatou.csv │ ├── pudong_nichengzhen.csv │ ├── pudong_sanlin.csv │ ├── pudong_shibo.csv │ ├── pudong_shuyuanzhen.csv │ ├── pudong_tangqiao.csv │ ├── pudong_tangzhen.csv │ ├── pudong_waigaoqiao.csv │ ├── pudong_wanxiangzhen.csv │ ├── pudong_weifang.csv │ ├── pudong_xinchang.csv │ ├── pudong_xuanqiao.csv │ ├── pudong_yangdong.csv │ ├── pudong_yangjing.csv │ ├── pudong_yuanshen.csv │ ├── pudong_yuqiao1.csv │ ├── pudong_zhangjiang.csv │ ├── pudong_zhoupu.csv │ └── pudong_zhuqiao.csv ├── 虹口 │ ├── hongkou_beiwaitan.csv │ ├── hongkou_jiangwanzhen.csv │ ├── hongkou_liangcheng.csv │ ├── hongkou_linpinglu.csv │ ├── hongkou_luxungongyuan.csv │ ├── hongkou_quyang.csv │ └── hongkou_sichuanbeilu.csv ├── 金山 │ ├── jinshan_caojing.csv │ ├── jinshan_fengjing.csv │ ├── jinshan_jinshan1.csv │ ├── jinshan_langxia.csv │ ├── jinshan_luxiang.csv │ ├── jinshan_shanyang.csv │ ├── jinshan_shihua.csv │ ├── jinshan_tinglin.csv │ ├── jinshan_zhangyan.csv │ └── jinshan_zhujing.csv ├── 长宁 │ ├── changning_beixinjing.csv │ ├── changning_gubei.csv │ ├── changning_hongqiao1.csv │ ├── changning_tianshan.csv │ ├── changning_xianxia.csv │ ├── changning_xijiao.csv │ ├── changning_xinhualu.csv │ ├── changning_zhenninglu.csv │ └── changning_zhongshangongyuan.csv ├── 闵行 │ ├── minhang_chunshen.csv │ ├── minhang_gumei.csv │ ├── minhang_hanghua.csv │ ├── minhang_huacao.csv │ ├── minhang_jinganxincheng.csv │ ├── minhang_jinhongqiao.csv │ ├── minhang_jinhui.csv │ ├── minhang_laominhang.csv │ ├── minhang_longbai.csv │ ├── minhang_maqiao.csv │ ├── minhang_meilong.csv │ ├── minhang_pujiang1.csv │ ├── minhang_qibao.csv │ ├── minhang_wujing.csv │ ├── minhang_xinzhuang5.csv │ └── minhang_zhuanqiao.csv ├── 青浦 │ ├── qingpu_baihe.csv │ ├── qingpu_chonggu.csv │ ├── qingpu_huaxin.csv │ ├── qingpu_jinze.csv │ ├── qingpu_liantang1.csv │ ├── qingpu_xianghuaqiao.csv │ ├── qingpu_xiayang.csv │ ├── qingpu_xujing.csv │ ├── qingpu_yingpu.csv │ ├── qingpu_zhaoxiang.csv │ └── qingpu_zhujiajiao.csv ├── 静安 │ ├── jingan_buyecheng.csv │ ├── jingan_caojiadu.csv │ ├── jingan_daning.csv │ ├── jingan_jiangninglu.csv │ ├── jingan_jingansi.csv │ ├── jingan_nanjingxilu.csv │ ├── jingan_pengpu.csv │ ├── jingan_xizangbeilu.csv │ ├── jingan_yangcheng.csv │ ├── jingan_yonghe.csv │ └── jingan_zhabeigongyuan.csv └── 黄埔 │ ├── huangpu_dapuqiao.csv │ ├── huangpu_dongjiadu.csv │ ├── huangpu_huaihaizhonglu.csv │ ├── huangpu_huangpubinjiang.csv │ ├── huangpu_laoximen.csv │ ├── huangpu_nanjingdonglu.csv │ ├── huangpu_penglaigongyuan.csv │ ├── huangpu_renminguangchang.csv │ ├── huangpu_shibobinjiang.csv │ ├── huangpu_wuliqiao.csv │ ├── huangpu_xintiandi.csv │ └── huangpu_yuyuan.csv ├── LJNTxiaoqu.py ├── LJNew_bouse.py ├── LJershoufang.xlsx ├── LJershoufang ├── LJershoufang.ipynb ├── LJershoufang.md ├── output_5_0.png ├── output_5_2.png ├── output_5_3.png ├── output_5_4.png ├── output_5_5.png ├── output_5_6.png ├── output_5_7.png └── 链家二手房数据分析.pdf ├── LJsecond_house ├── LJershoufang.ipynb ├── LJershoufang.md ├── output_5_0.png ├── output_5_2.png ├── output_5_3.png ├── output_5_4.png ├── output_5_5.png ├── output_5_6.png ├── output_5_7.png └── 链家二手房数据分析.pdf ├── LJzufang.csv ├── LJzufang.py ├── lianjiasecond_house.py ├── 上海小区爬虫最终版 ├── .gitignore ├── README.md ├── lib │ ├── __init__.py │ ├── const │ │ ├── __init__.py │ │ └── xpath.py │ ├── item │ │ ├── __init__.py │ │ └── xiaoqu.py │ ├── request │ │ ├── __init__.py │ │ ├── headers.py │ │ └── proxy.py │ ├── spider │ │ ├── __init__.py │ │ ├── base_spider.py │ │ └── xiaoqu_spider.py │ ├── utility │ │ ├── __init__.py │ │ ├── date.py │ │ ├── log.py │ │ ├── path.py │ │ ├── version.py │ │ └── writer.py │ └── zone │ │ ├── __init__.py │ │ ├── area.py │ │ ├── city.py │ │ └── district.py ├── log │ └── __init__.py ├── pic │ ├── district_top.png │ └── xiaoqu_top.png ├── requirements.txt ├── test │ └── date_test.py ├── tool │ ├── clean.py │ ├── ershou_image_with_coroutine.py │ ├── ershou_image_with_threads.py │ ├── ershou_number.py │ └── lianjia_xiaoqu.sql ├── xiaoqu.py ├── xiaoqu_to_chart.py └── xiaoqu_to_db.py ├── 小区 ├── LJxiaoqu.py ├── LJxiaoquHB.py ├── 上海链家小区信息(第1线程).csv ├── 上海链家小区信息(第1线程).xlsx └── 上海链家小区信息(第2线程).xlsx ├── 链家二手房.xlsx ├── 链家二手房数据整理.xlsx ├── 链家新房.csv ├── 链家新房.xlsx └── 链家租房.xlsx /.gitattributes: -------------------------------------------------------------------------------- 1 | *.js linguist-language=python 2 | *.css linguist-language=python 3 | *.html linguist-language=python 4 | -------------------------------------------------------------------------------- /11-10评论结果/11-10/pinglun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/11-10/pinglun.csv -------------------------------------------------------------------------------- /11-10评论结果/11-10/口碑评论.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/11-10/口碑评论.xlsx -------------------------------------------------------------------------------- /11-10评论结果/11-10/抓取评论详情结果.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/11-10/抓取评论详情结果.xlsx -------------------------------------------------------------------------------- /11-10评论结果/11-10/评论详情.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/11-10/评论详情.xlsx -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_20_19.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_20_19.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_20_34.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_20_34.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_20_49.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_20_49.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_21_04.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_21_04.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_21_19.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_21_19.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_21_49.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_21_49.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_22_27.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_22_27.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_22_42.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_22_42.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_22_59.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_22_59.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_23_14.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_23_14.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_23_28.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_23_28.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_23_43.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_23_43.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_24_09.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_24_09.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_24_24.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_24_24.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_24_39.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_24_39.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_24_54.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_24_54.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_25_20.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_25_20.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_26_09.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_26_09.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_26_36.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_26_36.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_26_50.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_26_50.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_27_17.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_27_17.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_27_43.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_27_43.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_27_58.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_27_58.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_28_26.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_28_26.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_28_41.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_28_41.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_29_07.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_29_07.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_29_21.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_29_21.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_29_36.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_29_36.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_29_51.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_29_51.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_30_30.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_30_30.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_30_57.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_30_57.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_31_12.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_31_12.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_31_39.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_31_39.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_31_54.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_31_54.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_32_08.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_32_08.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_32_24.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_32_24.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_32_50.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_32_50.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_33_28.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_33_28.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_33_43.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_33_43.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_33_57.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_33_57.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_34_34.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_34_34.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_35_00.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_35_00.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_35_13.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_35_13.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_35_27.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_35_27.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_35_41.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_35_41.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_36_07.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_36_07.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_36_21.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_36_21.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_36_34.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_36_34.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_36_48.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_36_48.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_37_02.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_37_02.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_37_16.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_37_16.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_37_30.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_37_30.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_38_32.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_38_32.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_38_58.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_38_58.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_39_12.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_39_12.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_39_26.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_39_26.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_39_39.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_39_39.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_39_54.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_39_54.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_40_08.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_40_08.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_40_22.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_40_22.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_40_37.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_40_37.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_40_51.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_40_51.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_41_41.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_41_41.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_41_56.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_41_56.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_42_10.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_42_10.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_42_25.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_42_25.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_43_27.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_43_27.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_43_54.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_43_54.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_44_44.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_44_44.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_45_10.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_45_10.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_45_25.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_45_25.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_45_52.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_45_52.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_46_07.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_46_07.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_46_33.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_46_33.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_46_48.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_46_48.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_47_03.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_47_03.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_47_17.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_47_17.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_47_32.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_47_32.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_47_47.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_47_47.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_48_02.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_48_02.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_48_17.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_48_17.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_48_32.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_48_32.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_48_46.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_48_46.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_49_01.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_49_01.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_49_40.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_49_40.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_49_54.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_49_54.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_50_20.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_50_20.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_50_37.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_50_37.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_50_52.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_50_52.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_51_07.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_51_07.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_51_47.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_51_47.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_52_01.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_52_01.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_52_53.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_52_53.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_53_19.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_53_19.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_53_45.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_53_45.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_54_10.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_54_10.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_54_24.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_54_24.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_54_38.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_54_38.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_55_03.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_55_03.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_55_40.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_55_40.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_56_06.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_56_06.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_56_32.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_56_32.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_56_46.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_56_46.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_56_59.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_56_59.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_57_13.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_57_13.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_58_14.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_58_14.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_58_29.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_58_29.png -------------------------------------------------------------------------------- /11-10评论结果/2019-11-10/2019-11-10-21_58_56.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/2019-11-10/2019-11-10-21_58_56.png -------------------------------------------------------------------------------- /11-10评论结果/Tableprocess.py: -------------------------------------------------------------------------------- 1 | import pandas as pd 2 | df1 = pd.read_excel(r'C:\Users\xiaoLiu\Desktop\口碑评论.xlsx', header=0, index_col=None) 3 | df2 = pd.read_excel(r'C:\Users\xiaoLiu\Desktop\评论详情.xlsx', header=0, index_col=None) 4 | df3 = pd.merge(df1, df2, how='outer', on='字段4_link', left_on=None, right_on=None, 5 | left_index=False, right_index=False, sort=False, 6 | suffixes=('_x', '_y'), copy=True, indicator=False) 7 | df3.to_csv('./lianjiepinglun.csv') -------------------------------------------------------------------------------- /11-10评论结果/koubeipinlun.py: -------------------------------------------------------------------------------- 1 | from aip import AipOcr 2 | import pandas as pd 3 | from sqlalchemy import create_engine 4 | import os 5 | import csv 6 | 7 | def getsceenshot(search_url,list): 8 | from selenium import webdriver 9 | import os 10 | import time 11 | from selenium.webdriver.common.by import By 12 | from selenium.webdriver.support.ui import WebDriverWait 13 | from selenium.webdriver.support import expected_conditions as EC 14 | from selenium import webdriver 15 | 16 | picture_time = time.strftime("%Y-%m-%d-%H_%M_%S", time.localtime(time.time())) 17 | directory_time = time.strftime("%Y-%m-%d", time.localtime(time.time())) 18 | 19 | try: 20 | File_Path = os.getcwd() + '\\' + directory_time + '\\' 21 | if not os.path.exists(File_Path): 22 | os.makedirs(File_Path) 23 | print("存入文件%s" % File_Path) 24 | else: 25 | print("写入新文件") 26 | except BaseException as msg: 27 | print("异常:%s" % msg) 28 | try: 29 | options = webdriver.ChromeOptions() 30 | options.add_argument('--headless') 31 | options.add_argument('--disable-gpu') 32 | driver = webdriver.Chrome(options=options) 33 | driver.maximize_window() 34 | 35 | js_height = "return document.body.clientHeight" 36 | 37 | # driver = webdriver.Chrome() 38 | driver.get(search_url) 39 | driver.execute_script(""" 40 | (function () { 41 | $('.subnav-title').hide(); 42 | $('.video-iframe').hide(); 43 | $('.mouth-remak').hide(); 44 | $('.contmain').hide(); 45 | $('.footer_auto').hide(); 46 | $('.mouth-title-end over-hid').hide(); 47 | $('.user-cont').hide(); 48 | $('.mouth-title-end').hide(); 49 | $('.nav-typebar-g12').hide(); 50 | $('.breadnav').hide(); 51 | $('.minitop').hide(); 52 | $('.topbar-header').hide(); 53 | $('.gotop02').hide(); 54 | $('.image-div').hide(); 55 | $('.video-container').hide(); 56 | $('.advbox1').hide(); 57 | $('.mouthcon-cont-left').hide(); 58 | $('.mouthcon-cont-right').attr('width','800'); 59 | $('.text-con').attr('font-size','18'); 60 | $('.img-list').hide(); 61 | })(); 62 | """) 63 | 64 | k = 1 65 | height = driver.execute_script(js_height) 66 | while True: 67 | if k * 500 < height: 68 | js_move = "window.scrollTo(0,{})".format(k * 500) 69 | print(js_move) 70 | driver.execute_script(js_move) 71 | time.sleep(0.2) 72 | height = driver.execute_script(js_height) 73 | k += 1 74 | else: 75 | break 76 | scroll_width = driver.execute_script('return document.body.parentNode.scrollWidth') 77 | scroll_height = driver.execute_script('return document.body.parentNode.scrollHeight') 78 | driver.set_window_size(scroll_width, scroll_height) 79 | driver.get_screenshot_as_file( 80 | '.\\' + directory_time + '\\' + picture_time + '.png') 81 | print("Process {} get one pic !!!".format(os.getpid())) 82 | list.append(search_url) 83 | print(list) 84 | 85 | 86 | except Exception as e: 87 | print(e) 88 | 89 | 90 | finally: 91 | driver.quit() 92 | print('爬取完成') 93 | return ('.\\' + directory_time + '\\' + picture_time + '.png') 94 | 95 | def getlinks(links): 96 | list=[] 97 | for link in links: 98 | getsceenshot(link,list) 99 | df=pd.Series(list) 100 | df.to_csv('./lianjie.csv',index=False) 101 | 102 | def getfile(rootdir): 103 | list = os.listdir(rootdir)#列出文件夹下所有的目录与文件 104 | lujing=[] 105 | for i in range(0,len(list)): 106 | rootdir = r'C:\Users\xiaoLiu\Desktop\11-10评论结果\2019-11-14' 107 | path = os.path.join(rootdir,list[i]) 108 | lujing.append(path) 109 | df=pd.Series(lujing) 110 | df.to_csv('./pictureName.csv',index=False) 111 | 112 | def get_file_content(filePath): 113 | """ 读取图片 """ 114 | with open(filePath, 'rb') as fp: 115 | return fp.read() 116 | 117 | def change_to_words(imgpath): 118 | APP_ID = '17523158' 119 | API_KEY = '9wyiqfr4Ob0llopVzzigi5T3' 120 | SECRET_KEY = 'Mo6Ctv7d6HstNesgPfiCOwxgeSw7nKyG' 121 | aipOcr = AipOcr(APP_ID, API_KEY, SECRET_KEY) 122 | for column in imgpath: 123 | options = {'detect_direction': 'true','language_type': 'CHN_ENG'} 124 | result = aipOcr.basicGeneral(get_file_content(column), options) 125 | words_result=result['words_result'] 126 | list_result=str() 127 | for i in range(len(words_result)): 128 | list_result += words_result[i]['words'] 129 | df=pd.Series(list_result) 130 | engine = create_engine("mysql+pymysql://root:960614abcd@localhost:3306/xiaoliu") 131 | df.to_sql(name = 'pinglun',con = engine,if_exists = 'append',index = False,index_label = False) 132 | 133 | if __name__ == '__main__': 134 | df = pd.read_excel(r'C:\Users\xiaoLiu\Desktop\未处理链接.xlsx', header=0, index_col=None) 135 | links = df['字段4_link'] 136 | getlinks(links) 137 | rootdir = r'C:\Users\xiaoLiu\Desktop\11-10评论结果\2019-11-14' 138 | getfile(rootdir) 139 | with open('./pictureName.csv','r',encoding='utf-8') as csvfile: 140 | reader = csv.reader(csvfile) 141 | imgpath = [row[0]for row in reader] 142 | change_to_words(imgpath) -------------------------------------------------------------------------------- /11-10评论结果/pinglun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/pinglun.csv -------------------------------------------------------------------------------- /11-10评论结果/口碑评论.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/口碑评论.xlsx -------------------------------------------------------------------------------- /11-10评论结果/抓取评论详情结果.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/抓取评论详情结果.xlsx -------------------------------------------------------------------------------- /11-10评论结果/评论详情.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/11-10评论结果/评论详情.xlsx -------------------------------------------------------------------------------- /5K壁纸爬虫.py: -------------------------------------------------------------------------------- 1 | from requests import get 2 | from filetype import guess 3 | from os import rename 4 | from os import makedirs 5 | from os.path import exists 6 | from json import loads 7 | from contextlib import closing 8 | 9 | 10 | # 文件下载器 11 | def Down_load(file_url, file_full_name, now_photo_count, all_photo_count): 12 | headers = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36"} 13 | 14 | # 开始下载图片 15 | with closing(get(file_url, headers=headers, stream=True)) as response: 16 | chunk_size = 1024 # 单次请求最大值 17 | content_size = int(response.headers['content-length']) # 文件总大小 18 | data_count = 0 # 当前已传输的大小 19 | with open(file_full_name, "wb") as file: 20 | for data in response.iter_content(chunk_size=chunk_size): 21 | file.write(data) 22 | done_block = int((data_count / content_size) * 50) 23 | data_count = data_count + len(data) 24 | now_jd = (data_count / content_size) * 100 25 | print("\r %s:[%s%s] %d%% %d/%d" % (file_full_name, done_block * '█', ' ' * (50 - 1 - done_block), now_jd, now_photo_count, all_photo_count), end=" ") 26 | 27 | # 下载完图片后获取图片扩展名,并为其增加扩展名 28 | file_type = guess(file_full_name) 29 | rename(file_full_name, file_full_name + '.' + file_type.extension) 30 | 31 | 32 | 33 | # 爬取不同类型图片 34 | def crawler_photo(type_id, photo_count): 35 | 36 | # 最新 1, 最热 2, 女生 3, 星空 4 37 | if(type_id == 1): 38 | url = 'https://service.paper.meiyuan.in/api/v2/columns/flow/5c68ffb9463b7fbfe72b0db0?page=1&per_page=' + str(photo_count) 39 | elif(type_id == 2): 40 | url = 'https://service.paper.meiyuan.in/api/v2/columns/flow/5c69251c9b1c011c41bb97be?page=1&per_page=' + str(photo_count) 41 | elif(type_id == 3): 42 | url = 'https://service.paper.meiyuan.in/api/v2/columns/flow/5c81087e6aee28c541eefc26?page=1&per_page=' + str(photo_count) 43 | elif(type_id == 4): 44 | url = 'https://service.paper.meiyuan.in/api/v2/columns/flow/5c81f64c96fad8fe211f5367?page=1&per_page=' + str(photo_count) 45 | 46 | # 获取图片列表数据 47 | headers = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36"} 48 | respond = get(url, headers=headers) 49 | photo_data = loads(respond.content) 50 | 51 | # 已经下载的图片张数 52 | now_photo_count = 1 53 | 54 | # 所有图片张数 55 | all_photo_count = len(photo_data) 56 | 57 | # 开始下载并保存5K分辨率壁纸 58 | for photo in photo_data: 59 | 60 | # 创建一个文件夹存放我们下载的图片 61 | if not exists('./' + str(type_id)): 62 | makedirs('./' + str(type_id)) 63 | 64 | # 准备下载的图片链接 65 | file_url = photo['urls']['raw'] 66 | 67 | # 准备下载的图片名称,不包含扩展名 68 | file_name_only = file_url.split('/') 69 | file_name_only = file_name_only[len(file_name_only) -1] 70 | 71 | # 准备保存到本地的完整路径 72 | file_full_name = './' + str(type_id) + '/' + file_name_only 73 | 74 | # 开始下载图片 75 | Down_load(file_url, file_full_name, now_photo_count, all_photo_count) 76 | now_photo_count = now_photo_count + 1 77 | 78 | 79 | 80 | if __name__ == '__main__': 81 | 82 | # 最新 1, 最热 2, 女生 3, 星空 4 83 | # 爬取类型为3的图片(女生),一共准备爬取20000张 84 | wall_paper_id = 1 85 | wall_paper_count = 10 86 | while(True): 87 | 88 | # 换行符 89 | print('\n\n') 90 | 91 | # 选择壁纸类型 92 | wall_paper_id = input("壁纸类型:最新壁纸 1, 最热壁纸 2, 女生壁纸 3, 星空壁纸 4\n请输入编号以便选择5K超清壁纸类型:") 93 | # 判断输入是否正确 94 | while(wall_paper_id != str(1) and wall_paper_id != str(2) and wall_paper_id != str(3) and wall_paper_id != str(4)): 95 | wall_paper_id = input("壁纸类型:最新壁纸 1, 最热壁纸 2, 女生壁纸 3, 星空壁纸 4\n请输入编号以便选择5K超清壁纸类型:") 96 | 97 | 98 | # 选择要下载的壁纸数量 99 | wall_paper_count = input("请输入要下载的5K超清壁纸的数量:") 100 | # 判断输入是否正确 101 | while(int(wall_paper_count) <= 0): 102 | wall_paper_count = input("请输入要下载的5K超清壁纸的数量:") 103 | 104 | 105 | # 开始爬取5K高清壁纸 106 | print("正在下载5K超清壁纸,请稍等……") 107 | crawler_photo(int(wall_paper_id), int(wall_paper_count)) 108 | print('\n下载5K高清壁纸成功!') -------------------------------------------------------------------------------- /NBA/NBA Player.py: -------------------------------------------------------------------------------- 1 | import pandas as pd 2 | import numpy as n 3 | 4 | # 读取网页中的数据表 5 | table = [] 6 | for i in range(1,7): 7 | table.append(pd.read_html('https://nba.hupu.com/stats/players/pts/%d' %i)[0]) 8 | 9 | # 所有数据纵向合并为数据框 10 | players = pd.concat(table) 11 | # 变量重命名 12 | columns=['排名','球员','球队','得分','命中-出手','命中率','命中-三分','三分命中率','命中-罚球','罚球命中率','场次','上场时间'] 13 | players.columns=columns 14 | 15 | 16 | players.drop(0,inplace=True) 17 | players.to_csv(r"C:\Users\MI\Desktop\球员信息.csv",encoding='utf_8_sig') 18 | 19 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # PythonSpyder_100_examples 2 | 自学Python爬虫路上的实战笔记,由浅到深逐步深入学习Python 爬虫 3 | -------------------------------------------------------------------------------- /多线程爬取表情包.py: -------------------------------------------------------------------------------- 1 | import requests 2 | import random 3 | import os 4 | from bs4 import BeautifulSoup 5 | import threading 6 | 7 | class crawler_pic(threading.Thread): 8 | begin_index = 0 # 起始页面 9 | end_index = 0 # 终止页 10 | grads = 20 # 爬取梯度:每个线程爬虫需要执行的爬取页数 11 | # 链接 12 | base_url = "http://www.win4000.com/wallpaper_big_154{}.html" 13 | # 图片保存根目录 14 | file_root = "D://pics_multi//" 15 | # 伪装浏览器 16 | UA = [ 17 | "Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50", 18 | "Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50", 19 | "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:2.0.1) Gecko/20100101 Firefox/4.0.1", 20 | "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1" 21 | ] 22 | # 随机构造头部信息 23 | headers = { 24 | "User-Agent": random.choice(UA) 25 | } 26 | def __init__(self, name, begin): 27 | threading.Thread.__init__(self) 28 | self.name = name 29 | self.begin_index = begin 30 | self.end_index = begin + self.grads 31 | # 获取 32 | def get_html(self, url): 33 | try: 34 | HTML = requests.get(url,headers=self.headers) 35 | HTML.raise_for_status() 36 | HTML.encoding = HTML.apparent_encoding 37 | return HTML.text 38 | except: 39 | print("In "+self.name+":ERROR Load "+url) 40 | return "NULL" 41 | # 将获取的图片存储至根目录下 42 | def store_pics(self,pic_urls): 43 | fileName = pic_urls[0]+"//" 44 | for picurl in pic_urls[1:]: 45 | # 构造图片存储地址 46 | path = self.file_root + fileName + picurl.split('/')[-1] 47 | print(path) 48 | 49 | try: 50 | # 需要逐层创建目录 51 | if not os.path.exists(self.file_root): 52 | os.mkdir(self.file_root) 53 | # 如无该目录,先行构建 54 | if not os.path.exists(self.file_root+fileName): 55 | os.mkdir(self.file_root+fileName) 56 | # 图片存在,不重复保存 57 | # 不存在,创建 58 | if not os.path.exists(path): 59 | # request获取图片内容 60 | pic = requests.get(picurl) 61 | with open(path, 'wb') as f: 62 | f.write(pic.content) 63 | f.close() 64 | print("图片:" + picurl + " 成功下载") 65 | else: 66 | print("图片已存在") 67 | except: 68 | print("爬取失败") 69 | return 1 70 | 71 | # 在html页面中获取图片链接,返回链接列表 72 | def get_pic_urls(self, HTML): 73 | 74 | pic_urls = ["filename"] 75 | soup = BeautifulSoup(HTML, "html.parser") 76 | """ 77 | 页面分析: 78 | 图片链接位于标签
--
  • -- [href:pic_url] 79 | 获取最上层:div 全部子孙标签 选取a 获取a的属性信息 80 | """ 81 | for tag in soup.find("div", attrs={"id": "picBox", "class": "picBox"}).descendants: 82 | if tag.name == 'img': 83 | pic_urls.append(tag.attrs['src']) 84 | pic_urls[0] = tag.attrs['title'] 85 | """ 86 | for a_tag in soup.find("div", attrs={"id": "picBox", "class": "picBox"}).findAll("a"): 87 | pic_urls.append(a_tag.attrs['href']) 88 | """ 89 | # 全局,记录图片数量 90 | global pic_num 91 | pic_num += len(pic_urls) - 1 92 | return pic_urls 93 | 94 | # 线程方法 95 | def run(self): 96 | # 爬取一遍分配的页面 97 | for i in range(self.begin_index,self.end_index): 98 | html = self.get_html(self.base_url.format(i)) 99 | # 页面爬取成功的情况下获取图片链接 100 | if html != "NULL": 101 | pic_urls = self.get_pic_urls(html) 102 | self.store_pics(pic_urls) 103 | """ 104 | for pic in pic_urls: 105 | print("in "+self.name+":"+pic) 106 | """ 107 | 108 | 109 | if __name__ == '__main__': 110 | 111 | threads = [] 112 | count = 0 113 | pic_num = 0 114 | # 构造爬虫 115 | for begin in range(700,900,20): 116 | threads.append(crawler_pic("Thread-begin:"+str(begin),begin)) 117 | 118 | # 开始爬取 119 | for thread in threads: 120 | thread.start() 121 | 122 | for thread in threads: 123 | thread.join() 124 | 125 | 126 | print(pic_num) -------------------------------------------------------------------------------- /大学排名爬虫.py: -------------------------------------------------------------------------------- 1 | import requests 2 | from bs4 import BeautifulSoup 3 | import bs4 4 | 5 | def getHTMLText(url): 6 | try: 7 | r = requests.get(url,timeout = 30) 8 | r.raise_for_status() 9 | r.encoding = r.apparent_encoding 10 | return r.text 11 | except: 12 | print("getError") 13 | return "" 14 | return "" 15 | 16 | def fillUnivList(ulist,html): 17 | soup = BeautifulSoup(html,"html.parser") 18 | """ 19 | 结构分析 20 | 排名位于tbody标签下 21 | 每个tr标签是一个学校的信息 22 | tr标签下有多个td标签,保存有该学校的各类指标 23 | """ 24 | # 遍历tr标签 tr标签是tbody标签的孩子 25 | for tr in soup.find('tbody').children: 26 | print(tr) 27 | if isinstance(tr,bs4.element.Tag): 28 | # 获取tr标签下的td标签 29 | tds = tr('td') 30 | # 获取相关指标 只需要第 0 1 3 个相关td标签,分别是学校名称,排名,分数 31 | ulist.append([tds[0].string,tds[1].string,tds[3].string]) 32 | 33 | # 打印前20的榜单 34 | def printUnivList(ulist,num): 35 | """ 36 | print("{:^10}\t{:^6}\t{:^10}".format("排名","学校名称","分数")) 37 | for i in range(num): 38 | u = ulist[i] 39 | print("{:^10}\t{:^6}\t{:^10}".format(u[0],u[1],u[2])) 40 | """ 41 | 42 | # 优化,解决中文不对齐问题 43 | #^num num代表占位 44 | tplt = "{0:^10}\t{1:{3}^10}\t{2:^10}" 45 | # chr(12288)是中文空白符 46 | print(tplt.format("排名","学校名称","分数",chr(12288))) 47 | for i in range(num): 48 | u = ulist[i] 49 | print(tplt.format(u[0],u[1],u[2],chr(12288))) 50 | def main(): 51 | unifo = [] 52 | url = "http://www.zuihaodaxue.cn/zuihaodaxuepaiming2019.html" 53 | html = getHTMLText(url) 54 | fillUnivList(unifo,html) 55 | printUnivList(unifo,100) 56 | 57 | main() -------------------------------------------------------------------------------- /安居客二手房.py: -------------------------------------------------------------------------------- 1 | import requests 2 | import time 3 | import csv 4 | import random 5 | from bs4 import BeautifulSoup 6 | 7 | headers = { 8 | 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0', 9 | 'Cookie':'sessid=F3A0D86E-9D3F-82F9-1BBA-D4A22CE6440F; aQQ_ajkguid=35DAB075-3352-357E-F3D9-D46F6E3E210A; lps=http%3A%2F%2Fwww.anjuke.com%2F%3Fpi%3DPZ-baidu-pc-all-biaoti%7Chttps%3A%2F%2Fsp0.baidu.com%2F9q9JcDHa2gU2pMbgoY3K%2Fadrc.php%3Ft%3D06KL00c00f7Hj1f0q3V-00PpAsK8poKI00000FYEi7C00000I5matL.THvs_oeHEtY0UWdVUv4_py4-g1PxuAT0T1d9PyPhPymLPW0sn1N-mW790ZRqwRuKfYFAPRwDfbRdrRPjfWKanWfsnYmsPYu7n1wKrHc0mHdL5iuVmv-b5HnzrHDvnH61njchTZFEuA-b5HDv0ARqpZwYTZnlQzqLILT8my4JIyV-QhPEUitOTAbqR7CVmh7GuZRVTAnVmyk_QyFGmyqYpfKWThnqPHRvP10%26tpl%3Dtpl_11534_19968_16032%26l%3D1514680221%26attach%3Dlocation%253D%2526linkName%253D%2525E6%2525A0%252587%2525E5%252587%252586%2525E5%2525A4%2525B4%2525E9%252583%2525A8-%2525E6%2525A0%252587%2525E9%2525A2%252598-%2525E4%2525B8%2525BB%2525E6%2525A0%252587%2525E9%2525A2%252598%2526linkText%253D%2525E5%2525AE%252589%2525E5%2525B1%252585%2525E5%2525AE%2525A2-%2525E5%252585%2525A8%2525E6%252588%2525BF%2525E6%2525BA%252590%2525E7%2525BD%252591%2525EF%2525BC%25258C%2525E6%252596%2525B0%2525E6%252588%2525BF%252520%2525E4%2525BA%25258C%2525E6%252589%25258B%2525E6%252588%2525BF%252520%2525E6%25258C%252591%2525E5%2525A5%2525BD%2525E6%252588%2525BF%2525E4%2525B8%25258A%2525E5%2525AE%252589%2525E5%2525B1%252585%2525E5%2525AE%2525A2%2525EF%2525BC%252581%2526xp%253Did%28%252522m3291618302_canvas%252522%29%25252FDIV%25255B1%25255D%25252FDIV%25255B1%25255D%25252FDIV%25255B1%25255D%25252FDIV%25255B1%25255D%25252FDIV%25255B1%25255D%25252FH2%25255B1%25255D%25252FA%25255B1%25255D%2526linkType%253D%2526checksum%253D129%26ie%3Dutf-8%26f%3D8%26tn%3Dmonline_3_dg%26wd%3D%25E5%25AE%2589%25E5%25B1%2585%25E5%25AE%25A2%26oq%3D%25E5%25AE%2589%25E5%25B1%2585%25E5%25AE%25A2%26rqlang%3Dcn; ctid=11; twe=2; _stat_guid=3050B320-8AA3-6F6A-9549-5BF2054466BB; _stat_rfpn=Ershou_Web_Property_List_FilterPage_tracklog; _prev_stat_guid=35DAB075-3352-357E-F3D9-D46F6E3E210A; 58tj_uuid=37c162b8-8162-4b45-aedf-dfa05080a10a; new_session=0; init_refer=https%253A%252F%252Fsp0.baidu.com%252F9q9JcDHa2gU2pMbgoY3K%252Fadrc.php%253Ft%253D06KL00c00f7Hj1f0q3V-00PpAsK8poKI00000FYEi7C00000I5matL.THvs_oeHEtY0UWdVUv4_py4-g1PxuAT0T1d9PyPhPymLPW0sn1N-mW790ZRqwRuKfYFAPRwDfbRdrRPjfWKanWfsnYmsPYu7n1wKrHc0mHdL5iuVmv-b5HnzrHDvnH61njchTZFEuA-b5HDv0ARqpZwYTZnlQzqLILT8my4JIyV-QhPEUitOTAbqR7CVmh7GuZRVTAnVmyk_QyFGmyqYpfKWThnqPHRvP10%2526tpl%253Dtpl_11534_19968_16032%2526l%253D1514680221%2526attach%253Dlocation%25253D%252526linkName%25253D%252525E6%252525A0%25252587%252525E5%25252587%25252586%252525E5%252525A4%252525B4%252525E9%25252583%252525A8-%252525E6%252525A0%25252587%252525E9%252525A2%25252598-%252525E4%252525B8%252525BB%252525E6%252525A0%25252587%252525E9%252525A2%25252598%252526linkText%25253D%252525E5%252525AE%25252589%252525E5%252525B1%25252585%252525E5%252525AE%252525A2-%252525E5%25252585%252525A8%252525E6%25252588%252525BF%252525E6%252525BA%25252590%252525E7%252525BD%25252591%252525EF%252525BC%2525258C%252525E6%25252596%252525B0%252525E6%25252588%252525BF%25252520%252525E4%252525BA%2525258C%252525E6%25252589%2525258B%252525E6%25252588%252525BF%25252520%252525E6%2525258C%25252591%252525E5%252525A5%252525BD%252525E6%25252588%252525BF%252525E4%252525B8%2525258A%252525E5%252525AE%25252589%252525E5%252525B1%25252585%252525E5%252525AE%252525A2%252525EF%252525BC%25252581%252526xp%25253Did%28%25252522m3291618302_canvas%25252522%29%2525252FDIV%2525255B1%2525255D%2525252FDIV%2525255B1%2525255D%2525252FDIV%2525255B1%2525255D%2525252FDIV%2525255B1%2525255D%2525252FDIV%2525255B1%2525255D%2525252FH2%2525255B1%2525255D%2525252FA%2525255B1%2525255D%252526linkType%25253D%252526checksum%25253D129%2526ie%253Dutf-8%2526f%253D8%2526tn%253Dmonline_3_dg%2526wd%253D%2525E5%2525AE%252589%2525E5%2525B1%252585%2525E5%2525AE%2525A2%2526oq%253D%2525E5%2525AE%252589%2525E5%2525B1%252585%2525E5%2525AE%2525A2%2526rqlang%253Dcn; new_uv=1; _ga=GA1.2.371539118.1571842678; _gid=GA1.2.389870108.1571842678; als=0; __xsptplus8=8.1.1571842677.1571846354.37%232%7Csp0.baidu.com%7C%7C%7C%25E5%25AE%2589%25E5%25B1%2585%25E5%25AE%25A2%7C%23%23A8ZibGF0HQquvvworKFYDClvSKK6Plsf%23; isp=true; Hm_lvt_c5899c8768ebee272710c9c5f365a6d8=1571842689; Hm_lpvt_c5899c8768ebee272710c9c5f365a6d8=1571844621; __xsptplusUT_8=1; _gat=1' 10 | } 11 | 12 | 13 | def parse_pages(url, num): 14 | response = requests.get(url=url, headers=headers) 15 | soup = BeautifulSoup(response.text, 'lxml') 16 | result_list = soup.find_all('li', class_='list-item') 17 | # print(len(result_list)) 18 | for result in result_list: 19 | # 标题 20 | title = result.find('a', class_='houseListTitle').text.strip() 21 | # print(title) 22 | # 户型 23 | layout = result.select('.details-item > span')[0].text 24 | # print(layout) 25 | # 面积 26 | cover = result.select('.details-item > span')[1].text 27 | # print(cover) 28 | # 楼层 29 | floor = result.select('.details-item > span')[2].text 30 | # print(floor) 31 | # 建造年份 32 | year = result.select('.details-item > span')[3].text 33 | # print(year) 34 | # 单价 35 | unit_price = result.find('span', class_='unit-price').text.strip() 36 | # print(unit_price) 37 | # 总价 38 | total_price = result.find('span', class_='price-det').text.strip() 39 | # print(total_price) 40 | # 关键字 41 | keyword = result.find('div', class_='tags-bottom').text.strip() 42 | # print(keyword) 43 | # 地址 44 | address = result.find('span', class_='comm-address').text.replace(' ', '').replace('\n', '') 45 | # print(address) 46 | # 详情页url 47 | details_url = result.find('a', class_='houseListTitle')['href'] 48 | # print(details_url) 49 | results = [title, layout, cover, floor, year, unit_price, total_price, keyword, address, details_url] 50 | with open('anjuke.csv', 'a', newline='', encoding='utf-8-sig') as f: 51 | w = csv.writer(f) 52 | w.writerow(results) 53 | 54 | # 判断是否还有下一页 55 | next_url = soup.find_all('a', class_='aNxt') 56 | if len(next_url) != 0: 57 | num += 1 58 | print('第' + str(num) + '页数据爬取完毕!') 59 | # 3-60秒之间随机暂停 60 | time.sleep(random.randint(3, 60)) 61 | parse_pages(next_url[0].attrs['href'], num) 62 | else: 63 | print('所有数据爬取完毕!') 64 | 65 | 66 | if __name__ == '__main__': 67 | with open('anjuke.csv', 'a', newline='', encoding='utf-8-sig') as fp: 68 | writer = csv.writer(fp) 69 | writer.writerow(['标题', '户型', '面积', '楼层', '建造年份', '单价', '总价', '关键字', '地址', '详情页地址']) 70 | start_num = 0 71 | start_url = 'https://shanghai.anjuke.com/sale/' 72 | parse_pages(start_url, start_num) 73 | 74 | -------------------------------------------------------------------------------- /王者荣耀英雄信息爬取.py: -------------------------------------------------------------------------------- 1 | #-*- coding: UTF-8 -*- 2 | from urllib.request import urlretrieve 3 | import requests 4 | import os 5 | 6 | 7 | def hero_imgs_download(url, header): 8 | req = requests.get(url = url, headers = header).json() 9 | hero_num = len(req['list']) 10 | print('一共有%d个英雄' % hero_num) 11 | hero_images_path = 'hero_images' 12 | for each_hero in req['list']: 13 | hero_photo_url = each_hero['cover'] 14 | hero_name = each_hero['name'] + '.jpg' 15 | filename = hero_images_path + '/' + hero_name 16 | if hero_images_path not in os.listdir(): 17 | os.makedirs(hero_images_path) 18 | urlretrieve(url = hero_photo_url, filename = filename) 19 | 20 | 21 | def hero_list(url, header): 22 | print('*' * 100) 23 | print('\t\t\t\t欢迎使用《王者荣耀》出装下助手!') 24 | print('*' * 100) 25 | req = requests.get(url = url, headers = header).json() 26 | flag = 0 27 | for each_hero in req['list']: 28 | flag += 1 29 | print('%s的ID为:%-7s' % (each_hero['name'], each_hero['hero_id']), end = '\t\t') 30 | if flag == 3: 31 | print('\n', end = '') 32 | flag = 0 33 | 34 | 35 | def seek_weapon(equip_id, weapon_info): 36 | for each_weapon in weapon_info: 37 | if each_weapon['equip_id'] == str(equip_id): 38 | weapon_name = each_weapon['name'] 39 | weapon_price = each_weapon['price'] 40 | return weapon_name, weapon_price 41 | 42 | 43 | def hero_info(url, header, weapon_info): 44 | req = requests.get(url = url, headers = header).json() 45 | print('\n历史上的%s:\n %s' % (req['info']['name'], req['info']['history_intro'])) 46 | for each_equip_choice in req['info']['equip_choice']: 47 | print('\n%s:\n %s' % (each_equip_choice['title'], each_equip_choice['description'])) 48 | total_price = 0 49 | flag = 0 50 | for each_weapon in each_equip_choice['list']: 51 | flag += 1 52 | weapon_name, weapon_price = seek_weapon(each_weapon['equip_id'], weapon_info) 53 | print('%s:%s' % (weapon_name, weapon_price), end = '\t') 54 | if flag == 3: 55 | print('\n', end = '') 56 | flag = 0 57 | total_price += int(weapon_price) 58 | print('神装套件价格共计:%d' % total_price) 59 | 60 | def hero_weapon(url, header): 61 | req = requests.get(url = url, headers = header).json() 62 | weapon_info_dict = req['list'] 63 | return weapon_info_dict 64 | 65 | 66 | if __name__ == '__main__': 67 | headers = {'Accept-Charset': 'UTF-8', 68 | 'Accept-Encoding': 'gzip,deflate', 69 | 'User-Agent': 'Dalvik/2.1.0 (Linux; U; Android 6.0.1; MI 5 MIUI/V8.1.6.0.MAACNDI)', 70 | 'X-Requested-With': 'XMLHttpRequest', 71 | 'Content-type': 'application/x-www-form-urlencoded', 72 | 'Connection': 'Keep-Alive', 73 | 'Host': 'gamehelper.gm825.com'} 74 | weapon_url = "http://gamehelper.gm825.com/wzry/equip/list?channel_id=90009a&app_id=h9044j&game_id=7622&game_name=%E7%8E%8B%E8%80%85%E8%8D%A3%E8%80%80&vcode=12.0.3&version_code=1203&cuid=2654CC14D2D3894DBF5808264AE2DAD7&ovr=6.0.1&device=Xiaomi_MI+5&net_type=1&client_id=1Yfyt44QSqu7PcVdDduBYQ%3D%3D&info_ms=fBzJ%2BCu4ZDAtl4CyHuZ%2FJQ%3D%3D&info_ma=XshbgIgi0V1HxXTqixI%2BKbgXtNtOP0%2Fn1WZtMWRWj5o%3D&mno=0&info_la=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&info_ci=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&mcc=0&clientversion=&bssid=VY%2BeiuZRJ%2FwaXmoLLVUrMODX1ZTf%2F2dzsWn2AOEM0I4%3D&os_level=23&os_id=dc451556fc0eeadb&resolution=1080_1920&dpi=480&client_ip=192.168.0.198&pdunid=a83d20d8" 75 | heros_url = "http://gamehelper.gm825.com/wzry/hero/list?channel_id=90009a&app_id=h9044j&game_id=7622&game_name=%E7%8E%8B%E8%80%85%E8%8D%A3%E8%80%80&vcode=12.0.3&version_code=1203&cuid=2654CC14D2D3894DBF5808264AE2DAD7&ovr=6.0.1&device=Xiaomi_MI+5&net_type=1&client_id=1Yfyt44QSqu7PcVdDduBYQ%3D%3D&info_ms=fBzJ%2BCu4ZDAtl4CyHuZ%2FJQ%3D%3D&info_ma=XshbgIgi0V1HxXTqixI%2BKbgXtNtOP0%2Fn1WZtMWRWj5o%3D&mno=0&info_la=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&info_ci=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&mcc=0&clientversion=&bssid=VY%2BeiuZRJ%2FwaXmoLLVUrMODX1ZTf%2F2dzsWn2AOEM0I4%3D&os_level=23&os_id=dc451556fc0eeadb&resolution=1080_1920&dpi=480&client_ip=192.168.0.198&pdunid=a83d20d8" 76 | hero_list(heros_url, headers) 77 | hero_id = input("请输入要查询的英雄ID:") 78 | hero_url = "http://gamehelper.gm825.com/wzry/hero/detail?hero_id={}&channel_id=90009a&app_id=h9044j&game_id=7622&game_name=%E7%8E%8B%E8%80%85%E8%8D%A3%E8%80%80&vcode=12.0.3&version_code=1203&cuid=2654CC14D2D3894DBF5808264AE2DAD7&ovr=6.0.1&device=Xiaomi_MI+5&net_type=1&client_id=1Yfyt44QSqu7PcVdDduBYQ%3D%3D&info_ms=fBzJ%2BCu4ZDAtl4CyHuZ%2FJQ%3D%3D&info_ma=XshbgIgi0V1HxXTqixI%2BKbgXtNtOP0%2Fn1WZtMWRWj5o%3D&mno=0&info_la=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&info_ci=9AChHTMC3uW%2BfY8%2BCFhcFw%3D%3D&mcc=0&clientversion=&bssid=VY%2BeiuZRJ%2FwaXmoLLVUrMODX1ZTf%2F2dzsWn2AOEM0I4%3D&os_level=23&os_id=dc451556fc0eeadb&resolution=1080_1920&dpi=480&client_ip=192.168.0.198&pdunid=a83d20d8".format(hero_id) 79 | weapon_info_dict = hero_weapon(weapon_url, headers) 80 | hero_info(hero_url, headers, weapon_info_dict) -------------------------------------------------------------------------------- /美桌网图片爬虫.py: -------------------------------------------------------------------------------- 1 | import requests 2 | import re 3 | import os 4 | import time 5 | url_root = 'http://www.win4000.com/wallpaper_big_154' 6 | # http://www.win4000.com/wallpaper_big_154(3bits).html 7 | user = {"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36"} 8 | pattern = re.compile(r'http://pic1.win4000.com/wallpaper/[\w|-]+/[\w]+.jpg') 9 | 10 | def get_picture_url(suffix): 11 | try: 12 | url = url_root + str(suffix) + ".html" 13 | print(url) 14 | r = requests.get(url,headers = user) 15 | r.raise_for_status() 16 | validpart = r.text.split('当前位置')[-1] 17 | validpart = validpart.split('listBox')[0] 18 | picurl_list = pattern.findall(validpart) 19 | return picurl_list 20 | except: 21 | print("ERROR") 22 | return ["NULL"] 23 | 24 | def store_pic(picurl_list): 25 | 26 | if "NULL" in picurl_list: 27 | return 0 28 | file_root = "D://pics//" 29 | 30 | for picurl in picurl_list: 31 | path = file_root + picurl.split('/')[-1] 32 | try: 33 | if not os.path.exists(file_root): 34 | os.mkdir(file_root) 35 | if not os.path.exists(path): 36 | pic = requests.get(picurl) 37 | with open(path,'wb') as f: 38 | f.write(pic.content) 39 | f.close() 40 | print("图片:"+picurl+" 成功下载") 41 | else: 42 | print("图片已存在") 43 | except: 44 | print("爬取失败") 45 | return 1 46 | 47 | if __name__ == '__main__': 48 | for suffix in range(800,900): 49 | store_pic(get_picture_url(suffix)) 50 | time.sleep(5) -------------------------------------------------------------------------------- /表情包爬虫.py: -------------------------------------------------------------------------------- 1 | import os 2 | from time import time 3 | 4 | import requests 5 | from bs4 import BeautifulSoup 6 | from queue import Queue 7 | from threading import Thread 8 | 9 | 10 | class DownloadBiaoqingbao(Thread): 11 | 12 | def __init__(self, queue, path): 13 | Thread.__init__(self) 14 | self.queue = queue 15 | self.path = '/home/wistbean/biaoqingbao/' 16 | if not os.path.exists(path): 17 | os.makedirs(path) 18 | 19 | def run(self): 20 | while True: 21 | url = self.queue.get() 22 | try: 23 | # print(url) 24 | download_biaoqingbaos(url, self.path) 25 | finally: 26 | self.queue.task_done() 27 | 28 | 29 | def download_biaoqingbaos(url, path): 30 | 31 | response = requests.get(url) 32 | soup = BeautifulSoup(response.content, 'lxml') 33 | img_list = soup.find_all('img', class_='ui image lazy') 34 | 35 | for img in img_list: 36 | image = img.get('data-original') 37 | title = img.get('title') 38 | print('下载图片: ', title) 39 | 40 | try: 41 | with open(path + title + os.path.splitext(image)[-1], 'wb') as f: 42 | img = requests.get(image).content 43 | f.write(img) 44 | except OSError: 45 | print('length failed') 46 | break 47 | 48 | 49 | if __name__ == '__main__': 50 | 51 | start = time() 52 | 53 | # 构建所有的链接 54 | _url = 'https://fabiaoqing.com/biaoqing/lists/page/{page}.html' 55 | urls = [_url.format(page=page) for page in range(1, 4328+1)] 56 | 57 | queue = Queue() 58 | path = '/home/wistbean/biaoqingbao/' 59 | 60 | # 创建线程 61 | for x in range(10): 62 | worker = DownloadBiaoqingbao(queue, path) 63 | worker.daemon = True 64 | worker.start() 65 | 66 | # 加入队列 67 | for url in urls: 68 | queue.put(url) 69 | 70 | queue.join() 71 | 72 | print('下载完毕耗时: ', time()-start) -------------------------------------------------------------------------------- /豆瓣电影排名爬虫.py: -------------------------------------------------------------------------------- 1 | import requests 2 | from bs4 import BeautifulSoup 3 | import xlwt 4 | 5 | 6 | def request_douban(url): 7 | try: 8 | response = requests.get(url) 9 | if response.status_code == 200: 10 | return response.text 11 | except requests.RequestException: 12 | return None 13 | 14 | 15 | book = xlwt.Workbook(encoding='utf-8', style_compression=0) 16 | 17 | sheet = book.add_sheet('豆瓣电影Top250', cell_overwrite_ok=True) 18 | sheet.write(0, 0, '名称') 19 | sheet.write(0, 1, '图片') 20 | sheet.write(0, 2, '排名') 21 | sheet.write(0, 3, '评分') 22 | sheet.write(0, 4, '作者') 23 | sheet.write(0, 5, '简介') 24 | 25 | n = 1 26 | 27 | 28 | def save_to_excel(soup): 29 | list = soup.find(class_='grid_view').find_all('li') 30 | 31 | for item in list: 32 | item_name = item.find(class_='title').string 33 | item_img = item.find('a').find('img').get('src') 34 | item_index = item.find(class_='').string 35 | item_score = item.find(class_='rating_num').string 36 | item_author = item.find('p').text 37 | if (item.find(class_='inq') != None): 38 | item_intr = item.find(class_='inq').string 39 | 40 | # print('爬取电影:' + item_index + ' | ' + item_name +' | ' + item_img +' | ' + item_score +' | ' + item_author +' | ' + item_intr ) 41 | print('爬取电影:' + item_index + ' | ' + item_name + ' | ' + item_score + ' | ' + item_intr) 42 | 43 | global n 44 | 45 | sheet.write(n, 0, item_name) 46 | sheet.write(n, 1, item_img) 47 | sheet.write(n, 2, item_index) 48 | sheet.write(n, 3, item_score) 49 | sheet.write(n, 4, item_author) 50 | sheet.write(n, 5, item_intr) 51 | 52 | n = n + 1 53 | 54 | 55 | def main(page): 56 | url = 'https://movie.douban.com/top250?start=' + str(page * 25) + '&filter=' 57 | html = request_douban(url) 58 | soup = BeautifulSoup(html, 'lxml') 59 | save_to_excel(soup) 60 | 61 | 62 | if __name__ == '__main__': 63 | 64 | for i in range(0, 10): 65 | main(i) 66 | 67 | book.save(u'豆瓣最受欢迎的250部电影.csv') -------------------------------------------------------------------------------- /豆瓣电影爬虫(GUI界面)/main.py: -------------------------------------------------------------------------------- 1 | from DoubanMovie import uiObject 2 | 3 | # main入口 4 | if __name__ == '__main__': 5 | 6 | ui = uiObject() 7 | ui.ui_process() -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/.idea/inspectionProfiles/profiles_settings.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 6 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/.idea/misc.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/.idea/modules.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/.idea/workspace.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/.idea/链家爬虫.iml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_anting.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_anting.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_fengzhuang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_fengzhuang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_huating.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_huating.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiadinglaocheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiadinglaocheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiadingxincheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiadingxincheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiangqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_jiangqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_juyuanxinqu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_juyuanxinqu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_malu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_malu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_nanxiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_nanxiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_shangda.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_shangda.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_waigang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_waigang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_xinchenglu1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_xinchenglu1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_xuxing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/嘉定/jiading_xuxing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_fengcheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_fengcheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_fengxianjinhui.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_fengxianjinhui.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_haiwan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_haiwan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_nanqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_nanqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_qingcun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_qingcun.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_situan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_situan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_xidu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_xidu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_zhelin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_zhelin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_zhuanghang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/奉贤/fengxian_zhuanghang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_dachangzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_dachangzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_dahua.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_dahua.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gongfu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gongfu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gongkang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gongkang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gucun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_gucun.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_luodian.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_luodian.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_luojing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_luojing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_songbao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_songbao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_songnan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_songnan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_tonghe.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_tonghe.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_yanghang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_yanghang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_yuepu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_yuepu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_zhangmiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/宝山/baoshan_zhangmiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/上海各区小区信息汇总.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/上海各区小区信息汇总.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/嘉定小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/嘉定小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/奉贤小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/奉贤小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/宝山小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/宝山小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/崇明小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/崇明小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/徐家汇小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/徐家汇小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/普陀小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/普陀小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/杨浦小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/杨浦小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/松江小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/松江小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/浦东小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/浦东小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/虹口小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/虹口小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/金山小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/金山小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/长宁区小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/长宁区小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/闵行小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/闵行小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/青浦小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/青浦小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/静安小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/静安小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/黄埔小区信息.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/小区信息/黄埔小区信息.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_baozhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_baozhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_changxingdao21211.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_changxingdao21211.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chenjiazhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chenjiazhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chongmingqita.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chongmingqita.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chongmingxincheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_chongmingxincheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_hengshadao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/崇明/chongming_hengshadao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_caohejing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_caohejing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_changqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_changqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_hengshanlu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_hengshanlu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_huadongligong.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_huadongligong.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_huajing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_huajing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_jianguoxilu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_jianguoxilu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_kangjian.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_kangjian.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_longhua.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_longhua.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_shanghainanzhan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_shanghainanzhan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_tianlin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_tianlin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_wantiguan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_wantiguan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xietulu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xietulu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xuhuibinjiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xuhuibinjiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xujiahui.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_xujiahui.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_zhiwuyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/徐家汇/xuhui_zhiwuyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_caoyang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_caoyang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changfeng1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changfeng1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changshoulu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changshoulu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changzheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_changzheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_ganquanyichuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_ganquanyichuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_guangxin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_guangxin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_taopu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_taopu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_wanli.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_wanli.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_wuning.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_wuning.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhenguang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhenguang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhenru.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhenru.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhongyuanliangwancheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/普陀/putuo_zhongyuanliangwancheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_anshan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_anshan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_dongwaitan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_dongwaitan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_gaojing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_gaojing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_huangxinggongyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_huangxinggongyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_kongjianglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_kongjianglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_wujiaochang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_wujiaochang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_xinjiangwancheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_xinjiangwancheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_zhongyuan1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_zhongyuan1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_zhoujiazuilu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/杨浦/yangpu_zhoujiazuilu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_chedun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_chedun.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_jiuting.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_jiuting.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_maogang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_maogang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_shenminbieshu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_shenminbieshu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_sheshan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_sheshan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_shihudang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_shihudang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_sijing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_sijing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjiangdaxuecheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjiangdaxuecheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjianglaocheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjianglaocheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjiangxincheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_songjiangxincheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xiaokunshan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xiaokunshan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xinbang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xinbang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xinqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_xinqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_yexie.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/松江/songjiang_yexie.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_beicai.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_beicai.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_biyun.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_biyun.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_caolu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_caolu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_chuansha.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_chuansha.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_datuanzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_datuanzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_gaodong.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_gaodong.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_gaohang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_gaohang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_geqing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_geqing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_hangtou.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_hangtou.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_huamu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_huamu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_huinan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_huinan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_jinqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_jinqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_jinyang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_jinyang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_kangqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_kangqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_laogangzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_laogangzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lianyang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lianyang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lingangxincheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lingangxincheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lujiazui.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_lujiazui.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_nanmatou.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_nanmatou.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_nichengzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_nichengzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_sanlin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_sanlin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_shibo.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_shibo.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_shuyuanzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_shuyuanzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_tangqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_tangqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_tangzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_tangzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_waigaoqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_waigaoqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_wanxiangzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_wanxiangzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_weifang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_weifang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_xinchang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_xinchang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_xuanqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_xuanqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yangdong.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yangdong.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yangjing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yangjing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yuanshen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yuanshen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yuqiao1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_yuqiao1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhangjiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhangjiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhoupu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhoupu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhuqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/浦东/pudong_zhuqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_beiwaitan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_beiwaitan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_jiangwanzhen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_jiangwanzhen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_liangcheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_liangcheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_linpinglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_linpinglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_luxungongyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_luxungongyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_quyang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_quyang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_sichuanbeilu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/虹口/hongkou_sichuanbeilu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_caojing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_caojing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_fengjing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_fengjing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_jinshan1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_jinshan1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_langxia.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_langxia.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_luxiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_luxiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_shanyang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_shanyang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_shihua.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_shihua.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_tinglin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_tinglin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_zhangyan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_zhangyan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_zhujing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/金山/jinshan_zhujing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_beixinjing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_beixinjing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_gubei.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_gubei.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_hongqiao1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_hongqiao1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_tianshan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_tianshan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xianxia.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xianxia.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xijiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xijiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xinhualu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_xinhualu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_zhenninglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_zhenninglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_zhongshangongyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/长宁/changning_zhongshangongyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_chunshen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_chunshen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_gumei.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_gumei.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_hanghua.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_hanghua.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_huacao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_huacao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinganxincheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinganxincheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinhongqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinhongqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinhui.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_jinhui.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_laominhang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_laominhang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_longbai.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_longbai.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_maqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_maqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_meilong.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_meilong.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_pujiang1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_pujiang1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_qibao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_qibao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_wujing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_wujing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_xinzhuang5.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_xinzhuang5.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_zhuanqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/闵行/minhang_zhuanqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_baihe.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_baihe.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_chonggu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_chonggu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_huaxin.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_huaxin.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_jinze.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_jinze.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_liantang1.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_liantang1.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xianghuaqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xianghuaqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xiayang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xiayang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xujing.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_xujing.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_yingpu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_yingpu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_zhaoxiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_zhaoxiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_zhujiajiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/青浦/qingpu_zhujiajiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_buyecheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_buyecheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_caojiadu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_caojiadu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_daning.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_daning.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_jiangninglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_jiangninglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_jingansi.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_jingansi.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_nanjingxilu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_nanjingxilu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_pengpu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_pengpu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_xizangbeilu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_xizangbeilu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_yangcheng.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_yangcheng.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_yonghe.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_yonghe.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_zhabeigongyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/静安/jingan_zhabeigongyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_dapuqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_dapuqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_dongjiadu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_dongjiadu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_huaihaizhonglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_huaihaizhonglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_huangpubinjiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_huangpubinjiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_laoximen.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_laoximen.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_nanjingdonglu.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_nanjingdonglu.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_penglaigongyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_penglaigongyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_renminguangchang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_renminguangchang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_shibobinjiang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_shibobinjiang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_wuliqiao.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_wuliqiao.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_xintiandi.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_xintiandi.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_yuyuan.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/20191030爬取链家上海小区信息/黄埔/huangpu_yuyuan.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJNew_bouse.py: -------------------------------------------------------------------------------- 1 | from bs4 import BeautifulSoup 2 | import requests 3 | import time 4 | import pandas as pd 5 | 6 | url = 'https://m.lianjia.com/nt/xiaoqu/' 7 | headers = {'Referer': url, 8 | 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0', 9 | 'Cookie': 'lianjia_uuid=443bb6e3-b556-47ea-967d-5444ea915dc6; Hm_lvt_9152f8221cb6243a53c83b956842be8a=1571809948,1572075513,1572091757,1572092040; _smt_uid=5dafea9d.5470d44f; UM_distinctid=16df72c7718be-086f464eb904c-4c312373-144000-16df72c7719794; _jzqa=1.3017324979856383500.1571809949.1572094819.1572175751.7; _jzqy=1.1571809949.1572091757.1.jzqsr=baidu|jzqct=%E9%93%BE%E5%AE%B6.-; sensorsdata2015jssdkcross=%7B%22distinct_id%22%3A%2216df72c7a34233-04d3d7bf7d75b18-4c312373-1327104-16df72c7a35214%22%2C%22%24device_id%22%3A%2216df72c7a34233-04d3d7bf7d75b18-4c312373-1327104-16df72c7a35214%22%2C%22props%22%3A%7B%22%24latest_traffic_source_type%22%3A%22%E7%9B%B4%E6%8E%A5%E6%B5%81%E9%87%8F%22%2C%22%24latest_referrer%22%3A%22%22%2C%22%24latest_referrer_host%22%3A%22%22%2C%22%24latest_search_keyword%22%3A%22%E6%9C%AA%E5%8F%96%E5%88%B0%E5%80%BC_%E7%9B%B4%E6%8E%A5%E6%89%93%E5%BC%80%22%2C%22%24latest_utm_source%22%3A%22baidu%22%2C%22%24latest_utm_medium%22%3A%22pinzhuan%22%2C%22%24latest_utm_campaign%22%3A%22sousuo%22%2C%22%24latest_utm_content%22%3A%22biaotimiaoshu%22%2C%22%24latest_utm_term%22%3A%22biaoti%22%7D%7D; _ga=GA1.2.1379447899.1571809952; _jzqx=1.1571814886.1572175751.4.jzqsr=sh%2Elianjia%2Ecom|jzqct=/ershoufang/pg100/.jzqsr=nt%2Elianjia%2Ecom|jzqct=/xiaoqu/8745128002049017/; select_city=320600; Hm_lpvt_9152f8221cb6243a53c83b956842be8a=1572177857; _jzqc=1; _gid=GA1.2.487758797.1572075516; srcid=eyJ0Ijoie1wiZGF0YVwiOlwiZjQ0ZDY4YmQzYWFiMTczNTkxNTgyYjBhNzkxMzc5N2I3ZTdmMjlhZTE2NzRlZTU4NjRlZjMyN2Q4MDVhZDdhMTRiOGJhNjNjMmEwYjNiZmIwMjI2ZGM5YWY5ZjQxMDA1MzkwYmJiMWFiMjBlYTJiNTkxMWY1ZmM4Y2ViYWRlOTg4NGQ5YjVhNWZhZTM2ZGRmMjJmZGQ3ZTAyMzUwNzM5NDc4OTNmMWQ3ZWM2ZWJkYzc1MjA2ZDRlZDE2YzA5MGQxYjVkZmRhNzNhYWI5NDMwYzUwYjIzNDI2ZmM5NTdkOTFjZTg0OTZhM2EzYzkwMmE5YzQwYjkxN2JlZWRlMmNmNzYxNGM0M2FlYTFlNjkyOGJkYzJhM2Q1M2ZhNmYxZTkyMjM0MjVmYTdmYTBhMGE4MTQ5NGI1MGY1Yjc5ZTJmNzMxNjUwMGE5NzdlMjZlMDA3YmU4Njg1NzliODQxNjIxZFwiLFwia2V5X2lkXCI6XCIxXCIsXCJzaWduXCI6XCI2NmFmZjk1M1wifSIsInIiOiJodHRwczovL20ubGlhbmppYS5jb20vbnQveGlhb3F1LyIsIm9zIjoid2ViIiwidiI6IjAuMSJ9; CNZZDATA1254525948=1706170696-1572075705-%7C1572172951; CNZZDATA1253491255=1450175482-1572074448-%7C1572172565; lianjia_ssid=91c55eb0-d2cf-7ad8-de18-58a73ecf6aab; _jzqckmp=1'} 10 | page = ('pg') 11 | 12 | # return demjson.encode(res) 13 | 14 | 15 | 16 | 17 | def areainfo(url): 18 | page = ('pg') 19 | for i in range(1, 31): # 获取1-100页的数据 20 | if i == 1: 21 | i = str(i) 22 | a = (url + page + i + '/') 23 | r = requests.get(url=a, headers=headers) 24 | print(a) 25 | htmlinfo = r.content 26 | else: 27 | i = str(i) 28 | a = (url + page + i + '/') 29 | print(a) 30 | r = requests.get(url=a, headers=headers) 31 | html2 = r.content 32 | htmlinfo = htmlinfo + html2 33 | time.sleep(0.5) 34 | return htmlinfo 35 | 36 | 37 | hlist = [] 38 | 39 | 40 | def listinfo(listhtml): 41 | areasoup = BeautifulSoup(listhtml, 'html.parser') 42 | ljhouse = areasoup.find_all('a', attrs={'class': 'pictext'}) 43 | 44 | for house in ljhouse: 45 | loupantitle = house.find("div", attrs={"class": "xiaoqu_head_title lazyload_ulog"}) 46 | loupanname = loupantitle.a.get_text() 47 | # loupantag = loupantitle.find_all("span") 48 | # wuye = loupantag[0].get_text() 49 | # xiaoshouzhuangtai = loupantag[1].get_text() 50 | location = house.find("div", attrs={"class": "xiaoqu_head_address"}).get_text() 51 | a = house.find("div", attrs={"class": "mod_box jichuxinxi"}) 52 | niandai=a.find("div", attrs={"class": "mod_cont"})[0].get_text() 53 | leixing=a.find("div", attrs={"class": "mod_cont"})[1].get_text() 54 | loudongshu=a.find("div", attrs={"class": "mod_cont"})[2].get_text() 55 | fangwushu=a.find("div", attrs={"class": "mod_cont"})[1].get_text() 56 | # price = house.find("div", attrs={"class": "main-price"}).get_text() 57 | # total = jiage.find("div", attrs={"class": "second"}) 58 | # totalprice = "暂无" 59 | # if total is not None: 60 | # totalprice = total.get_text() 61 | h = {'title': loupanname, 'location': location.replace("\n", ""), 62 | 'niandai': niandai.replace("\n", ""), 'leixing': leixing, 'loudongshu': loudongshu, 'fangwushu': fangwushu}; 63 | hlist.append(h) 64 | 65 | 66 | if __name__ == '__main__': 67 | url = url 68 | hlist.append( 69 | {'title': "楼盘名称", 'location': "位置", 70 | 'niandai': "建筑年代", 'leixing': "房屋类型", 'loudongshu': "楼栋数", 71 | 'fangwushu': "房屋数"}) 72 | areahtml = areainfo(url) 73 | listinfo(areahtml) 74 | # houseinfo = houseinfo.append(hlist) 75 | houseinfo = pd.DataFrame(hlist) 76 | houseinfo.to_csv('链家新房.csv', index=False, encoding="utf_8_sig") 77 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_0.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_0.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_2.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_3.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_3.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_4.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_4.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_5.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_5.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_6.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_6.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_7.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/output_5_7.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/链家二手房数据分析.pdf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJershoufang/链家二手房数据分析.pdf -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_0.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_0.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_2.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_3.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_3.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_4.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_4.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_5.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_5.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_6.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_6.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_7.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/output_5_7.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/链家二手房数据分析.pdf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJsecond_house/链家二手房数据分析.pdf -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJzufang.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/LJzufang.csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/LJzufang.py: -------------------------------------------------------------------------------- 1 | import requests 2 | import csv 3 | import random 4 | import time 5 | import socket 6 | import http.client 7 | from bs4 import BeautifulSoup 8 | import re 9 | 10 | 11 | def get_content(url, data=None): 12 | header = { 13 | 'Accept': 'application/json, text/javascript, */*; q=0.01', 14 | 'Accept-Encoding': 'gzip, deflate, br', 15 | 'Accept-Language': 'zh-CN,zh;q=0.9', 16 | 'Connection': 'keep-alive', 17 | 'Cookie': 'lianjia_uuid=8290c333-59db-490c-8808-8b2645f848c6; lianjia_ssid=55ca6233-79ad-4e5a-b366-831c546fe02e; srcid=eyJ0Ijoie1wiZGF0YVwiOlwiMjJmMmRhYzY2OTRjZTk2NDdjY2JlNDRiYTFhZTc1MDA0MjA3OTYwYTk2ZjlkZDE4MzFiYmJkZmEyNDc5MjhhZjU1NjZkYTJhMTU1NzkzNWU2M2IwYzY0ZjgzN2UwMDY4YzZiYTA3MWJkMzQ5MDc0MmI4NzU3YTY0MDhiNTFkMDc2MzhhNjI0MjI3YzBhNzk5YjYzYjg3MDE5ODM1ZjRlMWQ1ZDljNDBiMzczN2Q5MWQ1M2ZmMTQxYTZmNmE3MjQzNDBiZDk3YWI3MGVkMzdkM2FjYTQ3ZmViZjBmOWU1OTY3MDk1MmQ2OTgxMmQ4MmZkNjY5MzY5MjRhY2JmNTQwYzA3ZWMyMjA0MDBiNmQ5MDY5ZDZkYzQ2MTU2ODYwNTg1NjYxODljYTFkOTE3MDFlOWVkZTY2ZDllMWJiNjZlMGVmNmFmMGMyYjJkYThlNGFjYzhiNTY1YjY0NDFkNjhiYVwiLFwia2V5X2lkXCI6XCIxXCIsXCJzaWduXCI6XCI3MzE0NjlkOVwifSIsInIiOiJodHRwczovL3NoLmxpYW5qaWEuY29tL3p1ZmFuZy8iLCJvcyI6IndlYiIsInYiOiIwLjEifQ==', 18 | 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/77.0.3865.120 Safari/537.36' 19 | } 20 | timeout = random.choice(range(80, 180)) 21 | while True: 22 | try: 23 | req = requests.get(url, headers=header, timeout=timeout) 24 | req.encoding = 'utf-8' 25 | break 26 | 27 | except socket.timeout as e: 28 | print('3:', e) 29 | time.sleep(random.choice(range(8, 15))) 30 | 31 | except socket.error as e: 32 | print('4:', e) 33 | time.sleep(random.choice(range(20, 60))) 34 | 35 | except http.client.BadStatusLine as e: 36 | print('5:', e) 37 | time.sleep(random.choice(range(30, 80))) 38 | 39 | except http.client.IncompleteRead as e: 40 | print('6:', e) 41 | time.sleep(random.choice(range(5, 15))) 42 | return req.text 43 | 44 | 45 | def get_data(html_text): 46 | final = [] 47 | bs = BeautifulSoup(html_text, "html.parser") 48 | body = bs.body 49 | data = body.find('div', {'id': 'content'}).find('div', {'class': 'content__article'}) 50 | total = data.find('p', {'class': 'content__title'}).find('span', {'class': 'content__title--hl'}).string 51 | items = data.find('div', {'class': 'content__list'}).find_all('div', {'class': 'content__list--item'}) 52 | 53 | for item in items: 54 | temp = [] 55 | title = item.find('p', {'class': 'twoline'}).find('a').string 56 | price = str(item.find('span', {'class': 'content__list--item-price'}).text) 57 | infostr = str(item.find('p', {'class': 'content__list--item--des'}).text) 58 | try: 59 | type = str(item.find('p', {'class': 'content__list--item--brand'}).text) 60 | except: 61 | type = '' 62 | time = str(item.find('p', {'class': 'content__list--item--time'}).text) 63 | tag = str(item.find('p', {'class': 'content__list--item--bottom'}).text) 64 | title = re.sub(r'[\[\]\s]', '', title) 65 | infostr = re.sub(r'\s', '', infostr) 66 | info = infostr.split('/') 67 | type = re.sub(r'\s', '', type) 68 | tag = re.sub(r'\s', '', tag) 69 | address = info[0] 70 | size = info[1] 71 | fangxiang = info[2] 72 | format = info[3] 73 | # floor = info[4] 74 | temp.append(title) 75 | temp.append(price) 76 | temp.append(address) 77 | temp.append(size) 78 | temp.append(fangxiang) 79 | temp.append(format) 80 | # temp.append(floor) 81 | temp.append(type) 82 | temp.append(time) 83 | temp.append(tag) 84 | final.append(temp) 85 | return final 86 | 87 | 88 | def write_data(data, name): 89 | file_name = name 90 | 91 | with open(file_name, 'a', errors='ignore', newline='') as f: 92 | f_csv = csv.writer(f) 93 | f_csv.writerows([['title', 'price', 'address', 'size', 'fangxiang', 'format', 'type', 'time', 'tag']]) 94 | f_csv.writerows(data) 95 | 96 | 97 | if __name__ == '__main__': 98 | url = 'https://sh.lianjia.com/zufang/' 99 | # https://sh.lianjia.com/zufang/pujiang1/pg2rt200600000001l0/ 100 | result = [] 101 | for i in range(0, 99): 102 | html = get_content(url) 103 | d = get_data(html) 104 | result.extend(d) 105 | url = 'https://sh.lianjia.com/zufang/pg' + str(i+1) + '/#contentList' 106 | 107 | write_data(result, 'LJzufang.csv') -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/lianjiasecond_house.py: -------------------------------------------------------------------------------- 1 | import os 2 | import pandas as pd 3 | from selenium import webdriver 4 | from pyquery import PyQuery as pq 5 | driver = webdriver.Chrome() 6 | 7 | 8 | ''' 9 | 步骤: 10 | 1. 下载页面 11 | 2. 解析页面 12 | 3. 保存数据 13 | ''' 14 | 15 | 16 | def cached_url(url): 17 | """ 18 | 缓存网页, 避免重复下载浪费时间 19 | """ 20 | folder = 'cached' 21 | filename = url.split('/')[-2][2:] + '.html' # 以页面数命名 html 页面 22 | path = os.path.join(folder, filename) 23 | if os.path.exists(path): 24 | with open(path, 'rb') as f: 25 | s = f.read() 26 | return s 27 | else: 28 | if not os.path.exists(folder): 29 | # 建立 cached 文件夹 30 | os.makedirs(folder) 31 | 32 | driver.get(url) # 使用 Selenium 爬取页面 33 | with open(path, 'wb') as f: 34 | f.write(driver.page_source.encode()) 35 | return driver.page_source 36 | 37 | 38 | def house_from_div(div): 39 | """ 40 | 从一个 div 里面获取到一个二手房信息 41 | """ 42 | e = pq(div) 43 | 44 | # 小作用域变量用单字符 45 | m = {} 46 | m['name'] = e('.houseInfo').text() # 二手房名字 47 | m['price'] = e('.totalPrice').text() # 二手房总价 48 | m['unitprice'] = e('.unitPrice').text() # 二手房每平米单价 49 | m['position'] = e('.positionInfo').text() # 二手房位置 50 | m['follow'] = e('.followInfo').text() # 二手房关注信息 51 | m['url'] = e('a').attr('href') # 二手房页面链接 52 | 53 | return m 54 | 55 | 56 | def houses_from_url(url): 57 | """ 58 | 从 url 中下载网页并解析出页面内所有的房源 59 | """ 60 | # 页面只需要下载一次 61 | page = cached_url(url) 62 | e = pq(page) 63 | items = e('.info.clear') # 解析的class为info clear, 中间有空格, 使用“.info.clear” 64 | # 调用 house_from_div 65 | houses = [house_from_div(i) for i in items] 66 | return houses 67 | 68 | 69 | def append_to_csv(data): 70 | ''' 71 | 保存数据 72 | ''' 73 | file_name = './新房数据.csv' 74 | df = pd.DataFrame(data) 75 | df.to_csv(file_name, mode='a', encoding='gbk', header=False, index=False) 76 | 77 | 78 | def main(): 79 | for i in range(1, 101): 80 | # 一共 100 页房源信息 81 | url = 'https://sh.fang.lianjia.com/loupan/pg{}/'.format(i) 82 | houses = houses_from_url(url) 83 | print(houses) 84 | append_to_csv(houses) 85 | 86 | 87 | if __name__ == '__main__': 88 | main() -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/.gitignore: -------------------------------------------------------------------------------- 1 | *.pyc 2 | .idea/ 3 | data/ 4 | *.html 5 | *.csv 6 | log/log.txt 7 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/README.md: -------------------------------------------------------------------------------- 1 | # 链家网(lianjia.com)和贝壳网(ke.com)爬虫 2 | - 爬取链家网、贝壳网的各类房价数据(小区数据,挂牌二手房, 出租房,新房)。 3 | - **如果好用,请点星支持 !** 4 | - 支持北京上海广州深圳等国内21个主要城市;支持Python2和Python3; 基于页面的数据爬取,稳定可靠; 丰富的代码注释,帮助理解代码并且方便扩展功能。 5 | - 数据含义:城市-city, 区县-district, 板块-area, 小区-xiaoqu, 二手房-ershou, 租房-zufang, 新房-loupan。 6 | - 每个版块存储为一个csv文件,该文件可以作为原始数据进行进一步的处理和分析。 7 | - 支持图表展示。 8 | ![alt text](https://github.com/jumper2014/lianjia-spider/blob/master/pic/xiaoqu_top.png) 9 | ![alt text](https://github.com/jumper2014/lianjia-spider/blob/master/pic/district_top.png) 10 | - 如果链家和贝壳页面结构有调整,欢迎反馈,我将尽力保持更新。 11 | - 此代码仅供学习与交流,请勿用于商业用途,后果自负。 12 | 13 | ## 安装依赖 14 | - pip install -r requirements.txt 15 | - 运行前,请将当前目录加入到系统环境变量PYTHONPATH中。 16 | - 运行前,请指定要爬取的网站,见lib/spider/base_spider.py里面的SPIDER_NAME变量。 17 | - 清理数据,运行 python tool/clean.py 18 | 19 | ## 快速问答 20 | - Q: 如何降低爬取速度,避免被封IP?A:见base_spider.py里面的RANDOM_DELAY 21 | - Q: 如何减少并发的爬虫数? A: 见见base_spider.py的thread_pool_size 22 | - Q: 为何无法使用xiaoqu_to_chart.py? A: 该脚本现仅支持mac系统 23 | - Q: 有其他问题反馈途径么? A: 问题反馈QQ群号635276285。 24 | 25 | ## 小区房价数据爬取 26 | - 内容格式:采集日期,所属区县,板块名,小区名,挂牌均价,挂牌数 27 | - 内容如下:20180221,浦东,川沙,恒纬家苑,32176元/m2,3套在售二手房 28 | - 数据可以存入MySQL/MongoDB数据库,用于进一步数据分析,比如排序,计算区县和版块均价。 29 | - MySQL数据库结构可以通过导入tool/lianjia_xiaoqu.sql建立。 30 | - MySQL数据格式: 城市 日期 所属区县 版块名 小区名 挂牌均价 挂牌数 31 | - MySQL数据内容:上海 20180331 徐汇 衡山路 永嘉路621号 333333 0 32 | - MongoDB数据内容: { "_id" : ObjectId("5ac0309332e3885598b3b751"), "city" : "上海", "district" : "黄浦", "area" : "五里桥", "date" : "20180331", "price" : 81805, "sale" : 11, "xiaoqu" : "桥一小区" } 33 | - Excel数据内容:上海 20180331 徐汇 衡山路 永嘉路621号 333333 0 34 | - 运行, python xiaoqu.py 根据提示输入城市代码,回车确认,开始采集数据到csv文件 35 | - 运行, python xiaoqu.py city, 自动开始采集数据到csv文件 36 | ``` 37 | hz: 杭州, sz: 深圳, dl: 大连, fs: 佛山 38 | xm: 厦门, dg: 东莞, gz: 广州, bj: 北京 39 | cd: 成都, sy: 沈阳, jn: 济南, sh: 上海 40 | tj: 天津, qd: 青岛, cs: 长沙, su: 苏州 41 | cq: 重庆, wh: 武汉, hf: 合肥, yt: 烟台 42 | nj: 南京, 43 | ``` 44 | - 修改 xiaoqu_to_db.py 中的database变量,设置数据最终存入mysql/mongodb/Excel/json 45 | - python xiaoqu_to_db.py 根据提示将今天采集到的csv数据存入数据库。(默认导出为单一csv文件) 46 | - python xiaoqu_to_chart.py 将单一csv文件数据通过图表展示。 47 | 48 | ## 挂牌二手房数据爬取 49 | - 获取链家网挂牌二手房价数据,数据格式如下: 50 | - 20180405,浦东,万祥镇,祥安菊苑 3室2厅 258万,258万,祥安菊苑 | 3室2厅 | 126.58平米 | 南 | 毛坯 51 | - 运行,python ershou.py 根据提示输入城市代码,回车确认,开始采集数据到csv文件 52 | - 运行,python ershou.py city,自动开始采集数据到csv文件 53 | 54 | 55 | ## 出租房数据爬取 56 | - 获取链家网挂牌出租房数据,数据格式如下: 57 | - 20180407,浦东,御桥,仁和都市花园  ,3室2厅,100平米,8000 58 | - 运行,python zufang.py 根据提示输入城市代码,回车确认,开始采集数据到csv文件 59 | - 运行,python zufang.py city,自动开始采集数据到csv文件 60 | 61 | ## 新房数据爬取 62 | - 获取链家网新房数据,数据格式如下: 63 | - 20180407,上海星河湾,76000,1672万 64 | - 运行,python loupan.py 根据提示输入城市代码,回车确认,开始采集数据到csv文件 65 | - 运行,python loupan.py city,自动开始采集数据到csv文件 66 | 67 | ## 结果存储 68 | - 根目录下建立data目录存放结果数据文件 69 | - 小区房价数据存储目录为 data/site/xiaoqu/city/date 70 | - 二手房房价数据存储目录为 data/site/ershou/city/date 71 | - 出租房房价数据存储目录为 data/site/zufang/city/date 72 | - 新房房价数据存储目录为 data/site/loupan/city/date 73 | 74 | ## 性能 75 | - 300秒爬取上海市207个版块的2.7万条小区数据,平均每秒90条数据。 76 | ``` 77 | Total crawl 207 areas. 78 | Total cost 294.048109055 second to crawl 27256 data items. 79 | ``` 80 | - 1000秒爬取上海215个版块的7.5万条挂牌二手房数据,平均每秒75条数据。 81 | ``` 82 | Total crawl 215 areas. 83 | Total cost 1028.3090899 second to crawl 75448 data items. 84 | ``` 85 | - 300秒爬取上海215个版块的3.2万条出租房数据, 平均每秒150条数据。 86 | ``` 87 | Total crawl 215 areas. 88 | Total cost 299.7534770965576 second to crawl 32735 data items. 89 | ``` 90 | - 30秒爬取上海400个新盘数据。 91 | ``` 92 | Total crawl 400 loupan. 93 | Total cost 29.757128953933716 second 94 | ``` 95 | 96 | 97 | 98 | ### 更新记录 99 | - 2019/06/21 去除requirements.txt中的webbrower 100 | - 2018/11/05 增加工具下载二手房缩略图tool/download_ershou_image.py 101 | - 2018/11/01 增加二手房缩略图地址 102 | - 2018/10/28 xiaoqu_to_db.py改造成支持命令行参数自动运行。 103 | - 2018/10/25 将主要爬取代码抽取到spider类中。 104 | - 2018/10/22 文件名,目录,代码重构。 105 | - 2018/10/20 增加中间文件清理功能,能够爬取贝壳网的小区,新房,二手房和租房数据。 106 | - 2018/10/19 支持贝壳网小区数据爬取 107 | - 2018/10/15 增加Spider类,优化异常处理,功能无变动 108 | - 2018/10/14 允许用户通过命令行指定要爬取的城市,而不仅仅通过交互模式选择,用于支持自动爬取。 109 | - 2018/10/11 增加初步log功能。 110 | - 2018/10/09 图表展示区县均价排名。 111 | - 2018/10/07 小区房价导出到json文件, csv文件。图表展示最贵的小区。 112 | - 2018/10/05 增加Referer。增加透明代理服务器获取(未使用) 113 | - 2018/06/01 支持User-Agent 114 | - 2018/04/07 支持采集新房的基本房价信息 115 | - 2018/04/07 支持采集出租房的相关信息 116 | - 2018/04/05 支持采集挂牌二手房信息 117 | - 2018/04/02 支持将采集到的csv数据导入Excel 118 | - 2018/04/01 同时支持Python2和Python3 119 | - 2018/04/01 支持将采集到的csv数据导入MongoDB数据库 120 | - 2018/03/31 支持将采集到的csv数据导入MySQL数据库 121 | - 2018/03/27 修复bug: 版块下只有一页小区数据时未能正确爬取 122 | - 2018/03/27 增加5个城市,现在支持21个城市的小区数据爬取 123 | - 2018/03/10 自动获取城市的区县列表,现在支持16个城市小区数据爬取 124 | - 2018/03/06 支持北京二手房小区数据采集 125 | - 2018/02/21 应对链家前端页面更新,使用内置urllib2代替第三方requests库,提升性能,减少依赖 126 | - 2018/02/01 支持上海二手房小区数据采集 -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/__init__.py -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/const/__init__.py: -------------------------------------------------------------------------------- 1 | # coding=utf-8 2 | # -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/const/xpath.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 页面元素的XPATH 6 | 7 | from lib.spider.base_spider import SPIDER_NAME, LIANJIA_SPIDER, BEIKE_SPIDER 8 | 9 | if SPIDER_NAME == LIANJIA_SPIDER: 10 | ERSHOUFANG_QU_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div/a' 11 | ERSHOUFANG_BANKUAI_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div[2]/a' 12 | XIAOQU_QU_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div/a' 13 | XIAOQU_BANKUAI_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div[2]/a' 14 | DISTRICT_AREA_XPATH = '//div[3]/div[1]/dl[2]/dd/div/div[2]/a' 15 | CITY_DISTRICT_XPATH = '///div[3]/div[1]/dl[2]/dd/div/div/a' 16 | elif SPIDER_NAME == BEIKE_SPIDER: 17 | ERSHOUFANG_QU_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div/a' 18 | ERSHOUFANG_BANKUAI_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div[2]/a' 19 | XIAOQU_QU_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div/a' 20 | XIAOQU_BANKUAI_XPATH = '//*[@id="filter-options"]/dl[1]/dd/div[2]/a' 21 | DISTRICT_AREA_XPATH = '//div[3]/div[1]/dl[2]/dd/div/div[2]/a' 22 | CITY_DISTRICT_XPATH = '///div[3]/div[1]/dl[2]/dd/div/div/a' 23 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/item/__init__.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | 5 | 6 | if __name__ == '__main__': 7 | pass -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/item/xiaoqu.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | class XiaoQu(object): 4 | def __init__(self, district, area, name, price, on_sale, link,niandai,leixing,wuyefei,wuyegs,kaifashang,loushu,wushu): 5 | self.district = district 6 | self.area = area 7 | self.price = price 8 | self.name = name 9 | self.on_sale = on_sale 10 | self.link=link 11 | self.niandai=niandai 12 | self.leixing=leixing 13 | self.wuyefei=wuyefei 14 | self.loushu=loushu 15 | self.wushu=wushu 16 | 17 | def text(self): 18 | return self.district + "," + \ 19 | self.area + "," + \ 20 | self.name + "," + \ 21 | self.price + "," + \ 22 | self.on_sale + "," + \ 23 | self.link + "," + \ 24 | self.niandai + "," + \ 25 | self.leixing + "," + \ 26 | self.wuyefei + "," + \ 27 | self.loushu + "," + \ 28 | self.wushu -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/request/__init__.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | 5 | 6 | if __name__ == '__main__': 7 | pass -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/request/headers.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # USER AGENTS 可以自己添加 6 | 7 | import random 8 | from lib.spider.base_spider import SPIDER_NAME 9 | 10 | USER_AGENTS = [ 11 | "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; SV1; AcooBrowser; .NET CLR 1.1.4322; .NET CLR 2.0.50727)", 12 | "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0; Acoo Browser; SLCC1; .NET CLR 2.0.50727; Media Center PC 5.0; .NET CLR 3.0.04506)", 13 | "Mozilla/4.0 (compatible; MSIE 7.0; AOL 9.5; AOLBuild 4337.35; Windows NT 5.1; .NET CLR 1.1.4322; .NET CLR 2.0.50727)", 14 | "Mozilla/5.0 (Windows; U; MSIE 9.0; Windows NT 9.0; en-US)", 15 | "Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Win64; x64; Trident/5.0; .NET CLR 3.5.30729; .NET CLR 3.0.30729; .NET CLR 2.0.50727; Media Center PC 6.0)", 16 | "Mozilla/5.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0; WOW64; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; .NET CLR 1.0.3705; .NET CLR 1.1.4322)", 17 | "Mozilla/4.0 (compatible; MSIE 7.0b; Windows NT 5.2; .NET CLR 1.1.4322; .NET CLR 2.0.50727; InfoPath.2; .NET CLR 3.0.04506.30)", 18 | "Mozilla/5.0 (Windows; U; Windows NT 5.1; zh-CN) AppleWebKit/523.15 (KHTML, like Gecko, Safari/419.3) Arora/0.3 (Change: 287 c9dfb30)", 19 | "Mozilla/5.0 (X11; U; Linux; en-US) AppleWebKit/527+ (KHTML, like Gecko, Safari/419.3) Arora/0.6", 20 | "Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.8.1.2pre) Gecko/20070215 K-Ninja/2.1.1", 21 | "Mozilla/5.0 (Windows; U; Windows NT 5.1; zh-CN; rv:1.9) Gecko/20080705 Firefox/3.0 Kapiko/3.0", 22 | "Mozilla/5.0 (X11; Linux i686; U;) Gecko/20070322 Kazehakase/0.4.5", 23 | "Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.8) Gecko Fedora/1.9.0.8-1.fc10 Kazehakase/0.5.6", 24 | "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 Safari/535.11", 25 | "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_3) AppleWebKit/535.20 (KHTML, like Gecko) Chrome/19.0.1036.7 Safari/535.20", 26 | "Opera/9.80 (Macintosh; Intel Mac OS X 10.6.8; U; fr) Presto/2.9.168 Version/11.52", 27 | ] 28 | 29 | 30 | def create_headers(): 31 | headers = dict() 32 | headers["User-Agent"] = random.choice(USER_AGENTS) 33 | headers["Referer"] = "http://www.{0}.com".format(SPIDER_NAME) 34 | return headers 35 | 36 | 37 | if __name__ == '__main__': 38 | pass 39 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/request/proxy.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 用于获取代理 6 | 7 | from bs4 import BeautifulSoup 8 | import requests 9 | from lib.request.headers import create_headers 10 | 11 | proxys_src = [] 12 | proxys = [] 13 | 14 | 15 | def spider_proxyip(num=10): 16 | try: 17 | url = 'http://www.xicidaili.com/nt/1' 18 | req = requests.get(url, headers=create_headers()) 19 | source_code = req.content 20 | print(source_code) 21 | soup = BeautifulSoup(source_code, 'lxml') 22 | ips = soup.findAll('tr') 23 | 24 | for x in range(1, len(ips)): 25 | ip = ips[x] 26 | tds = ip.findAll("td") 27 | proxy_host = "{0}://".format(tds[5].contents[0]) + tds[1].contents[0] + ":" + tds[2].contents[0] 28 | proxy_temp = {tds[5].contents[0]: proxy_host} 29 | proxys_src.append(proxy_temp) 30 | if x >= num: 31 | break 32 | except Exception as e: 33 | print("spider_proxyip exception:") 34 | print(e) 35 | 36 | 37 | if __name__ == '__main__': 38 | spider_proxyip(10) 39 | print(proxys_src) 40 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/spider/__init__.py: -------------------------------------------------------------------------------- 1 | # coding=utf-8 2 | # -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/spider/base_spider.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 爬虫基类 6 | # 爬虫名常量,用来设置爬取哪个站点 7 | 8 | import threading 9 | from lib.zone.city import lianjia_cities, beike_cities 10 | from lib.utility.date import * 11 | import lib.utility.version 12 | import random 13 | 14 | thread_pool_size = 50 15 | 16 | # 防止爬虫被禁,随机延迟设定 17 | # 如果不想delay,就设定False, 18 | # 具体时间可以修改random_delay(),由于多线程,建议数值大于10 19 | RANDOM_DELAY = False 20 | LIANJIA_SPIDER = "lianjia" 21 | BEIKE_SPIDER = "lianjia" 22 | SPIDER_NAME = LIANJIA_SPIDER 23 | #SPIDER_NAME = BEIKE_SPIDER 24 | 25 | 26 | class BaseSpider(object): 27 | @staticmethod 28 | def random_delay(): 29 | if RANDOM_DELAY: 30 | time.sleep(random.randint(0, 16)) 31 | 32 | def __init__(self, name): 33 | self.name = name 34 | if self.name == LIANJIA_SPIDER: 35 | self.cities = lianjia_cities 36 | elif self.name == BEIKE_SPIDER: 37 | self.cities = beike_cities 38 | else: 39 | self.cities = None 40 | # 准备日期信息,爬到的数据存放到日期相关文件夹下 41 | self.date_string = get_date_string() 42 | print('Today date is: %s' % self.date_string) 43 | 44 | self.total_num = 0 # 总的小区个数,用于统计 45 | print("Target site is {0}.com".format(SPIDER_NAME)) 46 | self.mutex = threading.Lock() # 创建锁 47 | 48 | def create_prompt_text(self): 49 | """ 50 | 根据已有城市中英文对照表拼接选择提示信息 51 | :return: 拼接好的字串 52 | """ 53 | city_info = list() 54 | count = 0 55 | for en_name, ch_name in self.cities.items(): 56 | count += 1 57 | city_info.append(en_name) 58 | city_info.append(": ") 59 | city_info.append(ch_name) 60 | # if count % 4 == 0: 61 | # city_info.append("\n") 62 | # else: 63 | # city_info.append(", ") 64 | return 'Which city do you want to crawl?\n' + ''.join(city_info) 65 | 66 | def get_chinese_city(self, en): 67 | """ 68 | 拼音拼音名转中文城市名 69 | :param en: 拼音 70 | :return: 中文 71 | """ 72 | return self.cities.get(en, None) 73 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/__init__.py: -------------------------------------------------------------------------------- 1 | # coding=utf-8 2 | # -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/date.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 日期和时间的字符串辅助函数 6 | 7 | import time 8 | 9 | 10 | def get_time_string(): 11 | """ 12 | 获得形如20161010120000这样的年月日时分秒字符串 13 | :return: 14 | """ 15 | current = time.localtime() 16 | return time.strftime("%Y%m%d%H%M%S", current) 17 | 18 | 19 | def get_date_string(): 20 | """ 21 | 获得形如20161010这样的年月日字符串 22 | :return: 23 | """ 24 | current = time.localtime() 25 | return time.strftime("%Y%m%d", current) 26 | 27 | 28 | def get_year_month_string(): 29 | """ 30 | 获得形如201610这样的年月字符串 31 | :return: 32 | """ 33 | current = time.localtime() 34 | return time.strftime("%Y%m", current) 35 | 36 | 37 | if __name__ == "__main__": 38 | print(get_date_string()) 39 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/log.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | 6 | import logging 7 | from lib.utility.path import LOG_PATH 8 | 9 | logger = logging.getLogger(__name__) 10 | logger.setLevel(level=logging.INFO) 11 | handler = logging.FileHandler(LOG_PATH + "/log.txt") 12 | handler.setLevel(logging.INFO) 13 | formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s') 14 | handler.setFormatter(formatter) 15 | logger.addHandler(handler) 16 | 17 | if __name__ == '__main__': 18 | pass -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/path.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 获得当前目录结构,创建目录结构 6 | 7 | 8 | import inspect 9 | import os 10 | import sys 11 | 12 | 13 | def get_root_path(): 14 | file_path = os.path.abspath(inspect.getfile(sys.modules[__name__])) 15 | parent_path = os.path.dirname(file_path) 16 | lib_path = os.path.dirname(parent_path) 17 | root_path = os.path.dirname(lib_path) 18 | return root_path 19 | 20 | 21 | def create_data_path(): 22 | root_path = get_root_path() 23 | data_path = root_path + "/data" 24 | if not os.path.exists(data_path): 25 | os.makedirs(data_path) 26 | return data_path 27 | 28 | 29 | def create_site_path(site): 30 | data_path = create_data_path() 31 | site_path = data_path + "/" + site 32 | if not os.path.exists(site_path): 33 | os.makedirs(site_path) 34 | return site_path 35 | 36 | 37 | def create_city_path(site, city): 38 | site_path = create_site_path(site) 39 | city_path = site_path + "/" + city 40 | if not os.path.exists(city_path): 41 | os.makedirs(city_path) 42 | return city_path 43 | 44 | 45 | def create_date_path(site, city, date): 46 | city_path = create_city_path(site, city) 47 | date_path = city_path + "/" + date 48 | if not os.path.exists(date_path): 49 | os.makedirs(date_path) 50 | return date_path 51 | 52 | 53 | # const for path 54 | ROOT_PATH = get_root_path() 55 | DATA_PATH = ROOT_PATH + "/data" 56 | SAMPLE_PATH = ROOT_PATH + "/sample" 57 | LOG_PATH = ROOT_PATH + "/log" 58 | 59 | if __name__ == "__main__": 60 | create_date_path() 61 | create_date_path() 62 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/version.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 判断Python版本环境 6 | 7 | 8 | import sys 9 | 10 | if sys.version_info < (3, 0): # 如果小于Python3 11 | PYTHON_3 = False 12 | else: 13 | PYTHON_3 = True 14 | 15 | if not PYTHON_3: # 如果小于Python3 16 | reload(sys) 17 | sys.setdefaultencoding("utf-8") 18 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/utility/writer.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 保存结果到文件 6 | 7 | from lib.utility.path import * 8 | 9 | 10 | def write_urls_to_file(file_name, urls): 11 | file_name = DATA_PATH + "/" + file_name 12 | txt_file = open(file_name, 'w') 13 | for url in urls: 14 | txt_file.write(url+"\n") 15 | txt_file.close() 16 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/zone/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/zone/__init__.py -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/zone/area.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 板块信息相关函数 6 | 7 | from lib.zone.district import * 8 | from lib.const.xpath import * 9 | from lib.request.headers import * 10 | from lib.spider.base_spider import SPIDER_NAME 11 | 12 | 13 | def get_district_url(city, district): 14 | """ 15 | 拼接指定城市的区县url 16 | :param city: 城市 17 | :param district: 区县 18 | :return: 19 | """ 20 | return "http://{0}.{1}.com/xiaoqu/{2}".format(city, SPIDER_NAME, district) 21 | 22 | 23 | def get_areas(city, district): 24 | """ 25 | 通过城市和区县名获得下级板块名 26 | :param city: 城市 27 | :param district: 区县 28 | :return: 区县列表 29 | """ 30 | page = get_district_url(city, district) 31 | areas = list() 32 | try: 33 | headers = create_headers() 34 | response = requests.get(page, timeout=10, headers=headers) 35 | html = response.content 36 | root = etree.HTML(html) 37 | links = root.xpath(DISTRICT_AREA_XPATH) 38 | 39 | # 针对a标签的list进行处理 40 | for link in links: 41 | relative_link = link.attrib['href'] 42 | # 去掉最后的"/" 43 | relative_link = relative_link[:-1] 44 | # 获取最后一节 45 | area = relative_link.split("/")[-1] 46 | # 去掉区县名,防止重复 47 | if area != district: 48 | chinese_area = link.text 49 | chinese_area_dict[area] = chinese_area 50 | # print(chinese_area) 51 | areas.append(area) 52 | return areas 53 | except Exception as e: 54 | print(e) 55 | 56 | 57 | if __name__ == "__main__": 58 | print(get_areas("sh", "huangpu")) 59 | 60 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/zone/city.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 城市缩写和城市名的映射 6 | # 想抓取其他已有城市的话,需要把相关城市信息放入下面的字典中 7 | # 不过暂时只有下面这些城市在链家上是统一样式 8 | 9 | import sys 10 | from lib.utility.version import PYTHON_3 11 | from lib.utility.log import * 12 | 13 | cities = { 14 | 'bj': '北京', 15 | 'cd': '成都', 16 | 'cq': '重庆', 17 | 'cs': '长沙', 18 | 'dg': '东莞', 19 | 'dl': '大连', 20 | 'fs': '佛山', 21 | 'gz': '广州', 22 | 'hz': '杭州', 23 | 'hf': '合肥', 24 | 'jn': '济南', 25 | 'nj': '南京', 26 | 'qd': '青岛', 27 | 'sh': '上海', 28 | 'sz': '深圳', 29 | 'su': '苏州', 30 | 'sy': '沈阳', 31 | 'tj': '天津', 32 | 'wh': '武汉', 33 | 'xm': '厦门', 34 | 'yt': '烟台', 35 | } 36 | 37 | 38 | lianjia_cities = cities 39 | beike_cities = cities 40 | 41 | 42 | def create_prompt_text(): 43 | """ 44 | 根据已有城市中英文对照表拼接选择提示信息 45 | :return: 拼接好的字串 46 | """ 47 | city_info = list() 48 | count = 0 49 | for en_name, ch_name in cities.items(): 50 | count += 1 51 | city_info.append(en_name) 52 | city_info.append(": ") 53 | city_info.append(ch_name) 54 | if count % 4 == 0: 55 | city_info.append("\n") 56 | else: 57 | city_info.append(", ") 58 | return 'Which city do you want to crawl?\n' + ''.join(city_info) 59 | 60 | 61 | def get_chinese_city(en): 62 | """ 63 | 拼音拼音名转中文城市名 64 | :param en: 拼音 65 | :return: 中文 66 | """ 67 | return cities.get(en, None) 68 | 69 | 70 | def get_city(): 71 | city = None 72 | # 允许用户通过命令直接指定 73 | if len(sys.argv) < 2: 74 | print("Wait for your choice.") 75 | # 让用户选择爬取哪个城市的二手房小区价格数据 76 | prompt = create_prompt_text() 77 | # 判断Python版本 78 | if not PYTHON_3: # 如果小于Python3 79 | city = raw_input(prompt) 80 | else: 81 | city = input(prompt) 82 | elif len(sys.argv) == 2: 83 | city = str(sys.argv[1]) 84 | print("City is: {0}".format(city)) 85 | else: 86 | print("At most accept one parameter.") 87 | exit(1) 88 | 89 | chinese_city = get_chinese_city(city) 90 | if chinese_city is not None: 91 | message = 'OK, start to crawl ' + get_chinese_city(city) 92 | print(message) 93 | logger.info(message) 94 | else: 95 | print("No such city, please check your input.") 96 | exit(1) 97 | return city 98 | 99 | 100 | if __name__ == '__main__': 101 | print(get_chinese_city("sh")) 102 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/lib/zone/district.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 获得各城市的区县相关信息 6 | 7 | import requests 8 | from lxml import etree 9 | from lib.zone.city import cities 10 | from lib.const.xpath import * 11 | from lib.request.headers import * 12 | from lib.spider.base_spider import SPIDER_NAME 13 | 14 | chinese_city_district_dict = dict() # 城市代码和中文名映射 15 | chinese_area_dict = dict() # 版块代码和中文名映射 16 | area_dict = dict() 17 | 18 | 19 | def get_chinese_district(en): 20 | """ 21 | 拼音区县名转中文区县名 22 | :param en: 英文 23 | :return: 中文 24 | """ 25 | return chinese_city_district_dict.get(en, None) 26 | 27 | 28 | def get_districts(city): 29 | """ 30 | 获取各城市的区县中英文对照信息 31 | :param city: 城市 32 | :return: 英文区县名列表 33 | """ 34 | url = 'https://{0}.{1}.com/xiaoqu/'.format(city, SPIDER_NAME) 35 | headers = create_headers() 36 | response = requests.get(url, timeout=10, headers=headers) 37 | html = response.content 38 | root = etree.HTML(html) 39 | elements = root.xpath(CITY_DISTRICT_XPATH) 40 | en_names = list() 41 | ch_names = list() 42 | for element in elements: 43 | link = element.attrib['href'] 44 | en_names.append(link.split('/')[-2]) 45 | ch_names.append(element.text) 46 | 47 | # 打印区县英文和中文名列表 48 | for index, name in enumerate(en_names): 49 | chinese_city_district_dict[name] = ch_names[index] 50 | # print(name + ' -> ' + ch_names[index]) 51 | return en_names 52 | 53 | 54 | if __name__ == '__main__': 55 | for key in cities.keys(): 56 | # 寻找那些网页格式不合规的城市 57 | chinese_city_district_dict = dict() 58 | get_districts(key) 59 | if len(chinese_city_district_dict.items()) == 0: 60 | print(key) 61 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/log/__init__.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | 6 | 7 | if __name__ == '__main__': 8 | pass -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/pic/district_top.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/pic/district_top.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/pic/xiaoqu_top.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/pic/xiaoqu_top.png -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/requirements.txt: -------------------------------------------------------------------------------- 1 | threadpool 2 | bs4 3 | pymysql 4 | requests 5 | lxml 6 | records 7 | pymongo 8 | xlsxwriter 9 | numpy 10 | pandas 11 | pyecharts 12 | pyecharts-snapshot -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/test/date_test.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | 6 | import unittest 7 | from lib.utility.date import * 8 | 9 | 10 | class DateTest(unittest.TestCase): 11 | def setUp(self): 12 | pass 13 | 14 | def tearDown(self): 15 | pass 16 | 17 | def test_time_string(self): 18 | self.assertEqual(len(get_time_string()), 14) 19 | 20 | def test_date_string(self): 21 | self.assertEqual(len(get_date_string()), 8) 22 | 23 | def test_year_string(self): 24 | self.assertEqual(len(get_year_month_string()), 6) 25 | 26 | 27 | if __name__ == '__main__': 28 | unittest.main() 29 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/tool/clean.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 清理结果文件 6 | 7 | from lib.utility.path import * 8 | 9 | if __name__ == '__main__': 10 | # 删除日志 11 | os.system("rm -rf {0}/*.txt".format(LOG_PATH)) 12 | 13 | # 删除爬取的数据 14 | os.system("rm -rf {0}/*".format(DATA_PATH)) 15 | 16 | # 删除HTML 17 | os.system("rm -rf {0}/*.html".format(ROOT_PATH)) 18 | 19 | # 删除csv 20 | os.system("rm -rf {0}/*.csv".format(ROOT_PATH)) 21 | 22 | # 删除json 23 | os.system("rm -rf {0}/*.json".format(ROOT_PATH)) -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/tool/ershou_image_with_coroutine.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 协程下载图片,仅支持Python 3.6.1 6 | 7 | 8 | 9 | import aiohttp 10 | import aiofiles 11 | import asyncio 12 | import os 13 | import time 14 | from lib.zone.city import get_chinese_city 15 | from lib.request.headers import create_headers 16 | from lib.utility.date import get_date_string 17 | from lib.spider.base_spider import SPIDER_NAME 18 | from lib.utility.path import DATA_PATH 19 | 20 | 21 | def get_ershou_img_urls(city): 22 | urls = list() 23 | date = get_date_string() 24 | # 获得 csv 文件路径 25 | # date = "20180331" # 指定采集数据的日期 26 | # city = "sh" # 指定采集数据的城市 27 | csv_dir = "{0}/{1}/ershou/{2}/{3}".format(DATA_PATH, SPIDER_NAME, city, date) 28 | 29 | files = list() 30 | if not os.path.exists(csv_dir): 31 | print("{0} does not exist.".format(csv_dir)) 32 | print("Please run 'python ershou.py' firstly.") 33 | print("Bye.") 34 | exit(0) 35 | else: 36 | print('OK, start to process ' + get_chinese_city(city)) 37 | for csv in os.listdir(csv_dir): 38 | if csv[-3:] != "csv": 39 | continue 40 | data_csv = csv_dir + "/" + csv 41 | # print(data_csv) 42 | files.append(data_csv) 43 | 44 | # 清理数据 45 | count = 0 46 | for csv in files: 47 | with open(csv, 'r') as f: 48 | for line in f: 49 | count += 1 50 | text = line.strip() 51 | try: 52 | results = text.split("https://") 53 | except Exception as e: 54 | print(text) 55 | print(e) 56 | continue 57 | # 确保之前的步骤采集到了图片的url 58 | if len(results) > 1: 59 | url = results[-1] 60 | urls.append("https://"+url) 61 | print("https://"+url) 62 | print(len(urls)) 63 | return urls 64 | 65 | 66 | async def download_images(save_path: str, image_url: str): 67 | """ 68 | :param save_path: 保存图片的路径 69 | :param image_url: 图片的下载的url地址 70 | :return: 71 | """ 72 | async with aiohttp.ClientSession() as session: 73 | async with session.get(image_url, headers=create_headers()) as req: 74 | image = await req.read() 75 | fp = await aiofiles.open(save_path, 'wb') 76 | await fp.write(image) 77 | 78 | if __name__ == '__main__': 79 | # urls = ["https://img.ljcdn.com/370600-inspection/test-9925c97c-fc99-4d1a-97fa-2fd6d3209027.png!m_fill,w_280,h_210,f_jpg?from=ke.com", 80 | # "https://img.ljcdn.com/370600-inspection/df98f65c-427e-4d7d-91a7-425a5d682af5.jpg!m_fill,w_280,h_210,f_jpg?from=ke.com", 81 | # "https://img.ljcdn.com/370600-inspection/test-9925c97c-fc99-4d1a-97fa-2fd6d3209027.png!m_fill,w_280,h_210,f_jpg?from=ke.com", 82 | # "https://img.ljcdn.com/370600-inspection/df98f65c-427e-4d7d-91a7-425a5d682af5.jpg!m_fill,w_280,h_210,f_jpg?from=ke.com"] 83 | # 指定城市 84 | start = time.time() 85 | city = "yt" 86 | urls = get_ershou_img_urls(city) 87 | loop = asyncio.get_event_loop() 88 | date = get_date_string() 89 | csv_dir = "{0}/{1}/ershou/{2}/{3}".format(DATA_PATH, SPIDER_NAME, city, date) 90 | to_do = [download_images("{0}/{1}.jpg".format(csv_dir, i), urls[i]) for i in range(len(urls))] 91 | print("Start to download, please wait.") 92 | wait_future = asyncio.wait(to_do) 93 | resp = loop.run_until_complete(wait_future) 94 | loop.close() 95 | print("Download {0} images, cost {1} seconds.".format(len(urls), time.time() - start)) 96 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/tool/ershou_image_with_threads.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 多线程下载图片,Python 3.6.1 6 | 7 | 8 | import os 9 | import time 10 | from lib.zone.city import get_chinese_city 11 | from lib.request.headers import create_headers 12 | from lib.utility.date import get_date_string 13 | from lib.spider.base_spider import SPIDER_NAME 14 | from lib.utility.path import DATA_PATH 15 | from tomorrow import threads 16 | import requests 17 | 18 | 19 | def get_ershou_img_urls(city): 20 | urls = list() 21 | date = get_date_string() 22 | # 获得 csv 文件路径 23 | # date = "20180331" # 指定采集数据的日期 24 | # city = "sh" # 指定采集数据的城市 25 | csv_dir = "{0}/{1}/ershou/{2}/{3}".format(DATA_PATH, SPIDER_NAME, city, date) 26 | 27 | files = list() 28 | if not os.path.exists(csv_dir): 29 | print("{0} does not exist.".format(csv_dir)) 30 | print("Please run 'python ershou.py' firstly.") 31 | print("Bye.") 32 | exit(0) 33 | else: 34 | print('OK, start to process ' + get_chinese_city(city)) 35 | for csv in os.listdir(csv_dir): 36 | if csv[-3:] != "csv": 37 | continue 38 | data_csv = csv_dir + "/" + csv 39 | # print(data_csv) 40 | files.append(data_csv) 41 | 42 | # 清理数据 43 | count = 0 44 | for csv in files: 45 | with open(csv, 'r') as f: 46 | for line in f: 47 | count += 1 48 | text = line.strip() 49 | try: 50 | results = text.split("https://") 51 | except Exception as e: 52 | print(text) 53 | print(e) 54 | continue 55 | # 确保之前的步骤采集到了图片的url 56 | if len(results) > 1: 57 | url = results[-1] 58 | urls.append("https://"+url) 59 | print("https://"+url) 60 | print(len(urls)) 61 | return urls 62 | 63 | 64 | @threads(50) 65 | def download_images(save_path: str, image_url: str): 66 | """ 67 | :param save_path: 保存图片的路径 68 | :param image_url: 图片的下载的url地址 69 | :return: 70 | """ 71 | resp = requests.get(image_url) 72 | fp = open(save_path, 'wb') 73 | fp.write(resp.content) 74 | fp.close() 75 | 76 | if __name__ == '__main__': 77 | # urls = ["https://img.ljcdn.com/370600-inspection/test-9925c97c-fc99-4d1a-97fa-2fd6d3209027.png!m_fill,w_280,h_210,f_jpg?from=ke.com", 78 | # "https://img.ljcdn.com/370600-inspection/df98f65c-427e-4d7d-91a7-425a5d682af5.jpg!m_fill,w_280,h_210,f_jpg?from=ke.com", 79 | # "https://img.ljcdn.com/370600-inspection/test-9925c97c-fc99-4d1a-97fa-2fd6d3209027.png!m_fill,w_280,h_210,f_jpg?from=ke.com", 80 | # "https://img.ljcdn.com/370600-inspection/df98f65c-427e-4d7d-91a7-425a5d682af5.jpg!m_fill,w_280,h_210,f_jpg?from=ke.com"] 81 | # 指定城市 82 | start = time.time() 83 | city = "yt" 84 | urls = get_ershou_img_urls(city) 85 | 86 | date = get_date_string() 87 | csv_dir = "{0}/{1}/ershou/{2}/{3}".format(DATA_PATH, SPIDER_NAME, city, date) 88 | to_do = [download_images("{0}/{1}.jpg".format(csv_dir, i), urls[i]) for i in range(len(urls))] 89 | print("Start to download, please wait...") 90 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/tool/ershou_number.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: zengyuetian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 获取城市挂牌二手房数量 6 | 7 | import time 8 | from lib.spider.base_spider import SPIDER_NAME 9 | from bs4 import BeautifulSoup 10 | from lib.zone.city import cities 11 | import requests 12 | 13 | numbers = dict() 14 | 15 | 16 | def get_ershou_number(city): 17 | url = "https://{0}.{1}.com/ershoufang/".format(city, SPIDER_NAME) 18 | print(url) 19 | response = requests.get(url) 20 | html = response.content 21 | soup = BeautifulSoup(html, "lxml") 22 | element = soup.find('h2', class_='total') 23 | number = int(element.text.split(" ")[1].strip()) 24 | numbers[city] = number 25 | 26 | 27 | if __name__ == '__main__': 28 | start = time.time() 29 | for key, value in cities.items(): 30 | # print(key, value) 31 | get_ershou_number(key) 32 | for k, v in numbers.items(): 33 | print(cities[k], v) 34 | print("cost {0} seconds".format(time.time() - start)) 35 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/tool/lianjia_xiaoqu.sql: -------------------------------------------------------------------------------- 1 | # ************************************************************ 2 | # Sequel Pro SQL dump 3 | # Version 4541 4 | # 5 | # http://www.sequelpro.com/ 6 | # https://github.com/sequelpro/sequelpro 7 | # 8 | # Host: 127.0.0.1 (MySQL 5.7.21) 9 | # Database: lianjia 10 | # Generation Time: 2018-03-31 12:54:33 +0000 11 | # ************************************************************ 12 | 13 | 14 | /*!40101 SET @OLD_CHARACTER_SET_CLIENT=@@CHARACTER_SET_CLIENT */; 15 | /*!40101 SET @OLD_CHARACTER_SET_RESULTS=@@CHARACTER_SET_RESULTS */; 16 | /*!40101 SET @OLD_COLLATION_CONNECTION=@@COLLATION_CONNECTION */; 17 | /*!40101 SET NAMES utf8 */; 18 | /*!40014 SET @OLD_FOREIGN_KEY_CHECKS=@@FOREIGN_KEY_CHECKS, FOREIGN_KEY_CHECKS=0 */; 19 | /*!40101 SET @OLD_SQL_MODE=@@SQL_MODE, SQL_MODE='NO_AUTO_VALUE_ON_ZERO' */; 20 | /*!40111 SET @OLD_SQL_NOTES=@@SQL_NOTES, SQL_NOTES=0 */; 21 | 22 | 23 | # Dump of table xiaoqu 24 | # ------------------------------------------------------------ 25 | 26 | DROP TABLE IF EXISTS `xiaoqu`; 27 | 28 | CREATE TABLE `xiaoqu` ( 29 | `id` int(11) unsigned NOT NULL AUTO_INCREMENT, 30 | `city` varchar(10) DEFAULT NULL, 31 | `date` varchar(8) DEFAULT NULL, 32 | `district` varchar(50) DEFAULT NULL, 33 | `area` varchar(50) DEFAULT NULL, 34 | `xiaoqu` varchar(100) DEFAULT NULL, 35 | `price` int(11) DEFAULT NULL, 36 | `sale` int(11) DEFAULT NULL, 37 | PRIMARY KEY (`id`) 38 | ) ENGINE=InnoDB DEFAULT CHARSET=utf8; 39 | 40 | 41 | 42 | 43 | /*!40111 SET SQL_NOTES=@OLD_SQL_NOTES */; 44 | /*!40101 SET SQL_MODE=@OLD_SQL_MODE */; 45 | /*!40014 SET FOREIGN_KEY_CHECKS=@OLD_FOREIGN_KEY_CHECKS */; 46 | /*!40101 SET CHARACTER_SET_CLIENT=@OLD_CHARACTER_SET_CLIENT */; 47 | /*!40101 SET CHARACTER_SET_RESULTS=@OLD_CHARACTER_SET_RESULTS */; 48 | /*!40101 SET COLLATION_CONNECTION=@OLD_COLLATION_CONNECTION */; 49 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/xiaoqu.py: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env python 2 | # coding=utf-8 3 | # author: Zeng YueTian 4 | # 此代码仅供学习与交流,请勿用于商业用途。 5 | # 获得指定城市的小区数据 6 | # 这些数据包括: 7 | # 日期,所属区县,板块名,小区名,挂牌均价,挂牌数 8 | # 20180221,浦东,川沙,恒纬家苑,32176元/m2,3套在售二手房 9 | 10 | from lib.spider.xiaoqu_spider import * 11 | 12 | if __name__ == "__main__": 13 | spider = XiaoQuBaseSpider(SPIDER_NAME) 14 | spider.start() 15 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/上海小区爬虫最终版/xiaoqu_to_chart.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | import pandas as pd 4 | from pyecharts import Bar 5 | 6 | import os 7 | import time 8 | from lib.utility.version import PYTHON_3 9 | 10 | if __name__ == '__main__': 11 | try: 12 | import webbrowser as web 13 | auto_browse = True 14 | except Exception as e: 15 | auto_browse = False 16 | 17 | if auto_browse: 18 | try: 19 | if PYTHON_3: 20 | os.system("ps aux | grep python | grep http.server | grep -v grep | awk '{print $2}' | xargs kill") 21 | os.system("python -m http.server 8080 & > /dev/null 2>&1 ") 22 | else: 23 | os.system("ps aux | grep python | grep SimpleHTTPServer | grep -v grep | awk '{print $2}' | xargs kill") 24 | os.system("python -m SimpleHTTPServer 8080 & > /dev/null 2>&1 ") 25 | except Exception as e: 26 | print(e) 27 | 28 | # 注意,已经将分割符号转换成分号,因为有的小区名中有逗号 29 | df = pd.read_csv("xiaoqu.csv", encoding="utf-8", sep=";") 30 | 31 | # 打印总行数 32 | print("row number is {0}".format(len(df.index))) 33 | 34 | # 过滤房价为0的无效数据 35 | df = df[df.price > 0] 36 | # # 去除重复行 37 | # df = df.drop_duplicates() 38 | print("row number is {0}".format(len(df.index))) 39 | 40 | #################################################### 41 | # 最贵的小区排名 42 | #################################################### 43 | df.sort_values("price", ascending=False, inplace=True) 44 | num = 5 45 | print(df.head(num)) 46 | city = df["city_ch"][0] 47 | xqs = df["xiaoqu"][0:num] 48 | prices = df["price"][0:num] 49 | bar = Bar("{0}小区均价".format(city)) 50 | bar.add("小区均价前{0}名".format(num), xqs, prices, is_stack=True, is_label_show=True, xaxis_interval=0, xaxis_rotate=45) 51 | bar.render(path="xiaoqu.html") 52 | 53 | #################################################### 54 | # 区县均价排名 55 | #################################################### 56 | district_df = df.groupby('district').mean() 57 | district_df = district_df.round(0) 58 | district_df.sort_values("price", ascending=False, inplace=True) 59 | print(district_df) 60 | districts = district_df.index.tolist() 61 | prices = district_df["price"] 62 | bar = Bar("{0}区县均价".format(city)) 63 | bar.add("区县均价排名", districts, prices, is_stack=True, is_label_show=True, xaxis_interval=0, xaxis_rotate=45) 64 | bar.render(path="district.html") 65 | 66 | if auto_browse: 67 | web.open("http://localhost:8080/xiaoqu.html", new=0, autoraise=True) 68 | web.open("http://localhost:8080/district.html", new=0, autoraise=True) 69 | # 确保页面打开 70 | time.sleep(15) 71 | 72 | 73 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/LJxiaoquHB.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | # coding: utf-8 3 | import pandas as pd 4 | import numpy as np 5 | data1=pd.read_csv('housedata1.csv',encoding='gbk') 6 | data2=pd.read_csv('housedata2.csv',encoding='gbk') 7 | data=pd.concat([data1,data2],axis=0,ignore_index=False) 8 | data=pd.DataFrame(data) 9 | # data=str(data).encode('gbk') 10 | data=data.sort_values('area') 11 | data=data.reset_index() 12 | data=data.drop(labels='index',axis=1) 13 | # print(data.head()) 14 | print(data.loc[:,'area'].value_counts()) 15 | # for i,data['price'][i] in enumerate(data['price']): 16 | # data['price'][i]=int(data['price'][i].replace('元/平','')) 17 | #print(i,data['price'][i]) 18 | # print('changed_price\n',data['price'].head()) 19 | # print(data.head()) 20 | 21 | # print(type(data['price'][0]) 22 | data.to_csv('cleaned.csv', mode='a', encoding='gbk', header=False, index=False) 23 | 24 | # print(data.loc[:,'area'].value_counts()) 25 | # print(data.describe()) 26 | 27 | # area=data.groupby(by='area')['price'].mean() 28 | 29 | #print(data.loc[:,'price'].mean()) 30 | #area=data.groupby(by='area')['price'] 31 | # print(area) -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第1线程).csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第1线程).csv -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第1线程).xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第1线程).xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第2线程).xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/小区/上海链家小区信息(第2线程).xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/链家二手房.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/链家二手房.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/链家二手房数据整理.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/链家二手房数据整理.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/链家新房.csv: -------------------------------------------------------------------------------- 1 | title,location,niandai,leixing,loudongshu,fangwushu 2 | 楼盘名称,位置,建筑年代,房屋类型,楼栋数,房屋数 3 | -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/链家新房.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/链家新房.xlsx -------------------------------------------------------------------------------- /链家爬虫(含二手房、新房、租房、小区等爬虫)/链家租房.xlsx: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/liuzuoping/PythonSpyder_100_examples/1e137b396321b2643521582c2a761dc3fb3b025a/链家爬虫(含二手房、新房、租房、小区等爬虫)/链家租房.xlsx --------------------------------------------------------------------------------