- A+
所属分类:python

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
#coding=utf8 #麦子学院视频爬虫 import requests,random,time,re,os from bs4 import BeautifulSoup def get_down_url(url,name,dir_name): x=requests.get(url) req_url=re.findall(r"lessonUrl = \"(.*)\"",x.text)[0] print ("\t\t下载视频地址:",req_url) save_m4v(req_url,name,dir_name) def get_liebiao_url(URL): #获取视频所有的课程列表URL shop=BeautifulSoup(requests.get(URL).text,'html.parser') dir_name=shop.select("h1")[0].text for i in shop.select(".lesson-lists li"): print ("http://www.maiziedu.com"+i.select('a')[0]['href'],i.select('a')[0].text) get_down_url("http://www.maiziedu.com"+i.select('a')[0]['href'],i.select('a')[0].text,dir_name) def save_m4v(url,name,dir_name): name=str(name).split(':')[0]+'.'+str(url).split('.')[-1] print(name) try: os.makedirs(dir_name) except Exception: pass f=open(dir_name+"\\"+name,'wb') #x=requests.get(url).text f.write(requests.get(url).content) #print(x) f.close() get_liebiao_url("http://www.maiziedu.com/course/304/")#传入下载URL列表地址 |
爬取的一步视频截图:
- 我的微信
- 这是我的微信扫一扫
-
- 我的微信公众号
- 我的微信公众号扫一扫
-