


Let’s talk about our school’s website first:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/zhxt_bks.html
To check scores, you need to log in and then display The results of each subject are displayed, but only the results are displayed without the grade points, which is the weighted average score.
Obviously calculating grade points manually is a very troublesome thing. So we can use Python to make a crawler to solve this problem.
1. On the eve of the decisive battle
Let’s prepare a tool first: HttpFox plug-in.
This is an http protocol analysis plug-in that analyzes page request and response time, content, and COOKIE used by the browser.
Take me as an example, just install it on Firefox, the effect is as shown:
Okay View the corresponding information very intuitively.
Click start to start detection, click stop to pause detection, and click clear to clear the content.
Generally before use, click stop to pause, and then click clear to clear the screen to ensure that you see the data obtained by accessing the current page.
2. Go deep behind enemy lines
Let’s go to Shandong University’s score query website to see what information is sent when logging in.
First go to the login page, open httpfox, after clearing, click start to turn on the detection:
After entering the personal information, make sure httpfox is turned on. Then click OK to submit the information and log in.
You can see at this time that httpfox has detected three pieces of information:
At this time, click the stop button to ensure that what is captured is the feedback after visiting the page. Data so that we can simulate login when doing crawlers.
3. Pao Ding Jie Niu
At first glance, we got three data, two are GET and one is POST, but what are they? What and how it should be used, we still have no idea.
So, we need to check the captured content one by one.
Look at the POST information first:
Since it is the POST information, we can just look at the PostData.
You can see that there are two POST data, studid and pwd.
And it can be seen from the Redirect to of Type that after the POST is completed, it jumps to the bks_login2.loginmessage page.
It can be seen that this data is the form data submitted after clicking OK.
Click on the cookie label to see the cookie information:
Yes, an ACCOUNT cookie was received and will be automatically destroyed after the session ends.
So what information did you receive after submitting?
Let’s take a look at the next two GET data.
Look at the first one first. We click on the content tag to view the received content. Do you feel like eating it alive? -The HTML source code is undoubtedly exposed:
It seems that this is just the html source code of the page. Click on the cookie to view the cookie-related information:
Aha, it turns out that the content of the html page was received only after the cookie information was sent.
Let’s take a look at the last received message:
After a rough look, it should be just a css file called style.css, which doesn’t mean much to us. big effect.
4. Calmly respond
Now that we know what data we sent to the server and what data we received, the basic process is as follows:
First, we POST the student ID and password--->Then return the cookie value
Then send the cookie to the server--->Return the page information.
Get the data from the grades page, use regular expressions to extract the grades and credits separately and calculate the weighted average.
OK, it looks like a very simple sample paper. Then let’s try it out.
But before the experiment, there is still an unresolved problem, which is where is the POST data sent?
Look at the original page again:
It is obviously implemented using an html framework, that is to say, what we see in the address bar The address is not the address to submit the form on the right.
So how can I get the real address-. -Right click to view page source code:
Yes, that’s right, the one with name="w_right" is the login page we want.
The original address of the website is:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/zhxt_bks.html
So, the real form submission The address should be:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/xk_login.html
After entering it, it is true:
It’s actually the course selection system of Tsinghua University. . . My guess is that our school was too lazy to create a page, so we just borrowed it. . As a result, the title was not even changed. . .
But this page is still not the page we need, because the page our POST data is submitted to should be the page submitted in the ACTION of the form.
In other words, we need to check the source code to know where the POST data is sent:
Well, visually this is the address where the POST data is submitted. .
Arrange it into the address bar. The complete address should be as follows:
http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login
(The method of obtaining it is very simple. Just click on the link directly in Firefox browser to see the address of the link)
5. A little test of my skills
The next task is to use python to simulate sending a POST data and get the returned cookie value.
Regarding the operation of cookies, you can read this blog post:
http://blog.csdn.net/wxg694175346/article/details/8925978
We first prepare a POST data, then prepare a cookie to receive, and then write the source code as follows:
# -*- coding: utf-8 -*- #--------------------------------------- # 程序:山东大学爬虫 # 版本:0.1 # 作者:why # 日期:2013-07-12 # 语言:Python 2.7 # 操作:输入学号和密码 # 功能:输出成绩的加权平均值也就是绩点 #--------------------------------------- import urllib import urllib2 import cookielib cookie = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie)) #需要POST的数据# postdata=urllib.urlencode({ 'stuid':'201100300428', 'pwd':'921030' }) #自定义一个请求# req = urllib2.Request( url = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login', data = postdata ) #访问该链接# result = opener.open(req) #打印返回的内容# print result.read()
After this, look at the effect of the operation:
ok, in this way, we will calculate that the simulated login is successful.
6. Change the situation
The next task is to use a crawler to obtain the students’ scores.
Let’s look at the source website again.
After opening HTTPFOX, click to view the results and find that the following data was captured:
##Click on the first GET data and check the content. You can find that the Content is the content of the obtained results.
## so you can get The link to view the results is as follows:
http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopere
7. Everything is ready
Now everything is ready, so just apply the link to the crawler and see if you can see the results page.
As you can see from httpfox, we have to send a cookie to return the score information, so we use python to simulate sending a cookie to request the score information:
# -*- coding: utf-8 -*- #--------------------------------------- # 程序:山东大学爬虫 # 版本:0.1 # 作者:why # 日期:2013-07-12 # 语言:Python 2.7 # 操作:输入学号和密码 # 功能:输出成绩的加权平均值也就是绩点 #--------------------------------------- import urllib import urllib2 import cookielib #初始化一个CookieJar来处理Cookie的信息# cookie = cookielib.CookieJar() #创建一个新的opener来使用我们的CookieJar# opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie)) #需要POST的数据# postdata=urllib.urlencode({ 'stuid':'201100300428', 'pwd':'921030' }) #自定义一个请求# req = urllib2.Request( url = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login', data = postdata ) #访问该链接# result = opener.open(req) #打印返回的内容# print result.read() #打印cookie的值 for item in cookie: print 'Cookie:Name = '+item.name print 'Cookie:Value = '+item.value #访问该链接# result = opener.open('http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre') #打印返回的内容# print result.read()Press F5 to run and take a look at the captured data:
Since there is no problem, use regular expressions to convert the data Just process it a little and take out the credits and corresponding scores.
8. At your fingertips
Such a large amount of html source code is obviously not conducive to our processing. Next, we will use regular expressions to extract the necessary data.
For tutorials on regular expressions, you can read this blog post:
http://blog.csdn.net/wxg694175346/article/details/8929576
Let’s take a look at the source code of the score:
既然如此,用正则表达式就易如反掌了。 我们将代码稍稍整理一下,然后用正则来取出数据: 水平有限,,正则是有点丑,。运行的效果如图: ok,接下来的只是数据的处理问题了。。 9.凯旋而归 完整的代码如下,至此一个完整的爬虫项目便完工了。 以上就是 [Python]网络爬虫(十):一个爬虫的诞生全过程(以山东大学绩点运算为例)的内容,更多相关内容请关注PHP中文网(www.php.cn)!# -*- coding: utf-8 -*-
#---------------------------------------
# 程序:山东大学爬虫
# 版本:0.1
# 作者:why
# 日期:2013-07-12
# 语言:Python 2.7
# 操作:输入学号和密码
# 功能:输出成绩的加权平均值也就是绩点
#---------------------------------------
import urllib
import urllib2
import cookielib
import re
class SDU_Spider:
# 申明相关的属性
def __init__(self):
self.loginUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login' # 登录的url
self.resultUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre' # 显示成绩的url
self.cookieJar = cookielib.CookieJar() # 初始化一个CookieJar来处理Cookie的信息
self.postdata=urllib.urlencode({'stuid':'201100300428','pwd':'921030'}) # POST的数据
self.weights = [] #存储权重,也就是学分
self.points = [] #存储分数,也就是成绩
self.opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(self.cookieJar))
def sdu_init(self):
# 初始化链接并且获取cookie
myRequest = urllib2.Request(url = self.loginUrl,data = self.postdata) # 自定义一个请求
result = self.opener.open(myRequest) # 访问登录页面,获取到必须的cookie的值
result = self.opener.open(self.resultUrl) # 访问成绩页面,获得成绩的数据
# 打印返回的内容
# print result.read()
self.deal_data(result.read().decode('gbk'))
self.print_data(self.weights);
self.print_data(self.points);
# 将内容从页面代码中抠出来
def deal_data(self,myPage):
myItems = re.findall('<TR>.*?<p.*?<p.*?<p.*?<p.*?<p.*?>(.*?)</p>.*?<p.*?<p.*?>(.*?)</p>.*?</TR>',myPage,re.S) #获取到学分
for item in myItems:
self.weights.append(item[0].encode('gbk'))
self.points.append(item[1].encode('gbk'))
# 将内容从页面代码中抠出来
def print_data(self,items):
for item in items:
print item
#调用
mySpider = SDU_Spider()
mySpider.sdu_init()
# -*- coding: utf-8 -*-
#---------------------------------------
# 程序:山东大学爬虫
# 版本:0.1
# 作者:why
# 日期:2013-07-12
# 语言:Python 2.7
# 操作:输入学号和密码
# 功能:输出成绩的加权平均值也就是绩点
#---------------------------------------
import urllib
import urllib2
import cookielib
import re
import string
class SDU_Spider:
# 申明相关的属性
def __init__(self):
self.loginUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login' # 登录的url
self.resultUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre' # 显示成绩的url
self.cookieJar = cookielib.CookieJar() # 初始化一个CookieJar来处理Cookie的信息
self.postdata=urllib.urlencode({'stuid':'201100300428','pwd':'921030'}) # POST的数据
self.weights = [] #存储权重,也就是学分
self.points = [] #存储分数,也就是成绩
self.opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(self.cookieJar))
def sdu_init(self):
# 初始化链接并且获取cookie
myRequest = urllib2.Request(url = self.loginUrl,data = self.postdata) # 自定义一个请求
result = self.opener.open(myRequest) # 访问登录页面,获取到必须的cookie的值
result = self.opener.open(self.resultUrl) # 访问成绩页面,获得成绩的数据
# 打印返回的内容
# print result.read()
self.deal_data(result.read().decode('gbk'))
self.calculate_date();
# 将内容从页面代码中抠出来
def deal_data(self,myPage):
myItems = re.findall('<TR>.*?<p.*?<p.*?<p.*?<p.*?<p.*?>(.*?)</p>.*?<p.*?<p.*?>(.*?)</p>.*?</TR>',myPage,re.S) #获取到学分
for item in myItems:
self.weights.append(item[0].encode('gbk'))
self.points.append(item[1].encode('gbk'))
#计算绩点,如果成绩还没出来,或者成绩是优秀良好,就不运算该成绩
def calculate_date(self):
point = 0.0
weight = 0.0
for i in range(len(self.points)):
if(self.points[i].isdigit()):
point += string.atof(self.points[i])*string.atof(self.weights[i])
weight += string.atof(self.weights[i])
print point/weight
#调用
mySpider = SDU_Spider()
mySpider.sdu_init()

本篇文章给大家带来了关于Python的相关知识,其中主要介绍了关于Seaborn的相关问题,包括了数据可视化处理的散点图、折线图、条形图等等内容,下面一起来看一下,希望对大家有帮助。

本篇文章给大家带来了关于Python的相关知识,其中主要介绍了关于进程池与进程锁的相关问题,包括进程池的创建模块,进程池函数等等内容,下面一起来看一下,希望对大家有帮助。

本篇文章给大家带来了关于Python的相关知识,其中主要介绍了关于简历筛选的相关问题,包括了定义 ReadDoc 类用以读取 word 文件以及定义 search_word 函数用以筛选的相关内容,下面一起来看一下,希望对大家有帮助。

本篇文章给大家带来了关于Python的相关知识,其中主要介绍了关于数据类型之字符串、数字的相关问题,下面一起来看一下,希望对大家有帮助。

VS Code的确是一款非常热门、有强大用户基础的一款开发工具。本文给大家介绍一下10款高效、好用的插件,能够让原本单薄的VS Code如虎添翼,开发效率顿时提升到一个新的阶段。

本篇文章给大家带来了关于Python的相关知识,其中主要介绍了关于numpy模块的相关问题,Numpy是Numerical Python extensions的缩写,字面意思是Python数值计算扩展,下面一起来看一下,希望对大家有帮助。

pythn的中文意思是巨蟒、蟒蛇。1989年圣诞节期间,Guido van Rossum在家闲的没事干,为了跟朋友庆祝圣诞节,决定发明一种全新的脚本语言。他很喜欢一个肥皂剧叫Monty Python,所以便把这门语言叫做python。


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

EditPlus Chinese cracked version
Small size, syntax highlighting, does not support code prompt function

Safe Exam Browser
Safe Exam Browser is a secure browser environment for taking online exams securely. This software turns any computer into a secure workstation. It controls access to any utility and prevents students from using unauthorized resources.

Dreamweaver CS6
Visual web development tools

SublimeText3 Linux new version
SublimeText3 Linux latest version

mPDF
mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),
