此脚本为 python 脚本,实现原理为利用 selenium 定位相关元素,再配合点击事件完成浏览器的自动化.

Overview

Introduce

此脚本为 python 脚本,实现原理为利用 selenium 定位相关元素,再配合点击事件完成浏览器的自动化.

本打卡脚本适用于利用青柠疫服-高校疫情防空平台 https://wxyqfk.zhxy.net/#/poster 的每日疫情打卡


Usage

Download

可以利用 git 将此仓库下载到本地

git clone https://github.com/N0el4kLs/yqfkAutoCheck.git

Pre-use

安装脚本相关依赖

pip install -r requirements.txt
#pip3 install -r requirements.txt

info.ini Setting

用户需要在使用前先完善 info.ini 文件.

设置学校代码

选择自己的学校,点击确认

urlyxdm=xxxx 中的 xxxx 为学校代码 示例中的学院代码为 10646

完善配置文件中的[Url]部分

[Url]
loginpage = https://wxyqfk.zhxy.net/?yxdm=10646#/login

打卡用户信息设置

配置示例:

[PeopleList]
people = example 
// 这里填写用户表示,以此来查找用户的 LoginId SchoolCard 以及 PassWd
// 如果有多个用户,用户与用户之间用逗号隔开: people = example1,example2,example3

[LoginId]
example = 测试


[SchoolCard]
example = 123456789

[PassWd]
examole = 123456admin

Code Supplement

验证码

由于本人在实现识别图片验证码功能时是调用了别人的接口,所以此处代码需要用户自己填充.

填充代码解构: 输入:图片的路径的, 返回值: 为验证码.

需要修改的代码在QNcheck.py109-116

def getveriycode(self,imagedata):
        '''
        还原base64 图片信息,并获取图片验证码
        '''
        self.decodeImag(imagedata)
        
        # img_path = ./verifycode.png
        return USEROWNFUNC(r'./verifycode.png')  # 替换USEROWNFUNCTION(r'./verifycode.png') 为用户填充代码

邮件功能

本脚本提供了邮件发送功能,以提醒用户今日的打卡情况 此处使用的qq邮箱,所以需要用户填充相关信息 需要填充的代码位于 e_mail.py7-8

self.__my_sender =''   #发件人邮箱账号
self.__my_pass = ''    #发件人邮箱密码

QNcheck.py140中填入接受邮件的邮箱.

login_email = LoginEmail('user email') # 将 user eamil 更改为 接受邮件的邮箱

定位功能

QNcheck.py16-17行完善经纬度信息.

self.latitude = xxx    // 纬度
self.longitude = xxx  // 经度

RUN

运行脚本

python QNcheck.py
#python3 QNcheck.py
Owner
N0el4kLs
Just a Web Security Fresher
N0el4kLs
download NCERT books using scrapy

download_ncert_books download NCERT books using scrapy Downloading Books: You can either use the spider by cloning this repo and following the instruc

1 Dec 02, 2022
Scrap-mtg-top-8 - A top 8 mtg scraper using python

Scrap-mtg-top-8 - A top 8 mtg scraper using python

1 Jan 24, 2022
This script is intended to crawl license information of repositories through the GitHub API.

GithubLicenseCrawler This script is intended to crawl license information of repositories through the GitHub API. Taking a csv file with requirements.

schutera 4 Oct 25, 2022
feapder 是一款简单、快速、轻量级的爬虫框架。以开发快速、抓取快速、使用简单、功能强大为宗旨。支持分布式爬虫、批次爬虫、多模板爬虫,以及完善的爬虫报警机制。

feapder 是一款简单、快速、轻量级的爬虫框架。起名源于 fast、easy、air、pro、spider的缩写,以开发快速、抓取快速、使用简单、功能强大为宗旨,历时4年倾心打造。支持轻量爬虫、分布式爬虫、批次爬虫、爬虫集成,以及完善的爬虫报警机制。 之

boris 1.4k Dec 29, 2022
A web crawler script that crawls the target website and lists its links

A web crawler script that crawls the target website and lists its links || A web crawler script that lists links by scanning the target website.

2 Apr 29, 2022
Basic-html-scraper - A complete how to of web scraping with Python for beginners

basic-html-scraper Code from YT Video This video includes a complete how to of w

John 12 Oct 22, 2022
Screenhook is a script that captures an image of a web page and send it to a discord webhook.

screenshot from the web for discord webhooks screenhook is a script that captures an image of a web page and send it to a discord webhook.

Toast Energy 3 Jun 04, 2022
用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。

crawler_for_university 用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。 环境依赖 wxpy,requests,bs4等库 功能描述 该项目基于python,通过爬虫爬各高校的就业信息网,爬取招聘信

8 Aug 16, 2021
Scrapping Connections' info on Linkedin

Scrapping Connections' info on Linkedin

MohammadReza Ardestani 1 Feb 11, 2022
A simple code to fetch comments below an Instagram post and save them to a csv file

fetch_comments A simple code to fetch comments below an Instagram post and save them to a csv file usage First you have to enter your username and pas

2 Jul 14, 2022
12306抢票脚本

12306抢票脚本

罐子里的茶 457 Jan 05, 2023
This is a module that I had created along with my friend. It's a basic web scraping module

QuickInfo PYPI link : https://pypi.org/project/quickinfo/ This is the library that you've all been searching for, it's built for developers and allows

OneBit 2 Dec 13, 2021
A multithreaded tool for searching and downloading images from popular search engines. It is straightforward to set up and run!

🕳️ CygnusX1 Code by Trong-Dat Ngo. Overviews 🕳️ CygnusX1 is a multithreaded tool 🛠️ , used to search and download images from popular search engine

DatNgo 32 Dec 31, 2022
A simple python web scraper.

Dissec A simple python web scraper. It gets a website and its contents and parses them with the help of bs4. Installation To install the requirements,

11 May 06, 2022
Kusonime scraper using python3

Features Scrap from url Scrap from recommendation Search by query Todo [+] Search by genre Example # Get download url from kusonime import Scrap

MhankBarBar 2 Jan 28, 2022
A database scraper created with mechanical soup and sqlite

WebscrapingDatabases a database scraper created with mechanical soup and sqlite author: Mariya Sha Watch on YouTube: This repository was created to su

Mariya 30 Aug 08, 2022
Web scrapper para cotizar articulos

WebScrapper Este web scrapper esta desarrollado en python 3.10.0 para buscar en la pagina de cyber puerta articulos dentro del catalogo. El programa t

Jordan Gaona 1 Oct 27, 2021
A scrapy pipeline that provides an easy way to store files and images using various folder structures.

scrapy-folder-tree This is a scrapy pipeline that provides an easy way to store files and images using various folder structures. Supported folder str

Panagiotis Simakis 7 Oct 23, 2022
Automatically scrapes all menu items from the Taco Bell website

Automatically scrapes all menu items from the Taco Bell website. Returns as PANDAS dataframe.

Sasha 2 Jan 15, 2022
Minecraft Item Scraper

Minecraft Item Scraper To run, first ensure you have the BeautifulSoup module: pip install bs4 Then run, python minecraft_items.py folder-to-save-ima

Jaedan Calder 1 Dec 29, 2021