A web crawl script, search while crawling
该项目利用爬虫的原理进行搜索匹配所给关键词的网站,旨在提供全网搜索的功能,发现网络角落。
This project utilizes the principle of crawling to search and match websites with given keywords, aiming to provide a full network search function and discover network corners.
原理是在所给页面中找到所有链接,并访问他们,匹配是否符合关键词并记录,接着重复以上过程,从访问得到的内容中再得到链接,并执行搜索。
The principle is to find all the links on the given page, visit them, match whether they match the keywords, and record them. Then, repeat the above process to obtain the links from the accessed content, and perform a search.
v1.9
- Request timeout is set to 7 sec (default).
- Request timeout can be customized through command parameters.
- Simplify some code.
- A BUG Fixed: Default request timeout setting was not available in function "web_crawl".
Type this command to make sure that you have already installed the needed library. If not, it would install them automaticly.
pip install -r requirements.txt
Double click (or python3 ./SearSpy.py
) to run it.
Type the website's URL you want to search in.
Type your keyword (a word or a phrase is suggested)
Type the Recursion Depth (author suggested to fill the number between 10 and 20, larger number is okey but useless)
Type y
or Y
if you want to save all crawled links
Type y
or Y
to display all outputs
If you want to use advanced features, try python3 ./SearSpy.py -h
command!
This is a programe for turning csv format to json format.
Drag your csv file on its icon and release, or use python3 cv2json.py <path>
, or run it and input the file path.
This program is used to download all images in CSV files that link to the website.
Drag your csv file on its icon and release, or use python3 img_download.py <path>
, or run it and input the file path.
This is a SearSpy Wizard Program supports Chinese and English.
- make it faster (multi-thread)
- More customization options available