Skip to content

SatapasT/Wiki-web-scraper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 

Repository files navigation

Wiki Web Scraper

Wiki Web Scraper is a Python program designed for extracting and analyzing textual data from Wiki Fandom pages, specifically those related to TV shows or other topics. It utilizes the BeautifulSoup library for HTML parsing, the requests library for making HTTP requests, and the Googlesearch library for finding relevant Wikipedia pages.

Features

  1. Data Collection: The program retrieves textual data from the provided Wikipedia page and its linked pages.
  2. Keyword Analysis: It identifies and counts the occurrences of individual words (excluding non-alphabetic characters) within the collected data.
  3. More to Come: Stay tuned for additional features as the project continues to evolve!

Dependencies

Installation and Usage

pip install requests beautifulsoup4 alive-progress google

Run the script in a Python environment:

python main.py

Notes

The program's execution time may vary depending on the size and structure of the Wikipedia page and its linked pages.

Author

This program was created by SatapasT.

Happy Scrapping! :D

About

A wiki web scraper

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages