WebScraper is a Python project for extracting all links from a given website. It features both a command-line interface (CLI) and a modern, dark-themed graphical user interface (GUI) built with Tkinter.
- Extracts all links from a specified website
- Saves results to a timestamped text file
- Dark-themed, user-friendly GUI (Tkinter)
- CLI usage for quick scraping
- Error handling for invalid URLs and network issues
- Python 3.7+
- See
requirements.txtfor dependencies:- requests
- beautifulsoup4
- lxml
- pyfiglet
- Clone the repository:
git clone <repo-url> cd WebScraper
- Install dependencies:
pip install -r requirements.txt
Run the scraper from the terminal:
python main.py <site_url>- Example:
python main.py example.com
- The results will be printed to the console and saved in the
Results/folder.
- Start the GUI:
python Modules/gui.py
- Enter the website URL (e.g., https://example.com) in the input field.
- Click the "Scrape" button or press Enter to start scraping.
- All found links will be displayed in the results area.
- Click "Save Results" to export the links to a file in the
Results/folder.
This is the initial look of the application when you first open it.
After entering a URL and clicking 'Scrape', the application scans the website for links.
All found links are displayed in the results area, ready to be saved.
WebScraper/
├── main.py # CLI entry point
├── requirements.txt # Python dependencies
├── README.md # Project documentation
├── Results/ # Scraped results are saved here
├── Modules/
│ ├── scraper.py # Scraping logic
│ ├── greeting.py # Welcome message logic
│ └── gui.py # Tkinter GUI


