docs: update README to simplify usage instructions and remove outdated content

This commit is contained in:
patilk 2024-11-15 17:13:29 +01:00
parent cc79bb9d9d
commit 4bb6619911
Signed by: s500042
GPG Key ID: 1921AD722E7392EE

View File

@ -2,23 +2,20 @@
## Description ## Description
This project is a web scraper designed to extract data from websites. It can be customized to scrape various types of data and save it in different formats. This project is a web scraper designed to extract data from websites.
## Features ## Features
- Extracts data from web pages ☑️ Extracts data from web pages
<!-- - Supports multiple data formats (CSV, JSON, etc.)
- Customizable scraping rules
- Error handling and logging -->
## Installation ## Usage
### Using Docker ### With Docker
1. Clone the repository: 1. Clone the repository:
```bash ```bash
git clone https://git.wmi.amu.edu.pl/s500042/webscraper git clone https://git.wmi.amu.edu.pl/s500042/webscraper
``` ```
2. Navigate to the project directory: 2. Navigate to the project directory:
@ -27,20 +24,13 @@ git clone https://git.wmi.amu.edu.pl/s500042/webscraper
cd webscraper cd webscraper
``` ```
3. Build the Docker image and run it using script: 3. Build the Docker image and run it using `start.py` script:
- On Linux, ?Mac <!-- I haven't tested it yet -->
```bash
./start.sh
```
- Windows 🤡
```bash ```bash
python start.py python start.py
``` ```
This one will work just fine on Linux, but on Mac, you'll have to use On Mac, you'll have to use
```bash ```bash
python3 start.py python3 start.py
@ -51,31 +41,28 @@ python3 start.py
1. Clone the repository: 1. Clone the repository:
```bash ```bash
git clone https://github.com/yourusername/webscraper.git git clone https://github.com/yourusername/webscraper.git
``` ```
2. Navigate to the project directory: 2. Install the required dependencies:
```bash ```bash
cd webscraper/app pip install -r requirements.txt
```
3. Install the required dependencies:
```bash
pip install -r requirements.txt
``` ```
If you're on Arch Linux, you'll need to create a virtual environment. If you're on Arch Linux, you'll need to create a virtual environment.
Here's is a [Step by step guide](#) that will help you create it. Here's is a [Step by step guide](#) that will help you create it.
## Usage 3. Run `run-with-no-docker.py` script:
1. Configure the scraper by editing the `config.json` file.
2. Run the scraper:
```bash ```bash
python scraper.py python run-with-no-docker.py
```
On Mac you'll, need to use:
```bash
python3 run-with-no-dcoker.py
``` ```
## License ## License