mirror of
https://github.com/bluxmit/alnoda-workspaces.git
synced 2024-05-13 02:22:18 +12:00
50 lines
1.1 KiB
Markdown
50 lines
1.1 KiB
Markdown
# Get started
|
|
|
|
This tutorial shows how to scrap Google news on schedule. We use Python for the demonstration purpose.
|
|
|
|
To start, open workspace terminal and install Python package GNews
|
|
|
|
```
|
|
pip install gnews
|
|
```
|
|
|
|
Create python script
|
|
|
|
```
|
|
nano scraper.py
|
|
```
|
|
|
|
|
|
Paste the following code snippet. This is scrapper, that will output results to a file
|
|
|
|
```py
|
|
from gnews import GNews
|
|
import json
|
|
|
|
google_news = GNews()
|
|
usa_news = google_news.get_news('USA')
|
|
with open('usa_news.json', 'w') as f:
|
|
json.dump(usa_news, f)
|
|
```
|
|
|
|
Save the file with `Ctrl+s` and return to the terminal with `Ctrl+x`.
|
|
|
|
Execute the script
|
|
|
|
```
|
|
python3 scraper.py
|
|
```
|
|
|
|
Navigate to Quickstart and Open file browser. View and download the output file
|
|
|
|
<p align="center">
|
|
<img src="https://github.com/bluxmit/alnoda-workspaces/blob/main/workspaces/base-workspace/img/filebrowser-demo.gif?raw=true" alt="Filebrowser" width="750">
|
|
</p>
|
|
|
|
Schedule daily executions with Cronicle
|
|
|
|
<p align="center">
|
|
<img src="https://github.com/bluxmit/alnoda-workspaces/blob/main/workspaces/base-workspace/img/cronicle-demo.gif?raw=true" alt="Filebrowser" width="750">
|
|
</p>
|
|
|