Skip to content

Weekly Scrape

Weekly Scrape #34

Workflow file for this run

name: Weekly Scrape
# Run the workflow weekly
on:
workflow_dispatch:
schedule:
- cron: "0 0 * * 1"
permissions:
contents: write
jobs:
scrape_data:
runs-on: ubuntu-latest
steps:
- name: Checkout repository
uses: actions/checkout@v2
- name: Set up Python
uses: actions/setup-python@v4
with:
python-version: '3.11'
- name: Install python3-venv
run: sudo apt-get install -y python3-venv
- name: Check if venv module exists
run: python -m venv --help
- name: Debug environment
run: |
python --version
pwd
ls -la
- name: Debug .venv contents
run: ls -la /home/runner/work/gip_dashboard/gip_dashboard/.venv
- name: Create virtual environment
run: |
python -m venv /home/runner/work/gip_dashboard/gip_dashboard/.venv
ls -la /home/runner/work/gip_dashboard/gip_dashboard
- name: Install dependencies
run: |
python -m venv /home/runner/work/gip_dashboard/gip_dashboard/.venv
/home/runner/work/gip_dashboard/gip_dashboard/.venv/bin/python -m pip install --upgrade pip
/home/runner/work/gip_dashboard/gip_dashboard/.venv/bin/python -m pip install -r src/scripts/requirements.txt
# source .venv/bin/activate
# pip install -r src/scripts/requirements.txt # Assumes you have a requirements.txt file
- name: Run scraping script
env:
GEMINI_API_KEY: ${{ secrets.GEMINI_API_KEY }}
run: |
/home/runner/work/gip_dashboard/gip_dashboard/.venv/bin/python src/scripts/snapshot_crawler.py
#source .venv/bin/activate
#python src/scripts/snapshot_crawler.py
- name: Pull latest changes
run: |
git config --local user.name "github-actions[bot]"
git config --local user.email "github-actions[bot]@users.noreply.github.com"
git pull origin main
- name: Commit and push changes
run: |
git add .
git commit -m "Add generated YAML files"
git push origin main
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}