2024-02-29 22:16:39 +00:00
|
|
|
name: Scrape transfer and update file
|
2024-03-01 06:45:06 +00:00
|
|
|
run-name: Scrape transfer and update file
|
2024-02-29 22:16:39 +00:00
|
|
|
on:
|
|
|
|
workflow_dispatch:
|
2024-03-06 08:43:41 +00:00
|
|
|
schedule:
|
|
|
|
- cron: '0 10 * * *'
|
2024-02-29 22:16:39 +00:00
|
|
|
|
|
|
|
jobs:
|
2024-03-05 23:42:05 +00:00
|
|
|
setup:
|
|
|
|
name: Get number of pages and set up scrape page jobs
|
2024-02-29 22:16:39 +00:00
|
|
|
runs-on: ubuntu-latest
|
|
|
|
steps:
|
2024-03-05 23:27:51 +00:00
|
|
|
- name: Create matrix parameters
|
|
|
|
id: matrix-params
|
|
|
|
run: |
|
|
|
|
NUM_PAGES="$(curl -H 'User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:122.0) Gecko/20100101 Firefox/122.0' 'https://tes.collegesource.com/publicview/TES_publicview01.aspx?rid=f080a477-bff8-46df-a5b2-25e9affdd4ed&aid=27b576bb-cd07-4e57-84d0-37475fde70ce' | grep -e 'lblInstWithEQPaginationInfo' | grep -Poie '(?<=of )[0-9]*')"
|
|
|
|
MATRIX_PARAMS="$(seq -s "," 1 "$NUM_PAGES")"
|
|
|
|
MATRIX_PARAMS="\"page\": $(sed -e 's/,/}, {"page": /g' <<< "$MATRIX_PARAMS")"
|
2024-03-05 23:46:02 +00:00
|
|
|
echo "matrix-params={\"include\": [{"$MATRIX_PARAMS"}]}" | tee $GITHUB_OUTPUT
|
2024-03-05 23:27:51 +00:00
|
|
|
outputs:
|
|
|
|
matrix-params: ${{ steps.matrix-params.outputs.matrix-params }}
|
2024-03-06 06:43:26 +00:00
|
|
|
|
2024-03-05 23:46:02 +00:00
|
|
|
scrape-page:
|
2024-03-05 23:42:05 +00:00
|
|
|
name: Scrape page
|
2024-03-01 06:45:06 +00:00
|
|
|
runs-on: ubuntu-latest
|
2024-03-05 23:42:05 +00:00
|
|
|
needs: setup
|
2024-03-05 23:27:51 +00:00
|
|
|
strategy:
|
2024-03-05 23:46:02 +00:00
|
|
|
matrix: ${{ fromJson(needs.setup.outputs.matrix-params) }}
|
2024-03-06 08:36:33 +00:00
|
|
|
fail-fast: true
|
2024-03-01 06:45:06 +00:00
|
|
|
steps:
|
2024-03-05 23:42:05 +00:00
|
|
|
- name: Checkout scraping repo
|
|
|
|
uses: actions/checkout@v4
|
|
|
|
|
|
|
|
- name: Set up python
|
|
|
|
uses: actions/setup-python@v5
|
|
|
|
with:
|
|
|
|
python-version: '3.11'
|
|
|
|
cache: 'pip'
|
|
|
|
|
|
|
|
- name: Install dependencies
|
|
|
|
working-directory: transfer_scraper
|
|
|
|
run: |
|
|
|
|
python -m pip install --upgrade pip
|
|
|
|
pip install -r 'requirements.txt'
|
|
|
|
|
|
|
|
- name: Run scraper
|
|
|
|
working-directory: transfer_scraper
|
|
|
|
run: |
|
2024-03-06 00:05:41 +00:00
|
|
|
python3 main.py ${{ matrix.page }} transfer_${{ matrix.page }}.json
|
2024-03-05 23:42:05 +00:00
|
|
|
|
|
|
|
- name: Upload data to artifact
|
|
|
|
uses: actions/upload-artifact@v4
|
|
|
|
with:
|
|
|
|
name: transfer-page-${{ matrix.page }}
|
2024-03-06 00:06:49 +00:00
|
|
|
path: transfer_scraper/transfer_${{ matrix.page }}.json
|
2024-03-05 23:42:05 +00:00
|
|
|
|
2024-03-06 06:43:26 +00:00
|
|
|
commit-data:
|
|
|
|
name: Combine and commit data
|
|
|
|
runs-on: ubuntu-latest
|
|
|
|
needs: scrape-page
|
|
|
|
steps:
|
|
|
|
- name: Checkout data repo
|
|
|
|
uses: actions/checkout@v4
|
|
|
|
with:
|
|
|
|
repository: quatalog/data
|
|
|
|
path: data
|
2024-03-06 08:05:49 +00:00
|
|
|
token: ${{ secrets.PUSH_TOKEN }}
|
2024-03-06 06:43:26 +00:00
|
|
|
|
|
|
|
- name: Download partial JSONs
|
|
|
|
uses: actions/download-artifact@v4
|
|
|
|
with:
|
|
|
|
pattern: transfer-page-*
|
|
|
|
merge-multiple: true
|
|
|
|
path: new-data
|
|
|
|
|
|
|
|
- name: Combine JSONs
|
|
|
|
run: |
|
2024-03-06 18:10:23 +00:00
|
|
|
cat new-data/* | jq -s 'add | sort_by(.institution)' > data/transfer.json
|
2024-03-06 06:43:26 +00:00
|
|
|
|
|
|
|
- name: Commit data
|
|
|
|
working-directory: data
|
|
|
|
run: |
|
|
|
|
git config user.name "Quatalog Updater"
|
|
|
|
git config user.email "github_actions@quatalog.com"
|
|
|
|
git add transfer.json
|
|
|
|
git commit -m "$(date)" || exit 0
|
|
|
|
git push
|