|
1 | | -name: Daily Crawling at 6AM KST |
2 | | - |
3 | | -on: |
4 | | - schedule: |
5 | | - - cron: "0 21 * * *" # 한국시간 오전 6시 = UTC 21시 |
6 | | - workflow_dispatch: |
7 | | - |
8 | | -jobs: |
9 | | - crawl: |
10 | | - runs-on: ubuntu-latest |
11 | | - |
12 | | - steps: |
13 | | - - name: Checkout Repository |
14 | | - uses: actions/checkout@v3 |
15 | | - |
16 | | - - name: Set up Python 3.11 |
17 | | - uses: actions/setup-python@v4 |
18 | | - with: |
19 | | - python-version: "3.11" |
20 | | - |
21 | | - - name: Install Python dependencies |
22 | | - run: | |
23 | | - python -m pip install --upgrade pip |
24 | | - pip install -r requirements.txt |
25 | | -
|
26 | | - - name: Generate .env file from GitHub Secrets |
27 | | - run: | |
28 | | - echo "MYSQL_HOST=${{ secrets.MYSQL_HOST }}" >> .env |
29 | | - echo "MYSQL_USER=${{ secrets.MYSQL_USER }}" >> .env |
30 | | - echo "MYSQL_PASSWORD=${{ secrets.MYSQL_PASSWORD }}" >> .env |
31 | | - echo "MYSQL_DATABASE=${{ secrets.MYSQL_DATABASE }}" >> .env |
32 | | - echo "DISCORD_WEBHOOK_URL=${{ secrets.DISCORD_WEBHOOK_URL }}" >> .env |
33 | | - echo "OPEN_API_KEY=${{ secrets.OPEN_API_KEY }}" >> .env |
34 | | - echo "BIZ_INFO_API_KEY=${{ secrets.BIZ_INFO_API_KEY }}" >> .env |
35 | | -
|
36 | | - - name: Export OPENAI_API_KEY as environment variable |
37 | | - run: echo "OPENAI_API_KEY=${{ secrets.OPEN_API_KEY }}" >> $GITHUB_ENV |
38 | | - |
39 | | - - name: Run Crawling Script |
40 | | - run: | |
41 | | - python main.py |
| 1 | +# name: Daily Crawling at 6AM KST |
| 2 | + |
| 3 | +# on: |
| 4 | +# schedule: |
| 5 | +# - cron: "0 21 * * *" # 한국시간 오전 6시 = UTC 21시 |
| 6 | +# workflow_dispatch: |
| 7 | + |
| 8 | +# jobs: |
| 9 | +# crawl: |
| 10 | +# runs-on: ubuntu-latest |
| 11 | + |
| 12 | +# steps: |
| 13 | +# - name: Checkout Repository |
| 14 | +# uses: actions/checkout@v3 |
| 15 | + |
| 16 | +# - name: Set up Python 3.11 |
| 17 | +# uses: actions/setup-python@v4 |
| 18 | +# with: |
| 19 | +# python-version: "3.11" |
| 20 | + |
| 21 | +# - name: Install Python dependencies |
| 22 | +# run: | |
| 23 | +# python -m pip install --upgrade pip |
| 24 | +# pip install -r requirements.txt |
| 25 | + |
| 26 | +# - name: Generate .env file from GitHub Secrets |
| 27 | +# run: | |
| 28 | +# echo "MYSQL_HOST=${{ secrets.MYSQL_HOST }}" >> .env |
| 29 | +# echo "MYSQL_USER=${{ secrets.MYSQL_USER }}" >> .env |
| 30 | +# echo "MYSQL_PASSWORD=${{ secrets.MYSQL_PASSWORD }}" >> .env |
| 31 | +# echo "MYSQL_DATABASE=${{ secrets.MYSQL_DATABASE }}" >> .env |
| 32 | +# echo "DISCORD_WEBHOOK_URL=${{ secrets.DISCORD_WEBHOOK_URL }}" >> .env |
| 33 | +# echo "OPEN_API_KEY=${{ secrets.OPEN_API_KEY }}" >> .env |
| 34 | +# echo "BIZ_INFO_API_KEY=${{ secrets.BIZ_INFO_API_KEY }}" >> .env |
| 35 | + |
| 36 | +# - name: Export OPENAI_API_KEY as environment variable |
| 37 | +# run: echo "OPENAI_API_KEY=${{ secrets.OPEN_API_KEY }}" >> $GITHUB_ENV |
| 38 | + |
| 39 | +# - name: Run Crawling Script |
| 40 | +# run: | |
| 41 | +# python main.py |
0 commit comments