mirror of
https://github.com/PluralKit/PluralKit.git
synced 2026-02-14 01:30:13 +00:00
test
This commit is contained in:
parent
73de67d7e2
commit
9892cd9a3f
3 changed files with 13 additions and 13 deletions
2
.github/workflows/ci.yml
vendored
2
.github/workflows/ci.yml
vendored
|
|
@ -31,7 +31,7 @@ jobs:
|
||||||
|
|
||||||
GITHUB_APP_TOKEN: ${{ secrets.COMMIT_STATUS_TOKEN }}
|
GITHUB_APP_TOKEN: ${{ secrets.COMMIT_STATUS_TOKEN }}
|
||||||
DISCORD_WEBHOOK: ${{ secrets.DISCORD_WEBHOOK }}
|
DISCORD_WEBHOOK: ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
CUR_SHA: ${{ github.sha }}
|
GIT_SHA: ${{ github.sha }}
|
||||||
|
|
||||||
REPO_URL: https://github.com/${{ github.repository }}
|
REPO_URL: https://github.com/${{ github.repository }}
|
||||||
OLD_SHA: ${{ github.event.before }}
|
OLD_SHA: ${{ github.event.before }}
|
||||||
|
|
|
||||||
|
|
@ -50,11 +50,11 @@ def report_status(name, start_time, exit=None):
|
||||||
'Authorization': f'Bearer {must_get_env("GITHUB_APP_TOKEN")}',
|
'Authorization': f'Bearer {must_get_env("GITHUB_APP_TOKEN")}',
|
||||||
'content-type':'application/json'
|
'content-type':'application/json'
|
||||||
},
|
},
|
||||||
data=json.dumps(data)
|
data=bytes(json.dumps(data), 'UTF-8')
|
||||||
)
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
with urllib.request.urlopen(request) as response:
|
with urllib.request.urlopen(req) as response:
|
||||||
response_code = response.getcode()
|
response_code = response.getcode()
|
||||||
response_data = response.read()
|
response_data = response.read()
|
||||||
print(f"{response_code} updated status {data}: {response_data}")
|
print(f"{response_code} updated status {data}: {response_data}")
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,6 @@
|
||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
import os, sys, json, subprocess, random, time
|
import os, sys, json, subprocess, random, time, re
|
||||||
import urllib.request
|
import urllib.request
|
||||||
|
|
||||||
def must_get_env(name):
|
def must_get_env(name):
|
||||||
|
|
@ -22,18 +22,18 @@ def spawn_job(name):
|
||||||
'Authorization': f'Bearer {must_get_env("GITHUB_APP_TOKEN")}',
|
'Authorization': f'Bearer {must_get_env("GITHUB_APP_TOKEN")}',
|
||||||
'content-type':'application/json'
|
'content-type':'application/json'
|
||||||
},
|
},
|
||||||
data=json.dumps({
|
data=bytes(json.dumps({
|
||||||
'ref': must_get_env("GIT_SHA"),
|
'ref': must_get_env("GIT_SHA"),
|
||||||
'inputs': {
|
'inputs': {
|
||||||
'dispatchData': json.dumps({
|
'dispatchData': json.dumps({
|
||||||
'action': name,
|
'action': name,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
})
|
}), 'UTF-8')
|
||||||
)
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
with urllib.request.urlopen(request) as response:
|
with urllib.request.urlopen(req) as response:
|
||||||
response_code = response.getcode()
|
response_code = response.getcode()
|
||||||
response_data = response.read()
|
response_data = response.read()
|
||||||
print(f"{response_code} spawned job {name}: {response_data}")
|
print(f"{response_code} spawned job {name}: {response_data}")
|
||||||
|
|
@ -79,16 +79,16 @@ def create_jobs():
|
||||||
jobs = jobs | ["format_cs", "format_rs"]
|
jobs = jobs | ["format_cs", "format_rs"]
|
||||||
else:
|
else:
|
||||||
for key in modify_regexes.keys():
|
for key in modify_regexes.keys():
|
||||||
if re.match(key, changed_file, flags=re.MULTILINE) is not None:
|
if re.match(key, str(changed_files), flags=re.MULTILINE) is not None:
|
||||||
jobs = jobs | modify_regexes[key]
|
jobs = jobs | set(modify_regexes[key])
|
||||||
|
|
||||||
for key in changes:
|
for key in jobs:
|
||||||
if aliases.get(key) is not None:
|
if aliases.get(key) is not None:
|
||||||
jobs = jobs | aliases[key]
|
jobs = jobs | set(aliases[key])
|
||||||
jobs = jobs - [key]
|
jobs = jobs - set([key])
|
||||||
|
|
||||||
# test
|
# test
|
||||||
jobs = jobs | ["test"]
|
jobs = jobs | set(["test"])
|
||||||
|
|
||||||
# do this in a tx or something
|
# do this in a tx or something
|
||||||
for job in jobs:
|
for job in jobs:
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue