mirror of https://github.com/0xERR0R/blocky.git
chore(build): Automation workflows (#972)
* Added fork sync workflow(syncs the main branch with upstream every 30 minutes) * Added monthly workflow run deletion(all skipped or canceld and runs older than 30days will be deleted) * Removed Docker image build for main branch on forks * Added documentation how to enable sync and docker build workflows
This commit is contained in:
parent
2b1786a42d
commit
68a8476e48
|
@ -0,0 +1,37 @@
|
||||||
|
name: Delete workflow runs
|
||||||
|
|
||||||
|
on:
|
||||||
|
schedule:
|
||||||
|
- cron: '0 0 1 * *'
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
del_runs:
|
||||||
|
name: Delete workflow runs
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- name: Delete skipped
|
||||||
|
uses: Mattraks/delete-workflow-runs@v2
|
||||||
|
with:
|
||||||
|
token: ${{ github.token }}
|
||||||
|
repository: ${{ github.repository }}
|
||||||
|
retain_days: 0
|
||||||
|
keep_minimum_runs: 0
|
||||||
|
delete_run_by_conclusion_pattern: skipped
|
||||||
|
|
||||||
|
- name: Delete cancelled
|
||||||
|
uses: Mattraks/delete-workflow-runs@v2
|
||||||
|
with:
|
||||||
|
token: ${{ github.token }}
|
||||||
|
repository: ${{ github.repository }}
|
||||||
|
retain_days: 0
|
||||||
|
keep_minimum_runs: 0
|
||||||
|
delete_run_by_conclusion_pattern: cancelled
|
||||||
|
|
||||||
|
- name: Delete workflow runs(older than a month)
|
||||||
|
uses: Mattraks/delete-workflow-runs@v2
|
||||||
|
with:
|
||||||
|
token: ${{ github.token }}
|
||||||
|
repository: ${{ github.repository }}
|
||||||
|
retain_days: 30
|
||||||
|
keep_minimum_runs: 6
|
|
@ -35,15 +35,18 @@ jobs:
|
||||||
|
|
||||||
if [[ "${ENABLED,,}" != "true" ]]; then
|
if [[ "${ENABLED,,}" != "true" ]]; then
|
||||||
echo "enabled=0" >> $GITHUB_OUTPUT
|
echo "enabled=0" >> $GITHUB_OUTPUT
|
||||||
|
|
||||||
echo "Workflow is disabled"
|
echo "Workflow is disabled"
|
||||||
|
|
||||||
echo "### Workflow is disabled" >> $GITHUB_STEP_SUMMARY
|
echo "### Workflow is disabled" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "To enable this workflow by creating a secret 'DEVELOPMENT_DOCKER' with the value 'true'" >> $GITHUB_STEP_SUMMARY
|
echo "To enable this workflow by creating a secret 'DEVELOPMENT_DOCKER' with the value 'true'" >> $GITHUB_STEP_SUMMARY
|
||||||
else
|
else
|
||||||
echo "enabled=1" >> $GITHUB_OUTPUT
|
if [[ "${{ github.repository_owner }}" != "0xERR0R" && "${GITHUB_REF#refs/heads/}" == "main" ]]; then
|
||||||
|
echo "enabled=0" >> $GITHUB_OUTPUT
|
||||||
echo "Workflow is enabled"
|
echo "Workflow is disabled for main branch on forks"
|
||||||
|
else
|
||||||
|
echo "enabled=1" >> $GITHUB_OUTPUT
|
||||||
|
echo "Workflow is enabled"
|
||||||
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
docker:
|
docker:
|
||||||
|
|
|
@ -0,0 +1,36 @@
|
||||||
|
name: Sync Fork
|
||||||
|
|
||||||
|
on:
|
||||||
|
schedule:
|
||||||
|
- cron: '*/30 * * * *'
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: ${{ github.workflow }}
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
sync:
|
||||||
|
name: Sync with Upstream
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
if: github.repository_owner != '0xERR0R'
|
||||||
|
steps:
|
||||||
|
- name: Enabled Check
|
||||||
|
id: check
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
if [[ "${{ secrets.FORK_SYNC_TOKEN }}" != "" ]]; then
|
||||||
|
echo "enabled=1" >> $GITHUB_OUTPUT
|
||||||
|
|
||||||
|
echo "Workflow is enabled"
|
||||||
|
else
|
||||||
|
echo "enabled=0" >> $GITHUB_OUTPUT
|
||||||
|
|
||||||
|
echo "Workflow is disabled(create FORK_SYNC_TOKEN secret with repo write permission to enable it)"
|
||||||
|
fi
|
||||||
|
- name: Sync
|
||||||
|
if: ${{ steps.check.outputs.enabled == 1 }}
|
||||||
|
env:
|
||||||
|
GH_TOKEN: ${{ secrets.FORK_SYNC_TOKEN }}
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
gh repo sync ${{ github.repository }} -b main
|
|
@ -89,6 +89,5 @@ var _ = Describe("Chained grouped cache", func() {
|
||||||
Expect(cache.Contains("both", []string{"group1", "group2"})).Should(ConsistOf("group2"))
|
Expect(cache.Contains("both", []string{"group1", "group2"})).Should(ConsistOf("group2"))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -32,7 +32,6 @@ var _ = Describe("In-Memory grouped cache", func() {
|
||||||
Expect(cache.Contains("searchString", []string{"group1"})).Should(BeEmpty())
|
Expect(cache.Contains("searchString", []string{"group1"})).Should(BeEmpty())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
})
|
})
|
||||||
Describe("Cache creation", func() {
|
Describe("Cache creation", func() {
|
||||||
When("cache with 1 group was created", func() {
|
When("cache with 1 group was created", func() {
|
||||||
|
@ -127,6 +126,5 @@ var _ = Describe("In-Memory grouped cache", func() {
|
||||||
Expect(cache.Contains("both", []string{"group1", "group2"})).Should(ConsistOf("group2"))
|
Expect(cache.Contains("both", []string{"group1", "group2"})).Should(ConsistOf("group2"))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
|
@ -10,9 +10,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("BlockingConfig", func() {
|
var _ = Describe("BlockingConfig", func() {
|
||||||
var (
|
var cfg BlockingConfig
|
||||||
cfg BlockingConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("CachingConfig", func() {
|
var _ = Describe("CachingConfig", func() {
|
||||||
var (
|
var cfg CachingConfig
|
||||||
cfg CachingConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("ClientLookupConfig", func() {
|
var _ = Describe("ClientLookupConfig", func() {
|
||||||
var (
|
var cfg ClientLookupConfig
|
||||||
cfg ClientLookupConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("ConditionalUpstreamConfig", func() {
|
var _ = Describe("ConditionalUpstreamConfig", func() {
|
||||||
var (
|
var cfg ConditionalUpstreamConfig
|
||||||
cfg ConditionalUpstreamConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -10,9 +10,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("CustomDNSConfig", func() {
|
var _ = Describe("CustomDNSConfig", func() {
|
||||||
var (
|
var cfg CustomDNSConfig
|
||||||
cfg CustomDNSConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("FilteringConfig", func() {
|
var _ = Describe("FilteringConfig", func() {
|
||||||
var (
|
var cfg FilteringConfig
|
||||||
cfg FilteringConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("HostsFileConfig", func() {
|
var _ = Describe("HostsFileConfig", func() {
|
||||||
var (
|
var cfg HostsFileConfig
|
||||||
cfg HostsFileConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -7,9 +7,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("MetricsConfig", func() {
|
var _ = Describe("MetricsConfig", func() {
|
||||||
var (
|
var cfg MetricsConfig
|
||||||
cfg MetricsConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -7,9 +7,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("ParallelBestConfig", func() {
|
var _ = Describe("ParallelBestConfig", func() {
|
||||||
var (
|
var cfg ParallelBestConfig
|
||||||
cfg ParallelBestConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("QueryLogConfig", func() {
|
var _ = Describe("QueryLogConfig", func() {
|
||||||
var (
|
var cfg QueryLogConfig
|
||||||
cfg QueryLogConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -7,9 +7,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("RewriterConfig", func() {
|
var _ = Describe("RewriterConfig", func() {
|
||||||
var (
|
var cfg RewriterConfig
|
||||||
cfg RewriterConfig
|
|
||||||
)
|
|
||||||
|
|
||||||
suiteBeforeEach()
|
suiteBeforeEach()
|
||||||
|
|
||||||
|
|
|
@ -108,4 +108,16 @@ Main: [:material-docker:Docker Hub](https://hub.docker.com/r/spx01/blocky)
|
||||||
|
|
||||||
Mirror: [:material-github:GitHub Container Registry](https://ghcr.io/0xerr0r/blocky)
|
Mirror: [:material-github:GitHub Container Registry](https://ghcr.io/0xerr0r/blocky)
|
||||||
|
|
||||||
|
## Developer Information
|
||||||
|
|
||||||
|
### Docker Images
|
||||||
|
|
||||||
|
To enable Docker image creation on a GitHub fork create a secret with the name `DEVELOPMENT_DOCKER` and the value `true`.
|
||||||
|
This will trigger a workflow on every push of a branch starting with `fb-` and create an image with the branch name.
|
||||||
|
|
||||||
|
### Automatic fork sync
|
||||||
|
|
||||||
|
To enable automatic fork synchronisation create a secret with the name `FORK_SYNC_TOKEN` with an access token that has write permission to the fork repository.
|
||||||
|
The enabled workflow will sync the main branch every 30 minutes with its upstream.
|
||||||
|
|
||||||
--8<-- "docs/includes/abbreviations.md"
|
--8<-- "docs/includes/abbreviations.md"
|
||||||
|
|
|
@ -16,7 +16,7 @@ func TryAdapt[From, To any](inner SeriesParser[From], adapt func(From) (To, erro
|
||||||
|
|
||||||
// TryAdaptMethod returns a parser that wraps `inner` and tries to convert each parsed value
|
// TryAdaptMethod returns a parser that wraps `inner` and tries to convert each parsed value
|
||||||
// using the given method with pointer receiver of `To`.
|
// using the given method with pointer receiver of `To`.
|
||||||
func TryAdaptMethod[ToPtr *To, From any, To any](
|
func TryAdaptMethod[ToPtr *To, From, To any](
|
||||||
inner SeriesParser[From], method func(ToPtr, From) error,
|
inner SeriesParser[From], method func(ToPtr, From) error,
|
||||||
) SeriesParser[*To] {
|
) SeriesParser[*To] {
|
||||||
return TryAdapt(inner, func(from From) (*To, error) {
|
return TryAdapt(inner, func(from From) (*To, error) {
|
||||||
|
|
|
@ -11,9 +11,7 @@ import (
|
||||||
|
|
||||||
var _ = Describe("errorFilter", func() {
|
var _ = Describe("errorFilter", func() {
|
||||||
Describe("AllowErrors", func() {
|
Describe("AllowErrors", func() {
|
||||||
var (
|
var parser SeriesParser[struct{}]
|
||||||
parser SeriesParser[struct{}]
|
|
||||||
)
|
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
parser = newMockParser(func(res chan<- struct{}, err chan<- error) {
|
parser = newMockParser(func(res chan<- struct{}, err chan<- error) {
|
||||||
|
|
|
@ -86,9 +86,7 @@ var _ = Describe("Hosts", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
Describe("HostsIterator.ForEachHost", func() {
|
Describe("HostsIterator.ForEachHost", func() {
|
||||||
var (
|
var entry *HostsIterator
|
||||||
entry *HostsIterator
|
|
||||||
)
|
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
sutReader = linesReader(
|
sutReader = linesReader(
|
||||||
|
@ -227,9 +225,7 @@ var _ = Describe("HostsFile", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
Describe("HostsFileEntry.forEachHost", func() {
|
Describe("HostsFileEntry.forEachHost", func() {
|
||||||
var (
|
var entry *HostsFileEntry
|
||||||
entry *HostsFileEntry
|
|
||||||
)
|
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
sutReader = linesReader(
|
sutReader = linesReader(
|
||||||
|
@ -374,9 +370,7 @@ var _ = Describe("HostList", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
Describe("HostListEntry.forEachHost", func() {
|
Describe("HostListEntry.forEachHost", func() {
|
||||||
var (
|
var entry *HostListEntry
|
||||||
entry *HostListEntry
|
|
||||||
)
|
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
sutReader = linesReader(
|
sutReader = linesReader(
|
||||||
|
|
|
@ -11,9 +11,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("ForEach", func() {
|
var _ = Describe("ForEach", func() {
|
||||||
var (
|
var lines SeriesParser[string]
|
||||||
lines SeriesParser[string]
|
|
||||||
)
|
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
lines = Lines(linesReader(
|
lines = Lines(linesReader(
|
||||||
|
|
|
@ -15,9 +15,7 @@ const (
|
||||||
byteBits = 8
|
byteBits = 8
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var ErrInvalidArpaAddrLen = errors.New("arpa hostname is not of expected length")
|
||||||
ErrInvalidArpaAddrLen = errors.New("arpa hostname is not of expected length")
|
|
||||||
)
|
|
||||||
|
|
||||||
func ParseIPFromArpaAddr(arpa string) (net.IP, error) {
|
func ParseIPFromArpaAddr(arpa string) (net.IP, error) {
|
||||||
if strings.HasSuffix(arpa, IPv4PtrSuffix) {
|
if strings.HasSuffix(arpa, IPv4PtrSuffix) {
|
||||||
|
|
Loading…
Reference in New Issue