mirror of https://code.forgejo.org/forgejo/runner
151 lines
3.4 KiB
Go
151 lines
3.4 KiB
Go
package poller
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"time"
|
|
|
|
"gitea.com/gitea/act_runner/client"
|
|
runnerv1 "gitea.com/gitea/proto-go/runner/v1"
|
|
|
|
"github.com/bufbuild/connect-go"
|
|
log "github.com/sirupsen/logrus"
|
|
)
|
|
|
|
const (
|
|
errorRetryCounterLimit = 3
|
|
errorRetryTimeSleepSecs = 30
|
|
)
|
|
|
|
var ErrDataLock = errors.New("Data Lock Error")
|
|
|
|
func New(cli client.Client, dispatch func(context.Context, *runnerv1.Task) error) *Poller {
|
|
return &Poller{
|
|
Client: cli,
|
|
Dispatch: dispatch,
|
|
routineGroup: newRoutineGroup(),
|
|
metric: &metric{},
|
|
}
|
|
}
|
|
|
|
type Poller struct {
|
|
Client client.Client
|
|
Filter *client.Filter
|
|
Dispatch func(context.Context, *runnerv1.Task) error
|
|
|
|
routineGroup *routineGroup
|
|
metric *metric
|
|
errorRetryCounter int
|
|
}
|
|
|
|
func (p *Poller) Wait() {
|
|
p.routineGroup.Wait()
|
|
}
|
|
|
|
func (p *Poller) Poll(ctx context.Context, n int) error {
|
|
for i := 0; i < n; i++ {
|
|
func(i int) {
|
|
p.routineGroup.Run(func() {
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
log.Infof("stopped the runner: %d", i+1)
|
|
return
|
|
default:
|
|
if ctx.Err() != nil {
|
|
log.Infof("stopping the runner: %d", i+1)
|
|
return
|
|
}
|
|
if err := p.poll(ctx, i+1); err != nil {
|
|
log.WithField("thread", i+1).
|
|
WithError(err).Error("poll error")
|
|
if p.errorRetryCounter > errorRetryCounterLimit {
|
|
log.WithField("thread", i+1).Error("poller: too many errors, sleeping for 30 seconds")
|
|
// FIXME: it makes ctrl+c hang up
|
|
time.Sleep(time.Second * errorRetryTimeSleepSecs)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
})
|
|
}(i)
|
|
}
|
|
p.routineGroup.Wait()
|
|
return nil
|
|
}
|
|
|
|
func (p *Poller) poll(ctx context.Context, thread int) error {
|
|
l := log.WithField("thread", thread)
|
|
l.Info("poller: request stage from remote server")
|
|
|
|
reqCtx, cancel := context.WithTimeout(ctx, 5*time.Second)
|
|
defer cancel()
|
|
|
|
// request a new build stage for execution from the central
|
|
// build server.
|
|
resp, err := p.Client.FetchTask(reqCtx, connect.NewRequest(&runnerv1.FetchTaskRequest{}))
|
|
if err == context.Canceled || err == context.DeadlineExceeded {
|
|
l.WithError(err).Trace("poller: no stage returned")
|
|
p.errorRetryCounter++
|
|
return nil
|
|
}
|
|
|
|
if err != nil && err == ErrDataLock {
|
|
l.WithError(err).Info("task accepted by another runner")
|
|
p.errorRetryCounter++
|
|
return nil
|
|
}
|
|
|
|
if err != nil {
|
|
l.WithError(err).Error("cannot accept task")
|
|
p.errorRetryCounter++
|
|
return err
|
|
}
|
|
|
|
// exit if a nil or empty stage is returned from the system
|
|
// and allow the runner to retry.
|
|
if resp.Msg.Task == nil || resp.Msg.Task.Id == 0 {
|
|
return nil
|
|
}
|
|
|
|
p.errorRetryCounter = 0
|
|
|
|
runCtx, cancel := context.WithTimeout(ctx, time.Hour)
|
|
defer cancel()
|
|
|
|
// update runner status
|
|
// running: idle -> active
|
|
// stopped: active -> idle
|
|
if val := p.metric.IncBusyWorker(); val == 1 {
|
|
if _, err := p.Client.UpdateRunner(
|
|
ctx,
|
|
connect.NewRequest(&runnerv1.UpdateRunnerRequest{
|
|
Status: runnerv1.RunnerStatus_RUNNER_STATUS_ACTIVE,
|
|
}),
|
|
); err != nil {
|
|
return err
|
|
}
|
|
l.Info("update runner status to active")
|
|
}
|
|
|
|
defer func() {
|
|
if val := p.metric.DecBusyWorker(); val != 0 {
|
|
return
|
|
}
|
|
|
|
defer func() {
|
|
if _, err := p.Client.UpdateRunner(
|
|
ctx,
|
|
connect.NewRequest(&runnerv1.UpdateRunnerRequest{
|
|
Status: runnerv1.RunnerStatus_RUNNER_STATUS_IDLE,
|
|
}),
|
|
); err != nil {
|
|
log.Errorln("update status error:", err.Error())
|
|
}
|
|
l.Info("update runner status to idle")
|
|
}()
|
|
}()
|
|
|
|
return p.Dispatch(runCtx, resp.Msg.Task)
|
|
}
|