package main import ( "context" "encoding/json" "io" "io/ioutil" "os" "strconv" "sync" "time" "google.golang.org/grpc" "google.golang.org/grpc/metadata" "github.com/cncd/pipeline/pipeline" "github.com/cncd/pipeline/pipeline/backend" "github.com/cncd/pipeline/pipeline/backend/docker" "github.com/cncd/pipeline/pipeline/interrupt" "github.com/cncd/pipeline/pipeline/multipart" "github.com/cncd/pipeline/pipeline/rpc" "github.com/rs/zerolog" "github.com/rs/zerolog/log" "github.com/tevino/abool" "github.com/urfave/cli" oldcontext "golang.org/x/net/context" ) func loop(c *cli.Context) error { filter := rpc.Filter{ Labels: map[string]string{ "platform": c.String("platform"), }, Expr: c.String("drone-filter"), } hostname := c.String("hostname") if len(hostname) == 0 { hostname, _ = os.Hostname() } if c.BoolT("debug") { zerolog.SetGlobalLevel(zerolog.DebugLevel) } else { zerolog.SetGlobalLevel(zerolog.WarnLevel) } // TODO pass version information to grpc server // TODO authenticate to grpc server // grpc.Dial(target, )) conn, err := grpc.Dial( c.String("server"), grpc.WithInsecure(), grpc.WithPerRPCCredentials(&credentials{ username: c.String("username"), password: c.String("password"), }), ) if err != nil { return err } defer conn.Close() client := rpc.NewGrpcClient(conn) sigterm := abool.New() ctx := metadata.NewOutgoingContext( context.Background(), metadata.Pairs("hostname", hostname), ) ctx = interrupt.WithContextFunc(ctx, func() { println("ctrl+c received, terminating process") sigterm.Set() }) var wg sync.WaitGroup parallel := c.Int("max-procs") wg.Add(parallel) for i := 0; i < parallel; i++ { go func() { defer wg.Done() for { if sigterm.IsSet() { return } if err := run(ctx, client, filter); err != nil { log.Error().Err(err).Msg("pipeline done with error") return } } }() } wg.Wait() return nil } // NOTE we need to limit the size of the logs and files that we upload. // The maximum grpc payload size is 4194304. So until we implement streaming // for uploads, we need to set these limits below the maximum. const ( maxLogsUpload = 2000000 // this is per step maxFileUpload = 1000000 ) func run(ctx context.Context, client rpc.Peer, filter rpc.Filter) error { log.Debug(). Msg("request next execution") meta, _ := metadata.FromOutgoingContext(ctx) ctxmeta := metadata.NewOutgoingContext(context.Background(), meta) // get the next job from the queue work, err := client.Next(ctx, filter) if err != nil { return err } if work == nil { return nil } logger := log.With(). Str("repo", extractRepositoryName(work.Config)). Str("build", extractBuildNumber(work.Config)). Str("id", work.ID). Logger() logger.Debug(). Msg("received execution") // new docker engine engine, err := docker.NewEnv() if err != nil { logger.Error(). Err(err). Msg("cannot create docker client") return err } timeout := time.Hour if minutes := work.Timeout; minutes != 0 { timeout = time.Duration(minutes) * time.Minute } ctx, cancel := context.WithTimeout(ctxmeta, timeout) defer cancel() cancelled := abool.New() go func() { logger.Debug(). Msg("listen for cancel signal") if werr := client.Wait(ctx, work.ID); werr != nil { cancelled.SetTo(true) logger.Warn(). Err(werr). Msg("cancel signal received") cancel() } else { logger.Debug(). Msg("stop listening for cancel signal") } }() go func() { for { select { case <-ctx.Done(): logger.Debug(). Msg("pipeline done") return case <-time.After(time.Minute): logger.Debug(). Msg("pipeline lease renewed") client.Extend(ctx, work.ID) } } }() state := rpc.State{} state.Started = time.Now().Unix() err = client.Init(ctxmeta, work.ID, state) if err != nil { logger.Error(). Err(err). Msg("pipeline initialization failed") } var uploads sync.WaitGroup defaultLogger := pipeline.LogFunc(func(proc *backend.Step, rc multipart.Reader) error { loglogger := logger.With(). Str("image", proc.Image). Str("stage", proc.Alias). Logger() part, rerr := rc.NextPart() if rerr != nil { return rerr } uploads.Add(1) var secrets []string for _, secret := range work.Config.Secrets { if secret.Mask { secrets = append(secrets, secret.Value) } } loglogger.Debug().Msg("log stream opened") limitedPart := io.LimitReader(part, maxLogsUpload) logstream := rpc.NewLineWriter(client, work.ID, proc.Alias, secrets...) io.Copy(logstream, limitedPart) loglogger.Debug().Msg("log stream copied") file := &rpc.File{} file.Mime = "application/json+logs" file.Proc = proc.Alias file.Name = "logs.json" file.Data, _ = json.Marshal(logstream.Lines()) file.Size = len(file.Data) file.Time = time.Now().Unix() loglogger.Debug(). Msg("log stream uploading") if serr := client.Upload(ctxmeta, work.ID, file); serr != nil { loglogger.Error(). Err(serr). Msg("log stream upload error") } loglogger.Debug(). Msg("log stream upload complete") defer func() { loglogger.Debug(). Msg("log stream closed") uploads.Done() }() part, rerr = rc.NextPart() if rerr != nil { return nil } // TODO should be configurable limitedPart = io.LimitReader(part, maxFileUpload) file = &rpc.File{} file.Mime = part.Header().Get("Content-Type") file.Proc = proc.Alias file.Name = part.FileName() file.Data, _ = ioutil.ReadAll(limitedPart) file.Size = len(file.Data) file.Time = time.Now().Unix() file.Meta = map[string]string{} for key, value := range part.Header() { file.Meta[key] = value[0] } loglogger.Debug(). Str("file", file.Name). Str("mime", file.Mime). Msg("file stream uploading") if serr := client.Upload(ctxmeta, work.ID, file); serr != nil { loglogger.Error(). Err(serr). Str("file", file.Name). Str("mime", file.Mime). Msg("file stream upload error") } loglogger.Debug(). Str("file", file.Name). Str("mime", file.Mime). Msg("file stream upload complete") return nil }) defaultTracer := pipeline.TraceFunc(func(state *pipeline.State) error { proclogger := logger.With(). Str("image", state.Pipeline.Step.Image). Str("stage", state.Pipeline.Step.Alias). Int("exit_code", state.Process.ExitCode). Bool("exited", state.Process.Exited). Logger() procState := rpc.State{ Proc: state.Pipeline.Step.Alias, Exited: state.Process.Exited, ExitCode: state.Process.ExitCode, Started: time.Now().Unix(), // TODO do not do this Finished: time.Now().Unix(), } defer func() { proclogger.Debug(). Msg("update step status") if uerr := client.Update(ctxmeta, work.ID, procState); uerr != nil { proclogger.Debug(). Err(uerr). Msg("update step status error") } proclogger.Debug(). Msg("update step status complete") }() if state.Process.Exited { return nil } if state.Pipeline.Step.Environment == nil { state.Pipeline.Step.Environment = map[string]string{} } state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "success" state.Pipeline.Step.Environment["CI_BUILD_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10) state.Pipeline.Step.Environment["CI_BUILD_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10) state.Pipeline.Step.Environment["DRONE_BUILD_STATUS"] = "success" state.Pipeline.Step.Environment["DRONE_BUILD_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10) state.Pipeline.Step.Environment["DRONE_BUILD_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10) state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "success" state.Pipeline.Step.Environment["CI_JOB_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10) state.Pipeline.Step.Environment["CI_JOB_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10) state.Pipeline.Step.Environment["DRONE_JOB_STATUS"] = "success" state.Pipeline.Step.Environment["DRONE_JOB_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10) state.Pipeline.Step.Environment["DRONE_JOB_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10) if state.Pipeline.Error != nil { state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "failure" state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "failure" state.Pipeline.Step.Environment["DRONE_BUILD_STATUS"] = "failure" state.Pipeline.Step.Environment["DRONE_JOB_STATUS"] = "failure" } return nil }) err = pipeline.New(work.Config, pipeline.WithContext(ctx), pipeline.WithLogger(defaultLogger), pipeline.WithTracer(defaultTracer), pipeline.WithEngine(engine), ).Run() state.Finished = time.Now().Unix() state.Exited = true if err != nil { switch xerr := err.(type) { case *pipeline.ExitError: state.ExitCode = xerr.Code default: state.ExitCode = 1 state.Error = err.Error() } if cancelled.IsSet() { state.ExitCode = 137 } } logger.Debug(). Str("error", state.Error). Int("exit_code", state.ExitCode). Msg("pipeline complete") logger.Debug(). Msg("uploading logs") uploads.Wait() logger.Debug(). Msg("uploading logs complete") logger.Debug(). Str("error", state.Error). Int("exit_code", state.ExitCode). Msg("updating pipeline status") err = client.Done(ctxmeta, work.ID, state) if err != nil { logger.Error().Err(err). Msg("updating pipeline status failed") } else { logger.Debug(). Msg("updating pipeline status complete") } return nil } type credentials struct { username string password string } func (c *credentials) GetRequestMetadata(oldcontext.Context, ...string) (map[string]string, error) { return map[string]string{ "username": c.username, "password": c.password, }, nil } func (c *credentials) RequireTransportSecurity() bool { return false } // extract repository name from the configuration func extractRepositoryName(config *backend.Config) string { return config.Stages[0].Steps[0].Environment["DRONE_REPO"] } // extract build number from the configuration func extractBuildNumber(config *backend.Config) string { return config.Stages[0].Steps[0].Environment["DRONE_BUILD_NUMBER"] }