mirror of
https://github.com/woodpecker-ci/woodpecker.git
synced 2024-11-26 11:51:02 +00:00
put updated agent behind flag
This commit is contained in:
parent
d7758a23e0
commit
623be6d830
58 changed files with 4824 additions and 23 deletions
|
@ -1,6 +1,8 @@
|
||||||
package agent
|
package agent
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"fmt"
|
||||||
|
"math"
|
||||||
"os"
|
"os"
|
||||||
"os/signal"
|
"os/signal"
|
||||||
"strings"
|
"strings"
|
||||||
|
@ -60,7 +62,7 @@ var AgentCmd = cli.Command{
|
||||||
Value: "amd64",
|
Value: "amd64",
|
||||||
},
|
},
|
||||||
cli.StringFlag{
|
cli.StringFlag{
|
||||||
EnvVar: "DRONE_SERVER",
|
EnvVar: "DRONE_SERVER,DRONE_ENDPOINT",
|
||||||
Name: "drone-server",
|
Name: "drone-server",
|
||||||
Usage: "drone server address",
|
Usage: "drone server address",
|
||||||
Value: "ws://localhost:8000/ws/broker",
|
Value: "ws://localhost:8000/ws/broker",
|
||||||
|
@ -138,11 +140,55 @@ var AgentCmd = cli.Command{
|
||||||
Name: "extension",
|
Name: "extension",
|
||||||
Usage: "custom plugin extension endpoint",
|
Usage: "custom plugin extension endpoint",
|
||||||
},
|
},
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
cli.BoolFlag{
|
||||||
|
EnvVar: "DRONE_CANARY",
|
||||||
|
Name: "canary",
|
||||||
|
Usage: "enable experimental features at your own risk",
|
||||||
|
},
|
||||||
|
|
||||||
|
// cli.StringFlag{
|
||||||
|
// Name: "endpoint",
|
||||||
|
// EnvVar: "DRONE_ENDPOINT,DRONE_SERVER",
|
||||||
|
// Value: "ws://localhost:9999/ws/rpc",
|
||||||
|
// },
|
||||||
|
// cli.DurationFlag{
|
||||||
|
// Name: "backoff",
|
||||||
|
// EnvVar: "DRONE_BACKOFF",
|
||||||
|
// Value: time.Second * 15,
|
||||||
|
// },
|
||||||
|
cli.IntFlag{
|
||||||
|
Name: "retry-limit",
|
||||||
|
EnvVar: "DRONE_RETRY_LIMIT",
|
||||||
|
Value: math.MaxInt32,
|
||||||
|
},
|
||||||
|
cli.IntFlag{
|
||||||
|
Name: "max-procs",
|
||||||
|
EnvVar: "DRONE_MAX_PROCS",
|
||||||
|
Value: 1,
|
||||||
|
},
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "platform",
|
||||||
|
EnvVar: "DRONE_PLATFORM",
|
||||||
|
Value: "linux/amd64",
|
||||||
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
func start(c *cli.Context) {
|
func start(c *cli.Context) {
|
||||||
|
|
||||||
|
if c.Bool("canary") {
|
||||||
|
if err := loop(c); err != nil {
|
||||||
|
fmt.Println(err)
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
log := redlog.New(os.Stderr)
|
log := redlog.New(os.Stderr)
|
||||||
log.SetLevel(0)
|
log.SetLevel(0)
|
||||||
logger.SetLogger(log)
|
logger.SetLogger(log)
|
||||||
|
@ -187,7 +233,7 @@ func start(c *cli.Context) {
|
||||||
client.Ack(m.Ack)
|
client.Ack(m.Ack)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
r := pipeline{
|
r := pipelinet{
|
||||||
drone: client,
|
drone: client,
|
||||||
docker: docker,
|
docker: docker,
|
||||||
config: config{
|
config: config{
|
||||||
|
|
|
@ -22,13 +22,13 @@ type config struct {
|
||||||
extension []string
|
extension []string
|
||||||
}
|
}
|
||||||
|
|
||||||
type pipeline struct {
|
type pipelinet struct {
|
||||||
drone *stomp.Client
|
drone *stomp.Client
|
||||||
docker dockerclient.Client
|
docker dockerclient.Client
|
||||||
config config
|
config config
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *pipeline) run(w *model.Work) {
|
func (r *pipelinet) run(w *model.Work) {
|
||||||
|
|
||||||
// defer func() {
|
// defer func() {
|
||||||
// // r.drone.Ack(id, opts)
|
// // r.drone.Ack(id, opts)
|
||||||
|
|
191
drone/agent/exp.go
Normal file
191
drone/agent/exp.go
Normal file
|
@ -0,0 +1,191 @@
|
||||||
|
package agent
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"log"
|
||||||
|
"net/url"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline"
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend/docker"
|
||||||
|
"github.com/cncd/pipeline/pipeline/interrupt"
|
||||||
|
"github.com/cncd/pipeline/pipeline/multipart"
|
||||||
|
"github.com/cncd/pipeline/pipeline/rpc"
|
||||||
|
|
||||||
|
"github.com/codegangsta/cli"
|
||||||
|
"github.com/tevino/abool"
|
||||||
|
)
|
||||||
|
|
||||||
|
func loop(c *cli.Context) error {
|
||||||
|
endpoint, err := url.Parse(
|
||||||
|
c.String("drone-server"),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
client, err := rpc.NewClient(
|
||||||
|
endpoint.String(),
|
||||||
|
rpc.WithRetryLimit(
|
||||||
|
c.Int("retry-limit"),
|
||||||
|
),
|
||||||
|
rpc.WithBackoff(
|
||||||
|
c.Duration("backoff"),
|
||||||
|
),
|
||||||
|
rpc.WithToken(
|
||||||
|
c.String("drone-secret"),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer client.Close()
|
||||||
|
|
||||||
|
sigterm := abool.New()
|
||||||
|
ctx := context.Background()
|
||||||
|
ctx = interrupt.WithContextFunc(ctx, func() {
|
||||||
|
println("ctrl+c received, terminating process")
|
||||||
|
sigterm.Set()
|
||||||
|
})
|
||||||
|
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
parallel := c.Int("max-procs")
|
||||||
|
wg.Add(parallel)
|
||||||
|
|
||||||
|
for i := 0; i < parallel; i++ {
|
||||||
|
go func() {
|
||||||
|
defer wg.Done()
|
||||||
|
for {
|
||||||
|
if sigterm.IsSet() {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if err := run(ctx, client); err != nil {
|
||||||
|
log.Printf("build runner encountered error: exiting: %s", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
wg.Wait()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func run(ctx context.Context, client rpc.Peer) error {
|
||||||
|
log.Println("pipeline: request next execution")
|
||||||
|
|
||||||
|
// get the next job from the queue
|
||||||
|
work, err := client.Next(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if work == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
log.Printf("pipeline: received next execution: %s", work.ID)
|
||||||
|
|
||||||
|
// new docker engine
|
||||||
|
engine, err := docker.NewEnv()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
timeout := time.Hour
|
||||||
|
if minutes := work.Timeout; minutes != 0 {
|
||||||
|
timeout = time.Duration(minutes) * time.Minute
|
||||||
|
}
|
||||||
|
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
cancelled := abool.New()
|
||||||
|
go func() {
|
||||||
|
ok, _ := client.Notify(ctx, work.ID)
|
||||||
|
if ok {
|
||||||
|
cancelled.SetTo(true)
|
||||||
|
log.Printf("pipeline: cancel signal received: %s", work.ID)
|
||||||
|
cancel()
|
||||||
|
} else {
|
||||||
|
log.Printf("pipeline: cancel channel closed: %s", work.ID)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
log.Printf("pipeline: cancel ping loop: %s", work.ID)
|
||||||
|
return
|
||||||
|
case <-time.After(time.Minute):
|
||||||
|
log.Printf("pipeline: ping queue: %s", work.ID)
|
||||||
|
client.Extend(ctx, work.ID)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
state := rpc.State{}
|
||||||
|
state.Started = time.Now().Unix()
|
||||||
|
err = client.Update(context.Background(), work.ID, state)
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("pipeline: error updating pipeline status: %s: %s", work.ID, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
defaultLogger := pipeline.LogFunc(func(proc *backend.Step, rc multipart.Reader) error {
|
||||||
|
part, rerr := rc.NextPart()
|
||||||
|
if rerr != nil {
|
||||||
|
return rerr
|
||||||
|
}
|
||||||
|
writer := rpc.NewLineWriter(client, work.ID, proc.Alias)
|
||||||
|
io.Copy(writer, part)
|
||||||
|
|
||||||
|
defer func() {
|
||||||
|
log.Printf("pipeline: finish uploading logs: %s: step %s", work.ID, proc.Alias)
|
||||||
|
}()
|
||||||
|
|
||||||
|
part, rerr = rc.NextPart()
|
||||||
|
if rerr != nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
mime := part.Header().Get("Content-Type")
|
||||||
|
if serr := client.Save(context.Background(), work.ID, mime, part); serr != nil {
|
||||||
|
log.Printf("pipeline: cannot upload artifact: %s: %s: %s", work.ID, mime, serr)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
err = pipeline.New(work.Config,
|
||||||
|
pipeline.WithContext(ctx),
|
||||||
|
pipeline.WithLogger(defaultLogger),
|
||||||
|
pipeline.WithTracer(pipeline.DefaultTracer),
|
||||||
|
pipeline.WithEngine(engine),
|
||||||
|
).Run()
|
||||||
|
|
||||||
|
state.Finished = time.Now().Unix()
|
||||||
|
state.Exited = true
|
||||||
|
if err != nil {
|
||||||
|
state.Error = err.Error()
|
||||||
|
if xerr, ok := err.(*pipeline.ExitError); ok {
|
||||||
|
state.ExitCode = xerr.Code
|
||||||
|
}
|
||||||
|
if xerr, ok := err.(*pipeline.OomError); ok {
|
||||||
|
state.ExitCode = xerr.Code
|
||||||
|
}
|
||||||
|
if cancelled.IsSet() {
|
||||||
|
state.ExitCode = 130
|
||||||
|
} else if state.ExitCode == 0 {
|
||||||
|
state.ExitCode = 1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Printf("pipeline: execution complete: %s", work.ID)
|
||||||
|
|
||||||
|
err = client.Update(context.Background(), work.ID, state)
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("Pipeine: error updating pipeline status: %s: %s", work.ID, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
|
@ -2,6 +2,7 @@ package router
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"os"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
|
|
||||||
|
@ -119,9 +120,15 @@ func Load(middleware ...gin.HandlerFunc) http.Handler {
|
||||||
badges.GET("/cc.xml", server.GetCC)
|
badges.GET("/cc.xml", server.GetCC)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if os.Getenv("DRONE_CANARY") == "" {
|
||||||
e.POST("/hook", server.PostHook)
|
e.POST("/hook", server.PostHook)
|
||||||
e.POST("/api/hook", server.PostHook)
|
e.POST("/api/hook", server.PostHook)
|
||||||
|
} else {
|
||||||
|
e.POST("/hook", server.PostHook2)
|
||||||
|
e.POST("/api/hook", server.PostHook2)
|
||||||
|
}
|
||||||
|
|
||||||
|
if os.Getenv("DRONE_CANARY") == "" {
|
||||||
ws := e.Group("/ws")
|
ws := e.Group("/ws")
|
||||||
{
|
{
|
||||||
ws.GET("/broker", server.Broker)
|
ws.GET("/broker", server.Broker)
|
||||||
|
@ -133,6 +140,27 @@ func Load(middleware ...gin.HandlerFunc) http.Handler {
|
||||||
server.LogStream,
|
server.LogStream,
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
ws := e.Group("/ws")
|
||||||
|
{
|
||||||
|
ws.GET("/broker", server.RPCHandler)
|
||||||
|
ws.GET("/rpc", server.RPCHandler)
|
||||||
|
ws.GET("/feed", server.EventStream2)
|
||||||
|
ws.GET("/logs/:owner/:name/:build/:number",
|
||||||
|
session.SetRepo(),
|
||||||
|
session.SetPerm(),
|
||||||
|
session.MustPull,
|
||||||
|
server.LogStream2,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
info := e.Group("/api/info")
|
||||||
|
{
|
||||||
|
info.GET("/queue",
|
||||||
|
session.MustAdmin(),
|
||||||
|
server.GetQueueInfo,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
auth := e.Group("/authorize")
|
auth := e.Group("/authorize")
|
||||||
{
|
{
|
||||||
|
|
412
server/hook.go
412
server/hook.go
|
@ -1,9 +1,13 @@
|
||||||
package server
|
package server
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"net/url"
|
||||||
"regexp"
|
"regexp"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/square/go-jose"
|
"github.com/square/go-jose"
|
||||||
|
@ -15,7 +19,18 @@ import (
|
||||||
"github.com/drone/drone/shared/token"
|
"github.com/drone/drone/shared/token"
|
||||||
"github.com/drone/drone/store"
|
"github.com/drone/drone/store"
|
||||||
"github.com/drone/drone/yaml"
|
"github.com/drone/drone/yaml"
|
||||||
|
"github.com/drone/envsubst"
|
||||||
"github.com/drone/mq/stomp"
|
"github.com/drone/mq/stomp"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend"
|
||||||
|
yaml2 "github.com/cncd/pipeline/pipeline/frontend/yaml"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml/compiler"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml/linter"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml/matrix"
|
||||||
|
"github.com/cncd/pipeline/pipeline/rpc"
|
||||||
|
"github.com/cncd/pubsub"
|
||||||
|
"github.com/cncd/queue"
|
||||||
)
|
)
|
||||||
|
|
||||||
var skipRe = regexp.MustCompile(`\[(?i:ci *skip|skip *ci)\]`)
|
var skipRe = regexp.MustCompile(`\[(?i:ci *skip|skip *ci)\]`)
|
||||||
|
@ -105,8 +120,8 @@ func PostHook(c *gin.Context) {
|
||||||
// a small number of people will probably be upset by this, I'm not sure
|
// a small number of people will probably be upset by this, I'm not sure
|
||||||
// it is actually that big of a deal.
|
// it is actually that big of a deal.
|
||||||
if len(build.Email) == 0 {
|
if len(build.Email) == 0 {
|
||||||
author, err := store.GetUserLogin(c, build.Author)
|
author, uerr := store.GetUserLogin(c, build.Author)
|
||||||
if err == nil {
|
if uerr == nil {
|
||||||
build.Email = author.Email
|
build.Email = author.Email
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -164,9 +179,9 @@ func PostHook(c *gin.Context) {
|
||||||
log.Debugf("cannot parse .drone.yml.sig file. empty file")
|
log.Debugf("cannot parse .drone.yml.sig file. empty file")
|
||||||
} else {
|
} else {
|
||||||
build.Signed = true
|
build.Signed = true
|
||||||
output, err := signature.Verify([]byte(repo.Hash))
|
output, verr := signature.Verify([]byte(repo.Hash))
|
||||||
if err != nil {
|
if verr != nil {
|
||||||
log.Debugf("cannot verify .drone.yml.sig file. %s", err)
|
log.Debugf("cannot verify .drone.yml.sig file. %s", verr)
|
||||||
} else if string(output) != string(raw) {
|
} else if string(output) != string(raw) {
|
||||||
log.Debugf("cannot verify .drone.yml.sig file. no match")
|
log.Debugf("cannot verify .drone.yml.sig file. no match")
|
||||||
} else {
|
} else {
|
||||||
|
@ -212,7 +227,7 @@ func PostHook(c *gin.Context) {
|
||||||
}
|
}
|
||||||
|
|
||||||
client := stomp.MustFromContext(c)
|
client := stomp.MustFromContext(c)
|
||||||
client.SendJSON("/topic/events", model.Event{
|
client.SendJSON("topic/events", model.Event{
|
||||||
Type: model.Enqueued,
|
Type: model.Enqueued,
|
||||||
Repo: *repo,
|
Repo: *repo,
|
||||||
Build: *build,
|
Build: *build,
|
||||||
|
@ -247,3 +262,388 @@ func PostHook(c *gin.Context) {
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
// CANARY IMPLEMENTATION
|
||||||
|
//
|
||||||
|
// This file is a complete disaster because I'm trying to wedge in some
|
||||||
|
// experimental code. Please pardon our appearance during renovations.
|
||||||
|
//
|
||||||
|
|
||||||
|
func GetQueueInfo(c *gin.Context) {
|
||||||
|
c.IndentedJSON(200,
|
||||||
|
config.queue.Info(c),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
func PostHook2(c *gin.Context) {
|
||||||
|
remote_ := remote.FromContext(c)
|
||||||
|
|
||||||
|
tmprepo, build, err := remote_.Hook(c.Request)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to parse hook. %s", err)
|
||||||
|
c.AbortWithError(400, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if build == nil {
|
||||||
|
c.Writer.WriteHeader(200)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if tmprepo == nil {
|
||||||
|
log.Errorf("failure to ascertain repo from hook.")
|
||||||
|
c.Writer.WriteHeader(400)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// skip the build if any case-insensitive combination of the words "skip" and "ci"
|
||||||
|
// wrapped in square brackets appear in the commit message
|
||||||
|
skipMatch := skipRe.FindString(build.Message)
|
||||||
|
if len(skipMatch) > 0 {
|
||||||
|
log.Infof("ignoring hook. %s found in %s", skipMatch, build.Commit)
|
||||||
|
c.Writer.WriteHeader(204)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
repo, err := store.GetRepoOwnerName(c, tmprepo.Owner, tmprepo.Name)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to find repo %s/%s from hook. %s", tmprepo.Owner, tmprepo.Name, err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// get the token and verify the hook is authorized
|
||||||
|
parsed, err := token.ParseRequest(c.Request, func(t *token.Token) (string, error) {
|
||||||
|
return repo.Hash, nil
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to parse token from hook for %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(400, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if parsed.Text != repo.FullName {
|
||||||
|
log.Errorf("failure to verify token from hook. Expected %s, got %s", repo.FullName, parsed.Text)
|
||||||
|
c.AbortWithStatus(403)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if repo.UserID == 0 {
|
||||||
|
log.Warnf("ignoring hook. repo %s has no owner.", repo.FullName)
|
||||||
|
c.Writer.WriteHeader(204)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var skipped = true
|
||||||
|
if (build.Event == model.EventPush && repo.AllowPush) ||
|
||||||
|
(build.Event == model.EventPull && repo.AllowPull) ||
|
||||||
|
(build.Event == model.EventDeploy && repo.AllowDeploy) ||
|
||||||
|
(build.Event == model.EventTag && repo.AllowTag) {
|
||||||
|
skipped = false
|
||||||
|
}
|
||||||
|
|
||||||
|
if skipped {
|
||||||
|
log.Infof("ignoring hook. repo %s is disabled for %s events.", repo.FullName, build.Event)
|
||||||
|
c.Writer.WriteHeader(204)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
user, err := store.GetUser(c, repo.UserID)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to find repo owner %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(500, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// if there is no email address associated with the pull request,
|
||||||
|
// we lookup the email address based on the authors github login.
|
||||||
|
//
|
||||||
|
// my initial hesitation with this code is that it has the ability
|
||||||
|
// to expose your email address. At the same time, your email address
|
||||||
|
// is already exposed in the public .git log. So while some people will
|
||||||
|
// a small number of people will probably be upset by this, I'm not sure
|
||||||
|
// it is actually that big of a deal.
|
||||||
|
if len(build.Email) == 0 {
|
||||||
|
author, uerr := store.GetUserLogin(c, build.Author)
|
||||||
|
if uerr == nil {
|
||||||
|
build.Email = author.Email
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// if the remote has a refresh token, the current access token
|
||||||
|
// may be stale. Therefore, we should refresh prior to dispatching
|
||||||
|
// the job.
|
||||||
|
if refresher, ok := remote_.(remote.Refresher); ok {
|
||||||
|
ok, _ := refresher.Refresh(user)
|
||||||
|
if ok {
|
||||||
|
store.UpdateUser(c, user)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// fetch the build file from the database
|
||||||
|
cfg := ToConfig(c)
|
||||||
|
raw, err := remote_.File(user, repo, build, cfg.Yaml)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to get build config for %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
sec, err := remote_.File(user, repo, build, cfg.Shasum)
|
||||||
|
if err != nil {
|
||||||
|
log.Debugf("cannot find build secrets for %s. %s", repo.FullName, err)
|
||||||
|
// NOTE we don't exit on failure. The sec file is optional
|
||||||
|
}
|
||||||
|
|
||||||
|
axes, err := matrix.Parse(raw)
|
||||||
|
if err != nil {
|
||||||
|
c.String(500, "Failed to parse yaml file or calculate matrix. %s", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if len(axes) == 0 {
|
||||||
|
axes = append(axes, matrix.Axis{})
|
||||||
|
}
|
||||||
|
|
||||||
|
netrc, err := remote_.Netrc(user, repo)
|
||||||
|
if err != nil {
|
||||||
|
c.String(500, "Failed to generate netrc file. %s", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// verify the branches can be built vs skipped
|
||||||
|
branches, err := yaml2.ParseBytes(raw)
|
||||||
|
if err != nil {
|
||||||
|
c.String(500, "Failed to parse yaml file. %s", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if !branches.Branches.Match(build.Branch) && build.Event != model.EventTag && build.Event != model.EventDeploy {
|
||||||
|
c.String(200, "Branch does not match restrictions defined in yaml")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
signature, err := jose.ParseSigned(string(sec))
|
||||||
|
if err != nil {
|
||||||
|
log.Debugf("cannot parse .drone.yml.sig file. %s", err)
|
||||||
|
} else if len(sec) == 0 {
|
||||||
|
log.Debugf("cannot parse .drone.yml.sig file. empty file")
|
||||||
|
} else {
|
||||||
|
build.Signed = true
|
||||||
|
output, verr := signature.Verify([]byte(repo.Hash))
|
||||||
|
if verr != nil {
|
||||||
|
log.Debugf("cannot verify .drone.yml.sig file. %s", verr)
|
||||||
|
} else if string(output) != string(raw) {
|
||||||
|
log.Debugf("cannot verify .drone.yml.sig file. no match")
|
||||||
|
} else {
|
||||||
|
build.Verified = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// update some build fields
|
||||||
|
build.Status = model.StatusPending
|
||||||
|
build.RepoID = repo.ID
|
||||||
|
|
||||||
|
// and use a transaction
|
||||||
|
var jobs []*model.Job
|
||||||
|
for num, axis := range axes {
|
||||||
|
jobs = append(jobs, &model.Job{
|
||||||
|
BuildID: build.ID,
|
||||||
|
Number: num + 1,
|
||||||
|
Status: model.StatusPending,
|
||||||
|
Environment: axis,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
err = store.CreateBuild(c, build, jobs...)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to save commit for %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(500, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
c.JSON(200, build)
|
||||||
|
|
||||||
|
uri := fmt.Sprintf("%s/%s/%d", httputil.GetURL(c.Request), repo.FullName, build.Number)
|
||||||
|
err = remote_.Status(user, repo, build, uri)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("error setting commit status for %s/%d", repo.FullName, build.Number)
|
||||||
|
}
|
||||||
|
|
||||||
|
// get the previous build so that we can send
|
||||||
|
// on status change notifications
|
||||||
|
last, _ := store.GetBuildLastBefore(c, repo, build.Branch, build.ID)
|
||||||
|
secs, err := store.GetMergedSecretList(c, repo)
|
||||||
|
if err != nil {
|
||||||
|
log.Debugf("Error getting secrets for %s#%d. %s", repo.FullName, build.Number, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
// new code here
|
||||||
|
//
|
||||||
|
|
||||||
|
message := pubsub.Message{}
|
||||||
|
message.Data, _ = json.Marshal(model.Event{
|
||||||
|
Type: model.Enqueued,
|
||||||
|
Repo: *repo,
|
||||||
|
Build: *build,
|
||||||
|
})
|
||||||
|
message.Labels = map[string]string{
|
||||||
|
"repo": repo.FullName,
|
||||||
|
"private": strconv.FormatBool(repo.IsPrivate),
|
||||||
|
}
|
||||||
|
// TODO remove global reference
|
||||||
|
config.pubsub.Publish(c, "topic/events", message)
|
||||||
|
|
||||||
|
//
|
||||||
|
// workspace
|
||||||
|
//
|
||||||
|
|
||||||
|
var (
|
||||||
|
link, _ = url.Parse(repo.Link)
|
||||||
|
base = "/drone"
|
||||||
|
path = "src/" + link.Host + "/" + repo.FullName
|
||||||
|
)
|
||||||
|
|
||||||
|
for _, job := range jobs {
|
||||||
|
|
||||||
|
metadata := metadataFromStruct(repo, build, last, job, "linux/amd64")
|
||||||
|
environ := metadata.Environ()
|
||||||
|
|
||||||
|
secrets := map[string]string{}
|
||||||
|
for _, sec := range secs {
|
||||||
|
if !sec.MatchEvent(build.Event) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if build.Verified || sec.SkipVerify {
|
||||||
|
secrets[sec.Name] = sec.Value
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sub := func(name string) string {
|
||||||
|
if v, ok := environ[name]; ok {
|
||||||
|
return v
|
||||||
|
}
|
||||||
|
return secrets[name]
|
||||||
|
}
|
||||||
|
if s, err := envsubst.Eval(string(raw), sub); err != nil {
|
||||||
|
raw = []byte(s)
|
||||||
|
}
|
||||||
|
parsed, err := yaml2.ParseBytes(raw)
|
||||||
|
if err != nil {
|
||||||
|
// TODO
|
||||||
|
}
|
||||||
|
|
||||||
|
lerr := linter.New(
|
||||||
|
linter.WithTrusted(repo.IsTrusted),
|
||||||
|
).Lint(parsed)
|
||||||
|
if lerr != nil {
|
||||||
|
// TODO
|
||||||
|
}
|
||||||
|
|
||||||
|
ir := compiler.New(
|
||||||
|
compiler.WithEnviron(environ),
|
||||||
|
compiler.WithEscalated("plugins/docker", "plugins/gcr", "plugins/ecr"),
|
||||||
|
compiler.WithLocal(false),
|
||||||
|
compiler.WithNetrc(netrc.Login, netrc.Password, netrc.Machine),
|
||||||
|
compiler.WithPrefix(
|
||||||
|
fmt.Sprintf(
|
||||||
|
"%d_%d",
|
||||||
|
job.ID,
|
||||||
|
time.Now().Unix(),
|
||||||
|
),
|
||||||
|
),
|
||||||
|
compiler.WithProxy(),
|
||||||
|
compiler.WithVolumes(), // todo set global volumes
|
||||||
|
compiler.WithWorkspace(base, path),
|
||||||
|
).Compile(parsed)
|
||||||
|
|
||||||
|
task := new(queue.Task)
|
||||||
|
task.ID = fmt.Sprint(job.ID)
|
||||||
|
task.Labels = map[string]string{}
|
||||||
|
task.Labels["platform"] = "linux/amd64"
|
||||||
|
// TODO set proper platform
|
||||||
|
// TODO set proper labels
|
||||||
|
task.Data, _ = json.Marshal(rpc.Pipeline{
|
||||||
|
ID: fmt.Sprint(job.ID),
|
||||||
|
Config: ir,
|
||||||
|
Timeout: repo.Timeout,
|
||||||
|
})
|
||||||
|
|
||||||
|
config.logger.Open(context.Background(), task.ID)
|
||||||
|
config.queue.Push(context.Background(), task)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// use helper funciton to return ([]backend.Config, error)
|
||||||
|
|
||||||
|
type builder struct {
|
||||||
|
secs []*model.Secret
|
||||||
|
repo *model.Repo
|
||||||
|
build *model.Build
|
||||||
|
last *model.Build
|
||||||
|
jobs []*model.Job
|
||||||
|
link string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *builder) Build() ([]*backend.Config, error) {
|
||||||
|
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// return the metadata from the cli context.
|
||||||
|
func metadataFromStruct(repo *model.Repo, build, last *model.Build, job *model.Job, link string) frontend.Metadata {
|
||||||
|
return frontend.Metadata{
|
||||||
|
Repo: frontend.Repo{
|
||||||
|
Name: repo.Name,
|
||||||
|
Link: repo.Link,
|
||||||
|
Remote: repo.Clone,
|
||||||
|
Private: repo.IsPrivate,
|
||||||
|
},
|
||||||
|
Curr: frontend.Build{
|
||||||
|
Number: build.Number,
|
||||||
|
Created: build.Created,
|
||||||
|
Started: build.Started,
|
||||||
|
Finished: build.Finished,
|
||||||
|
Status: build.Status,
|
||||||
|
Event: build.Event,
|
||||||
|
Link: build.Link,
|
||||||
|
Target: build.Deploy,
|
||||||
|
Commit: frontend.Commit{
|
||||||
|
Sha: build.Commit,
|
||||||
|
Ref: build.Ref,
|
||||||
|
Refspec: build.Refspec,
|
||||||
|
Branch: build.Branch,
|
||||||
|
Message: build.Message,
|
||||||
|
Author: frontend.Author{
|
||||||
|
Name: build.Author,
|
||||||
|
Email: build.Email,
|
||||||
|
Avatar: build.Avatar,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Prev: frontend.Build{
|
||||||
|
Number: last.Number,
|
||||||
|
Created: last.Created,
|
||||||
|
Started: last.Started,
|
||||||
|
Finished: last.Finished,
|
||||||
|
Status: last.Status,
|
||||||
|
Event: last.Event,
|
||||||
|
Link: last.Link,
|
||||||
|
Target: last.Deploy,
|
||||||
|
Commit: frontend.Commit{
|
||||||
|
Sha: last.Commit,
|
||||||
|
Ref: last.Ref,
|
||||||
|
Refspec: last.Refspec,
|
||||||
|
Branch: last.Branch,
|
||||||
|
Message: last.Message,
|
||||||
|
Author: frontend.Author{
|
||||||
|
Name: last.Author,
|
||||||
|
Email: last.Email,
|
||||||
|
Avatar: last.Avatar,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Job: frontend.Job{
|
||||||
|
Number: job.Number,
|
||||||
|
Matrix: job.Environment,
|
||||||
|
},
|
||||||
|
Sys: frontend.System{
|
||||||
|
Name: "drone",
|
||||||
|
Link: link,
|
||||||
|
Arch: "linux/amd64",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
221
server/rpc.go
Normal file
221
server/rpc.go
Normal file
|
@ -0,0 +1,221 @@
|
||||||
|
package server
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"strconv"
|
||||||
|
|
||||||
|
"github.com/Sirupsen/logrus"
|
||||||
|
"github.com/cncd/logging"
|
||||||
|
"github.com/cncd/pipeline/pipeline/rpc"
|
||||||
|
"github.com/cncd/pubsub"
|
||||||
|
"github.com/cncd/queue"
|
||||||
|
"github.com/gin-gonic/gin"
|
||||||
|
|
||||||
|
"github.com/drone/drone/model"
|
||||||
|
"github.com/drone/drone/remote"
|
||||||
|
"github.com/drone/drone/store"
|
||||||
|
)
|
||||||
|
|
||||||
|
// This file is a complete disaster because I'm trying to wedge in some
|
||||||
|
// experimental code. Please pardon our appearance during renovations.
|
||||||
|
|
||||||
|
var config = struct {
|
||||||
|
pubsub pubsub.Publisher
|
||||||
|
queue queue.Queue
|
||||||
|
logger logging.Log
|
||||||
|
secret string
|
||||||
|
host string
|
||||||
|
}{
|
||||||
|
pubsub.New(),
|
||||||
|
queue.New(),
|
||||||
|
logging.New(),
|
||||||
|
os.Getenv("DRONE_SECRET"),
|
||||||
|
os.Getenv("DRONE_HOST"),
|
||||||
|
}
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
config.pubsub.Create(context.Background(), "topic/events")
|
||||||
|
}
|
||||||
|
|
||||||
|
// func SetupRPC() gin.HandlerFunc {
|
||||||
|
// return func(c *gin.Context) {
|
||||||
|
// c.Next()
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
func RPCHandler(c *gin.Context) {
|
||||||
|
|
||||||
|
fmt.Println(c.Request.Header.Write(os.Stdout))
|
||||||
|
if secret := c.Request.Header.Get("Authorization"); secret != "Bearer "+config.secret {
|
||||||
|
log.Printf("Unable to connect agent. Invalid authorization token %q does not match %q", secret, config.secret)
|
||||||
|
c.String(401, "Unable to connect agent. Invalid authorization token")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
peer := RPC{
|
||||||
|
remote: remote.FromContext(c),
|
||||||
|
store: store.FromContext(c),
|
||||||
|
queue: config.queue,
|
||||||
|
pubsub: config.pubsub,
|
||||||
|
logger: config.logger,
|
||||||
|
host: config.host,
|
||||||
|
}
|
||||||
|
rpc.NewServer(&peer).ServeHTTP(c.Writer, c.Request)
|
||||||
|
}
|
||||||
|
|
||||||
|
type RPC struct {
|
||||||
|
remote remote.Remote
|
||||||
|
queue queue.Queue
|
||||||
|
pubsub pubsub.Publisher
|
||||||
|
logger logging.Log
|
||||||
|
store store.Store
|
||||||
|
host string
|
||||||
|
}
|
||||||
|
|
||||||
|
// Next implements the rpc.Next function
|
||||||
|
func (s *RPC) Next(c context.Context) (*rpc.Pipeline, error) {
|
||||||
|
filter := func(*queue.Task) bool { return true }
|
||||||
|
task, err := s.queue.Poll(c, filter)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
} else if task == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
pipeline := new(rpc.Pipeline)
|
||||||
|
err = json.Unmarshal(task.Data, pipeline)
|
||||||
|
return pipeline, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Notify implements the rpc.Notify function
|
||||||
|
func (s *RPC) Notify(c context.Context, id string) (bool, error) {
|
||||||
|
err := s.queue.Wait(c, id)
|
||||||
|
return (err == queue.ErrCancel), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Extend implements the rpc.Extend function
|
||||||
|
func (s *RPC) Extend(c context.Context, id string) error {
|
||||||
|
return s.queue.Extend(c, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update implements the rpc.Update function
|
||||||
|
func (s *RPC) Update(c context.Context, id string, state rpc.State) error {
|
||||||
|
jobID, err := strconv.ParseInt(id, 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
job, err := s.store.GetJob(jobID)
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("error: cannot find job with id %d: %s", jobID, err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
build, err := s.store.GetBuild(job.BuildID)
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("error: cannot find build with id %d: %s", job.BuildID, err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
repo, err := s.store.GetRepo(build.RepoID)
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("error: cannot find repo with id %d: %s", build.RepoID, err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if build.Status != model.StatusRunning {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
job.Started = state.Started
|
||||||
|
job.Finished = state.Finished
|
||||||
|
job.ExitCode = state.ExitCode
|
||||||
|
job.Status = model.StatusRunning
|
||||||
|
job.Error = state.Error
|
||||||
|
|
||||||
|
if build.Status == model.StatusPending {
|
||||||
|
build.Started = job.Started
|
||||||
|
build.Status = model.StatusRunning
|
||||||
|
s.store.UpdateBuild(build)
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Printf("pipeline: update %s: exited=%v, exit_code=%d", id, state.Exited, state.ExitCode)
|
||||||
|
|
||||||
|
if state.Exited {
|
||||||
|
|
||||||
|
job.Status = model.StatusSuccess
|
||||||
|
if job.ExitCode != 0 || job.Error != "" {
|
||||||
|
job.Status = model.StatusFailure
|
||||||
|
}
|
||||||
|
|
||||||
|
// save the logs
|
||||||
|
var buf bytes.Buffer
|
||||||
|
if serr := s.logger.Snapshot(context.Background(), id, &buf); serr != nil {
|
||||||
|
log.Printf("error: snapshotting logs: %s", serr)
|
||||||
|
}
|
||||||
|
if werr := s.store.WriteLog(job, &buf); werr != nil {
|
||||||
|
log.Printf("error: persisting logs: %s", werr)
|
||||||
|
}
|
||||||
|
|
||||||
|
// close the logger
|
||||||
|
s.logger.Close(c, id)
|
||||||
|
s.queue.Done(c, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
// hackity hack
|
||||||
|
cc := context.WithValue(c, "store", s.store)
|
||||||
|
ok, uerr := store.UpdateBuildJob(cc, build, job)
|
||||||
|
if uerr != nil {
|
||||||
|
log.Printf("error: updating job: %s", uerr)
|
||||||
|
}
|
||||||
|
if ok {
|
||||||
|
// get the user because we transfer the user form the server to agent
|
||||||
|
// and back we lose the token which does not get serialized to json.
|
||||||
|
user, uerr := s.store.GetUser(repo.UserID)
|
||||||
|
if uerr != nil {
|
||||||
|
logrus.Errorf("Unable to find user. %s", err)
|
||||||
|
} else {
|
||||||
|
s.remote.Status(user, repo, build,
|
||||||
|
fmt.Sprintf("%s/%s/%d", s.host, repo.FullName, build.Number))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
message := pubsub.Message{}
|
||||||
|
message.Data, _ = json.Marshal(model.Event{
|
||||||
|
Type: func() model.EventType {
|
||||||
|
// HACK we don't even really care about the event type.
|
||||||
|
// so we should just simplify how events are triggered.
|
||||||
|
// WTF was this being used for?????????????????????????
|
||||||
|
if job.Status == model.StatusRunning {
|
||||||
|
return model.Started
|
||||||
|
}
|
||||||
|
return model.Finished
|
||||||
|
}(),
|
||||||
|
Repo: *repo,
|
||||||
|
Build: *build,
|
||||||
|
Job: *job,
|
||||||
|
})
|
||||||
|
message.Labels = map[string]string{
|
||||||
|
"repo": repo.FullName,
|
||||||
|
"private": strconv.FormatBool(repo.IsPrivate),
|
||||||
|
}
|
||||||
|
s.pubsub.Publish(c, "topic/events", message)
|
||||||
|
log.Println("finish rpc.update")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Save implements the rpc.Save function
|
||||||
|
func (s *RPC) Save(c context.Context, id, mime string, file io.Reader) error { return nil }
|
||||||
|
|
||||||
|
// Log implements the rpc.Log function
|
||||||
|
func (s *RPC) Log(c context.Context, id string, line *rpc.Line) error {
|
||||||
|
entry := new(logging.Entry)
|
||||||
|
entry.Data, _ = json.Marshal(line)
|
||||||
|
fmt.Println(string(entry.Data))
|
||||||
|
s.logger.Write(c, id, entry)
|
||||||
|
return nil
|
||||||
|
}
|
158
server/stream.go
158
server/stream.go
|
@ -1,10 +1,13 @@
|
||||||
package server
|
package server
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/cncd/logging"
|
||||||
|
"github.com/cncd/pubsub"
|
||||||
"github.com/drone/drone/cache"
|
"github.com/drone/drone/cache"
|
||||||
"github.com/drone/drone/model"
|
"github.com/drone/drone/model"
|
||||||
"github.com/drone/drone/router/middleware/session"
|
"github.com/drone/drone/router/middleware/session"
|
||||||
|
@ -194,3 +197,158 @@ func reader(ws *websocket.Conn) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
// CANARY IMPLEMENTATION
|
||||||
|
//
|
||||||
|
// This file is a complete disaster because I'm trying to wedge in some
|
||||||
|
// experimental code. Please pardon our appearance during renovations.
|
||||||
|
//
|
||||||
|
|
||||||
|
func LogStream2(c *gin.Context) {
|
||||||
|
repo := session.Repo(c)
|
||||||
|
buildn, _ := strconv.Atoi(c.Param("build"))
|
||||||
|
jobn, _ := strconv.Atoi(c.Param("number"))
|
||||||
|
|
||||||
|
build, err := store.GetBuildNumber(c, repo, buildn)
|
||||||
|
if err != nil {
|
||||||
|
logrus.Debugln("stream cannot get build number.", err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
job, err := store.GetJobNumber(c, build, jobn)
|
||||||
|
if err != nil {
|
||||||
|
logrus.Debugln("stream cannot get job number.", err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if job.Status != model.StatusRunning {
|
||||||
|
logrus.Debugln("stream not found.")
|
||||||
|
c.AbortWithStatus(404)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
ws, err := upgrader.Upgrade(c.Writer, c.Request, nil)
|
||||||
|
if err != nil {
|
||||||
|
if _, ok := err.(websocket.HandshakeError); !ok {
|
||||||
|
logrus.Errorf("Cannot upgrade websocket. %s", err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
logrus.Debugf("Successfull upgraded websocket")
|
||||||
|
|
||||||
|
ticker := time.NewTicker(pingPeriod)
|
||||||
|
logc := make(chan []byte, 10)
|
||||||
|
|
||||||
|
ctx, cancel := context.WithCancel(
|
||||||
|
context.Background(),
|
||||||
|
)
|
||||||
|
defer func() {
|
||||||
|
cancel()
|
||||||
|
ticker.Stop()
|
||||||
|
close(logc)
|
||||||
|
logrus.Debugf("Successfully closing websocket")
|
||||||
|
}()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
// TODO remove global variable
|
||||||
|
config.logger.Tail(ctx, fmt.Sprint(job.ID), func(entries ...*logging.Entry) {
|
||||||
|
for _, entry := range entries {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
logc <- entry.Data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
cancel()
|
||||||
|
}()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return
|
||||||
|
case buf, ok := <-logc:
|
||||||
|
if ok {
|
||||||
|
ws.SetWriteDeadline(time.Now().Add(writeWait))
|
||||||
|
ws.WriteMessage(websocket.TextMessage, buf)
|
||||||
|
}
|
||||||
|
case <-ticker.C:
|
||||||
|
err := ws.WriteControl(websocket.PingMessage, []byte{}, time.Now().Add(writeWait))
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
reader(ws)
|
||||||
|
}
|
||||||
|
|
||||||
|
func EventStream2(c *gin.Context) {
|
||||||
|
ws, err := upgrader.Upgrade(c.Writer, c.Request, nil)
|
||||||
|
if err != nil {
|
||||||
|
if _, ok := err.(websocket.HandshakeError); !ok {
|
||||||
|
logrus.Errorf("Cannot upgrade websocket. %s", err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
logrus.Debugf("Successfull upgraded websocket")
|
||||||
|
|
||||||
|
user := session.User(c)
|
||||||
|
repo := map[string]bool{}
|
||||||
|
if user != nil {
|
||||||
|
repo, _ = cache.GetRepoMap(c, user)
|
||||||
|
}
|
||||||
|
|
||||||
|
ticker := time.NewTicker(pingPeriod)
|
||||||
|
eventc := make(chan []byte, 10)
|
||||||
|
|
||||||
|
ctx, cancel := context.WithCancel(
|
||||||
|
context.Background(),
|
||||||
|
)
|
||||||
|
defer func() {
|
||||||
|
cancel()
|
||||||
|
ticker.Stop()
|
||||||
|
close(eventc)
|
||||||
|
logrus.Debugf("Successfully closing websocket")
|
||||||
|
}()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
// TODO remove this from global config
|
||||||
|
config.pubsub.Subscribe(c, "topic/events", func(m pubsub.Message) {
|
||||||
|
name := m.Labels["repo"]
|
||||||
|
priv := m.Labels["private"]
|
||||||
|
if repo[name] || priv == "false" {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
eventc <- m.Data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
cancel()
|
||||||
|
}()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return
|
||||||
|
case buf, ok := <-eventc:
|
||||||
|
if ok {
|
||||||
|
ws.SetWriteDeadline(time.Now().Add(writeWait))
|
||||||
|
ws.WriteMessage(websocket.TextMessage, buf)
|
||||||
|
}
|
||||||
|
case <-ticker.C:
|
||||||
|
err := ws.WriteControl(websocket.PingMessage, []byte{}, time.Now().Add(writeWait))
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
reader(ws)
|
||||||
|
}
|
||||||
|
|
29
vendor/github.com/cncd/logging/LICENSE
generated
vendored
Normal file
29
vendor/github.com/cncd/logging/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
BSD 3-Clause License
|
||||||
|
|
||||||
|
Copyright (c) 2017, Brad Rydzewski
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this
|
||||||
|
list of conditions and the following disclaimer.
|
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
|
||||||
|
* Neither the name of the copyright holder nor the names of its
|
||||||
|
contributors may be used to endorse or promote products derived from
|
||||||
|
this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
6
vendor/github.com/cncd/logging/README
generated
vendored
Normal file
6
vendor/github.com/cncd/logging/README
generated
vendored
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
Go package provides a common interface for storing and streaming logs.
|
||||||
|
|
||||||
|
Documentation:
|
||||||
|
|
||||||
|
http://godoc.org/github.com/cncd/logging
|
||||||
|
http://godoc.org/github.com/cncd/logging/gcp
|
143
vendor/github.com/cncd/logging/log.go
generated
vendored
Normal file
143
vendor/github.com/cncd/logging/log.go
generated
vendored
Normal file
|
@ -0,0 +1,143 @@
|
||||||
|
package logging
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"sync"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TODO (bradrydzewski) writing to subscribers is currently a blocking
|
||||||
|
// operation and does not protect against slow clients from locking
|
||||||
|
// the stream. This should be resolved.
|
||||||
|
|
||||||
|
// TODO (bradrydzewski) implement a mux.Info to fetch information and
|
||||||
|
// statistics for the multiplexier. Streams, subscribers, etc
|
||||||
|
// mux.Info()
|
||||||
|
|
||||||
|
// TODO (bradrydzewski) refactor code to place publisher and subscriber
|
||||||
|
// operations in separate files with more encapsulated logic.
|
||||||
|
// sub.push()
|
||||||
|
// sub.join()
|
||||||
|
// sub.start()... event loop
|
||||||
|
|
||||||
|
type subscriber struct {
|
||||||
|
handler Handler
|
||||||
|
}
|
||||||
|
|
||||||
|
type stream struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
path string
|
||||||
|
hist []*Entry
|
||||||
|
subs map[*subscriber]struct{}
|
||||||
|
done chan struct{}
|
||||||
|
wait sync.WaitGroup
|
||||||
|
}
|
||||||
|
|
||||||
|
type log struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
streams map[string]*stream
|
||||||
|
}
|
||||||
|
|
||||||
|
// New returns a new logger.
|
||||||
|
func New() Log {
|
||||||
|
return &log{
|
||||||
|
streams: map[string]*stream{},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *log) Open(c context.Context, path string) error {
|
||||||
|
l.Lock()
|
||||||
|
_, ok := l.streams[path]
|
||||||
|
if !ok {
|
||||||
|
l.streams[path] = &stream{
|
||||||
|
path: path,
|
||||||
|
subs: make(map[*subscriber]struct{}),
|
||||||
|
done: make(chan struct{}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
l.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *log) Write(c context.Context, path string, entry *Entry) error {
|
||||||
|
l.Lock()
|
||||||
|
s, ok := l.streams[path]
|
||||||
|
l.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
s.Lock()
|
||||||
|
s.hist = append(s.hist, entry)
|
||||||
|
for sub := range s.subs {
|
||||||
|
go sub.handler(entry)
|
||||||
|
}
|
||||||
|
s.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *log) Tail(c context.Context, path string, handler Handler) error {
|
||||||
|
l.Lock()
|
||||||
|
s, ok := l.streams[path]
|
||||||
|
l.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
|
||||||
|
sub := &subscriber{
|
||||||
|
handler: handler,
|
||||||
|
}
|
||||||
|
s.Lock()
|
||||||
|
if len(s.hist) != 0 {
|
||||||
|
sub.handler(s.hist...)
|
||||||
|
}
|
||||||
|
s.subs[sub] = struct{}{}
|
||||||
|
s.Unlock()
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-c.Done():
|
||||||
|
case <-s.done:
|
||||||
|
}
|
||||||
|
|
||||||
|
s.Lock()
|
||||||
|
delete(s.subs, sub)
|
||||||
|
s.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *log) Close(c context.Context, path string) error {
|
||||||
|
l.Lock()
|
||||||
|
s, ok := l.streams[path]
|
||||||
|
l.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
|
||||||
|
s.Lock()
|
||||||
|
close(s.done)
|
||||||
|
s.Unlock()
|
||||||
|
|
||||||
|
l.Lock()
|
||||||
|
delete(l.streams, path)
|
||||||
|
l.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *log) Snapshot(c context.Context, path string, w io.Writer) error {
|
||||||
|
l.Lock()
|
||||||
|
s, ok := l.streams[path]
|
||||||
|
l.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
s.Lock()
|
||||||
|
for _, entry := range s.hist {
|
||||||
|
w.Write(entry.Data)
|
||||||
|
w.Write(cr)
|
||||||
|
}
|
||||||
|
s.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var cr = []byte{'\n'}
|
80
vendor/github.com/cncd/logging/logging.go
generated
vendored
Normal file
80
vendor/github.com/cncd/logging/logging.go
generated
vendored
Normal file
|
@ -0,0 +1,80 @@
|
||||||
|
package logging
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"errors"
|
||||||
|
"io"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ErrNotFound is returned when the log does not exist.
|
||||||
|
var ErrNotFound = errors.New("stream: not found")
|
||||||
|
|
||||||
|
// Entry defines a log entry.
|
||||||
|
type Entry struct {
|
||||||
|
// ID identifies this message.
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
|
||||||
|
// Data is the actual data in the entry.
|
||||||
|
Data []byte `json:"data"`
|
||||||
|
|
||||||
|
// Tags represents the key-value pairs the
|
||||||
|
// entry is tagged with.
|
||||||
|
Tags map[string]string `json:"tags,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handler defines a callback function for handling log entries.
|
||||||
|
type Handler func(...*Entry)
|
||||||
|
|
||||||
|
// Log defines a log multiplexer.
|
||||||
|
type Log interface {
|
||||||
|
// Open opens the log.
|
||||||
|
Open(c context.Context, path string) error
|
||||||
|
|
||||||
|
// Write writes the entry to the log.
|
||||||
|
Write(c context.Context, path string, entry *Entry) error
|
||||||
|
|
||||||
|
// Tail tails the log.
|
||||||
|
Tail(c context.Context, path string, handler Handler) error
|
||||||
|
|
||||||
|
// Close closes the log.
|
||||||
|
Close(c context.Context, path string) error
|
||||||
|
|
||||||
|
// Snapshot snapshots the stream to Writer w.
|
||||||
|
Snapshot(c context.Context, path string, w io.Writer) error
|
||||||
|
|
||||||
|
// Info returns runtime information about the multiplexer.
|
||||||
|
// Info(c context.Context) (interface{}, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// // global streamer
|
||||||
|
// var global = New()
|
||||||
|
//
|
||||||
|
// // Set sets a default global logger.
|
||||||
|
// func Set(log Log) {
|
||||||
|
// global = log
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Open opens the log stream.
|
||||||
|
// func Open(c context.Context, path string) error {
|
||||||
|
// return global.Open(c, path)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Write writes the log entry to the stream.
|
||||||
|
// func Write(c context.Context, path string, entry *Entry) error {
|
||||||
|
// return global.Write(c, path, entry)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Tail tails the log stream.
|
||||||
|
// func Tail(c context.Context, path string, handler Handler) error {
|
||||||
|
// return global.Tail(c, path, handler)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Close closes the log stream.
|
||||||
|
// func Close(c context.Context, path string) error {
|
||||||
|
// return global.Close(c, path)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Snapshot snapshots the stream to Writer w.
|
||||||
|
// func Snapshot(c context.Context, path string, w io.Writer) error {
|
||||||
|
// return global.Snapshot(c, path, w)
|
||||||
|
// }
|
29
vendor/github.com/cncd/pipeline/LICENSE
generated
vendored
Normal file
29
vendor/github.com/cncd/pipeline/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
BSD 3-Clause License
|
||||||
|
|
||||||
|
Copyright (c) 2017, Brad Rydzewski
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this
|
||||||
|
list of conditions and the following disclaimer.
|
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
|
||||||
|
* Neither the name of the copyright holder nor the names of its
|
||||||
|
contributors may be used to endorse or promote products derived from
|
||||||
|
this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
21
vendor/github.com/cncd/pipeline/pipeline/backend/backend.go
generated
vendored
Normal file
21
vendor/github.com/cncd/pipeline/pipeline/backend/backend.go
generated
vendored
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
package backend
|
||||||
|
|
||||||
|
import "io"
|
||||||
|
|
||||||
|
// Engine defines a container orchestration backend and is used
|
||||||
|
// to create and manage container resources.
|
||||||
|
type Engine interface {
|
||||||
|
// Setup the pipeline environment.
|
||||||
|
Setup(*Config) error
|
||||||
|
// Start the pipeline step.
|
||||||
|
Exec(*Step) error
|
||||||
|
// Kill the pipeline step.
|
||||||
|
Kill(*Step) error
|
||||||
|
// Wait for the pipeline step to complete and returns
|
||||||
|
// the completion results.
|
||||||
|
Wait(*Step) (*State, error)
|
||||||
|
// Tail the pipeline step logs.
|
||||||
|
Tail(*Step) (io.ReadCloser, error)
|
||||||
|
// Destroy the pipeline environment.
|
||||||
|
Destroy(*Config) error
|
||||||
|
}
|
130
vendor/github.com/cncd/pipeline/pipeline/backend/docker/convert.go
generated
vendored
Normal file
130
vendor/github.com/cncd/pipeline/pipeline/backend/docker/convert.go
generated
vendored
Normal file
|
@ -0,0 +1,130 @@
|
||||||
|
package docker
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/base64"
|
||||||
|
"encoding/json"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
|
||||||
|
"github.com/docker/docker/api/types/container"
|
||||||
|
)
|
||||||
|
|
||||||
|
// returns a container configuration.
|
||||||
|
func toConfig(proc *backend.Step) *container.Config {
|
||||||
|
config := &container.Config{
|
||||||
|
Image: proc.Image,
|
||||||
|
Labels: proc.Labels,
|
||||||
|
WorkingDir: proc.WorkingDir,
|
||||||
|
AttachStdout: true,
|
||||||
|
AttachStderr: true,
|
||||||
|
}
|
||||||
|
if len(proc.Environment) != 0 {
|
||||||
|
config.Env = toEnv(proc.Environment)
|
||||||
|
}
|
||||||
|
if len(proc.Command) != 0 {
|
||||||
|
config.Cmd = proc.Command
|
||||||
|
}
|
||||||
|
if len(proc.Entrypoint) != 0 {
|
||||||
|
config.Entrypoint = proc.Entrypoint
|
||||||
|
}
|
||||||
|
if len(proc.Volumes) != 0 {
|
||||||
|
config.Volumes = toVol(proc.Volumes)
|
||||||
|
}
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
// returns a container host configuration.
|
||||||
|
func toHostConfig(proc *backend.Step) *container.HostConfig {
|
||||||
|
config := &container.HostConfig{
|
||||||
|
Resources: container.Resources{
|
||||||
|
CPUQuota: proc.CPUQuota,
|
||||||
|
CPUShares: proc.CPUShares,
|
||||||
|
CpusetCpus: proc.CPUSet,
|
||||||
|
Memory: proc.MemLimit,
|
||||||
|
MemorySwap: proc.MemSwapLimit,
|
||||||
|
},
|
||||||
|
Privileged: proc.Privileged,
|
||||||
|
ShmSize: proc.ShmSize,
|
||||||
|
}
|
||||||
|
// if len(proc.VolumesFrom) != 0 {
|
||||||
|
// config.VolumesFrom = proc.VolumesFrom
|
||||||
|
// }
|
||||||
|
// if len(proc.Network) != 0 {
|
||||||
|
// config.NetworkMode = container.NetworkMode(
|
||||||
|
// proc.Network,
|
||||||
|
// )
|
||||||
|
// }
|
||||||
|
if len(proc.DNS) != 0 {
|
||||||
|
config.DNS = proc.DNS
|
||||||
|
}
|
||||||
|
if len(proc.DNSSearch) != 0 {
|
||||||
|
config.DNSSearch = proc.DNSSearch
|
||||||
|
}
|
||||||
|
if len(proc.ExtraHosts) != 0 {
|
||||||
|
config.ExtraHosts = proc.ExtraHosts
|
||||||
|
}
|
||||||
|
if len(proc.Devices) != 0 {
|
||||||
|
config.Devices = toDev(proc.Devices)
|
||||||
|
}
|
||||||
|
if len(proc.Volumes) != 0 {
|
||||||
|
config.Binds = proc.Volumes
|
||||||
|
}
|
||||||
|
// if proc.OomKillDisable {
|
||||||
|
// config.OomKillDisable = &proc.OomKillDisable
|
||||||
|
// }
|
||||||
|
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
// helper function that converts a slice of volume paths to a set of
|
||||||
|
// unique volume names.
|
||||||
|
func toVol(paths []string) map[string]struct{} {
|
||||||
|
set := map[string]struct{}{}
|
||||||
|
for _, path := range paths {
|
||||||
|
parts := strings.Split(path, ":")
|
||||||
|
if len(parts) < 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
set[parts[1]] = struct{}{}
|
||||||
|
}
|
||||||
|
return set
|
||||||
|
}
|
||||||
|
|
||||||
|
// helper function that converts a key value map of environment variables to a
|
||||||
|
// string slice in key=value format.
|
||||||
|
func toEnv(env map[string]string) []string {
|
||||||
|
var envs []string
|
||||||
|
for k, v := range env {
|
||||||
|
envs = append(envs, k+"="+v)
|
||||||
|
}
|
||||||
|
return envs
|
||||||
|
}
|
||||||
|
|
||||||
|
// helper function that converts a slice of device paths to a slice of
|
||||||
|
// container.DeviceMapping.
|
||||||
|
func toDev(paths []string) []container.DeviceMapping {
|
||||||
|
var devices []container.DeviceMapping
|
||||||
|
for _, path := range paths {
|
||||||
|
parts := strings.Split(path, ":")
|
||||||
|
if len(parts) < 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
devices = append(devices, container.DeviceMapping{
|
||||||
|
PathOnHost: parts[0],
|
||||||
|
PathInContainer: parts[1],
|
||||||
|
CgroupPermissions: "rwm",
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return devices
|
||||||
|
}
|
||||||
|
|
||||||
|
// helper function that serializes the auth configuration as JSON
|
||||||
|
// base64 payload.
|
||||||
|
func encodeAuthToBase64(authConfig backend.Auth) (string, error) {
|
||||||
|
buf, err := json.Marshal(authConfig)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return base64.URLEncoding.EncodeToString(buf), nil
|
||||||
|
}
|
202
vendor/github.com/cncd/pipeline/pipeline/backend/docker/docker.go
generated
vendored
Normal file
202
vendor/github.com/cncd/pipeline/pipeline/backend/docker/docker.go
generated
vendored
Normal file
|
@ -0,0 +1,202 @@
|
||||||
|
package docker
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
|
||||||
|
"github.com/docker/docker/api/types"
|
||||||
|
"github.com/docker/docker/api/types/network"
|
||||||
|
"github.com/docker/docker/api/types/volume"
|
||||||
|
"github.com/docker/docker/client"
|
||||||
|
"github.com/docker/docker/pkg/stdcopy"
|
||||||
|
)
|
||||||
|
|
||||||
|
type engine struct {
|
||||||
|
client client.APIClient
|
||||||
|
}
|
||||||
|
|
||||||
|
// New returns a new Docker Engine using the given client.
|
||||||
|
func New(cli client.APIClient) backend.Engine {
|
||||||
|
return &engine{
|
||||||
|
client: cli,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewEnv returns a new Docker Engine using the client connection
|
||||||
|
// environment variables.
|
||||||
|
func NewEnv() (backend.Engine, error) {
|
||||||
|
cli, err := client.NewEnvClient()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return New(cli), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Setup(conf *backend.Config) error {
|
||||||
|
for _, vol := range conf.Volumes {
|
||||||
|
_, err := e.client.VolumeCreate(noContext, volume.VolumesCreateBody{
|
||||||
|
Name: vol.Name,
|
||||||
|
Driver: vol.Driver,
|
||||||
|
DriverOpts: vol.DriverOpts,
|
||||||
|
// Labels: defaultLabels,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for _, network := range conf.Networks {
|
||||||
|
_, err := e.client.NetworkCreate(noContext, network.Name, types.NetworkCreate{
|
||||||
|
Driver: network.Driver,
|
||||||
|
Options: network.DriverOpts,
|
||||||
|
// Labels: defaultLabels,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Exec(proc *backend.Step) error {
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
config := toConfig(proc)
|
||||||
|
hostConfig := toHostConfig(proc)
|
||||||
|
|
||||||
|
// create pull options with encoded authorization credentials.
|
||||||
|
pullopts := types.ImagePullOptions{}
|
||||||
|
if proc.AuthConfig.Username != "" && proc.AuthConfig.Password != "" {
|
||||||
|
pullopts.RegistryAuth, _ = encodeAuthToBase64(proc.AuthConfig)
|
||||||
|
}
|
||||||
|
|
||||||
|
// automatically pull the latest version of the image if requested
|
||||||
|
// by the process configuration.
|
||||||
|
if proc.Pull {
|
||||||
|
rc, perr := e.client.ImagePull(ctx, config.Image, pullopts)
|
||||||
|
if perr == nil {
|
||||||
|
io.Copy(ioutil.Discard, rc)
|
||||||
|
rc.Close()
|
||||||
|
}
|
||||||
|
// fix for drone/drone#1917
|
||||||
|
if perr != nil && proc.AuthConfig.Password != "" {
|
||||||
|
return perr
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := e.client.ContainerCreate(ctx, config, hostConfig, nil, proc.Name)
|
||||||
|
if client.IsErrImageNotFound(err) {
|
||||||
|
// automatically pull and try to re-create the image if the
|
||||||
|
// failure is caused because the image does not exist.
|
||||||
|
rc, perr := e.client.ImagePull(ctx, config.Image, pullopts)
|
||||||
|
if perr != nil {
|
||||||
|
return perr
|
||||||
|
}
|
||||||
|
io.Copy(ioutil.Discard, rc)
|
||||||
|
rc.Close()
|
||||||
|
|
||||||
|
_, err = e.client.ContainerCreate(ctx, config, hostConfig, nil, proc.Name)
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, net := range proc.Networks {
|
||||||
|
err = e.client.NetworkConnect(ctx, net.Name, proc.Name, &network.EndpointSettings{
|
||||||
|
Aliases: net.Aliases,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// if proc.Network != "host" { // or bridge, overlay, none, internal, container:<name> ....
|
||||||
|
// err = e.client.NetworkConnect(ctx, proc.Network, proc.Name, &network.EndpointSettings{
|
||||||
|
// Aliases: proc.NetworkAliases,
|
||||||
|
// })
|
||||||
|
// if err != nil {
|
||||||
|
// return err
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
return e.client.ContainerStart(ctx, proc.Name, startOpts)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Kill(proc *backend.Step) error {
|
||||||
|
return e.client.ContainerKill(noContext, proc.Name, "9")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Wait(proc *backend.Step) (*backend.State, error) {
|
||||||
|
_, err := e.client.ContainerWait(noContext, proc.Name)
|
||||||
|
if err != nil {
|
||||||
|
// todo
|
||||||
|
}
|
||||||
|
|
||||||
|
info, err := e.client.ContainerInspect(noContext, proc.Name)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if info.State.Running {
|
||||||
|
// todo
|
||||||
|
}
|
||||||
|
|
||||||
|
return &backend.State{
|
||||||
|
Exited: true,
|
||||||
|
ExitCode: info.State.ExitCode,
|
||||||
|
OOMKilled: info.State.OOMKilled,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Tail(proc *backend.Step) (io.ReadCloser, error) {
|
||||||
|
logs, err := e.client.ContainerLogs(noContext, proc.Name, logsOpts)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
rc, wc := io.Pipe()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
stdcopy.StdCopy(wc, wc, logs)
|
||||||
|
logs.Close()
|
||||||
|
wc.Close()
|
||||||
|
rc.Close()
|
||||||
|
}()
|
||||||
|
return rc, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *engine) Destroy(conf *backend.Config) error {
|
||||||
|
for _, stage := range conf.Stages {
|
||||||
|
for _, step := range stage.Steps {
|
||||||
|
e.client.ContainerKill(noContext, step.Name, "9")
|
||||||
|
e.client.ContainerRemove(noContext, step.Name, removeOpts)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for _, volume := range conf.Volumes {
|
||||||
|
e.client.VolumeRemove(noContext, volume.Name, true)
|
||||||
|
}
|
||||||
|
for _, network := range conf.Networks {
|
||||||
|
e.client.NetworkRemove(noContext, network.Name)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
noContext = context.Background()
|
||||||
|
|
||||||
|
startOpts = types.ContainerStartOptions{}
|
||||||
|
|
||||||
|
removeOpts = types.ContainerRemoveOptions{
|
||||||
|
RemoveVolumes: true,
|
||||||
|
RemoveLinks: false,
|
||||||
|
Force: false,
|
||||||
|
}
|
||||||
|
|
||||||
|
logsOpts = types.ContainerLogsOptions{
|
||||||
|
Follow: true,
|
||||||
|
ShowStdout: true,
|
||||||
|
ShowStderr: true,
|
||||||
|
Details: false,
|
||||||
|
Timestamps: false,
|
||||||
|
}
|
||||||
|
)
|
44
vendor/github.com/cncd/pipeline/pipeline/backend/docker/pool.go
generated
vendored
Normal file
44
vendor/github.com/cncd/pipeline/pipeline/backend/docker/pool.go
generated
vendored
Normal file
|
@ -0,0 +1,44 @@
|
||||||
|
package docker
|
||||||
|
|
||||||
|
// import (
|
||||||
|
// "context"
|
||||||
|
//
|
||||||
|
// "github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
// )
|
||||||
|
//
|
||||||
|
// // Pool manages a pool of Docker clients.
|
||||||
|
// type Pool struct {
|
||||||
|
// queue chan (backend.Engine)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // NewPool returns a Pool.
|
||||||
|
// func NewPool(engines ...backend.Engine) *Pool {
|
||||||
|
// return &Pool{
|
||||||
|
// queue: make(chan backend.Engine, len(engines)),
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Reserve requests the next available Docker client in the pool.
|
||||||
|
// func (p *Pool) Reserve(c context.Context) backend.Engine {
|
||||||
|
// select {
|
||||||
|
// case <-c.Done():
|
||||||
|
// case engine := <-p.queue:
|
||||||
|
// return engine
|
||||||
|
// }
|
||||||
|
// return nil
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Release releases the Docker client back to the pool.
|
||||||
|
// func (p *Pool) Release(engine backend.Engine) {
|
||||||
|
// p.queue <- engine
|
||||||
|
// }
|
||||||
|
|
||||||
|
// pool := docker.Pool(
|
||||||
|
// docker.FromEnvironmentMust(),
|
||||||
|
// docker.FromEnvironmentMust(),
|
||||||
|
// docker.FromEnvironmentMust(),
|
||||||
|
// docker.FromEnvironmentMust(),
|
||||||
|
// )
|
||||||
|
//
|
||||||
|
// client := pool.Reserve()
|
||||||
|
// defer pool.Release(client)
|
103
vendor/github.com/cncd/pipeline/pipeline/backend/types.go
generated
vendored
Normal file
103
vendor/github.com/cncd/pipeline/pipeline/backend/types.go
generated
vendored
Normal file
|
@ -0,0 +1,103 @@
|
||||||
|
package backend
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Config defines the runtime configuration of a pipeline.
|
||||||
|
Config struct {
|
||||||
|
Stages []*Stage `json:"pipeline"` // pipeline stages
|
||||||
|
Networks []*Network `json:"networks"` // network definitions
|
||||||
|
Volumes []*Volume `json:"volumes"` // volume definitions
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stage denotes a collection of one or more steps.
|
||||||
|
Stage struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Alias string `json:"alias,omitempty"`
|
||||||
|
Steps []*Step `json:"steps,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Step defines a container process.
|
||||||
|
Step struct {
|
||||||
|
Name string `json:"name"`
|
||||||
|
Alias string `json:"alias,omitempty"`
|
||||||
|
Image string `json:"image,omitempty"`
|
||||||
|
Pull bool `json:"pull,omitempty"`
|
||||||
|
Detached bool `json:"detach,omitempty"`
|
||||||
|
Privileged bool `json:"privileged,omitempty"`
|
||||||
|
WorkingDir string `json:"working_dir,omitempty"`
|
||||||
|
Environment map[string]string `json:"environment,omitempty"`
|
||||||
|
Labels map[string]string `json:"labels,omitempty"`
|
||||||
|
Entrypoint []string `json:"entrypoint,omitempty"`
|
||||||
|
Command []string `json:"command,omitempty"`
|
||||||
|
ExtraHosts []string `json:"extra_hosts,omitempty"`
|
||||||
|
Volumes []string `json:"volumes,omitempty"`
|
||||||
|
Devices []string `json:"devices,omitempty"`
|
||||||
|
Networks []Conn `json:"networks,omitempty"`
|
||||||
|
DNS []string `json:"dns,omitempty"`
|
||||||
|
DNSSearch []string `json:"dns_search,omitempty"`
|
||||||
|
MemSwapLimit int64 `json:"memswap_limit,omitempty"`
|
||||||
|
MemLimit int64 `json:"mem_limit,omitempty"`
|
||||||
|
ShmSize int64 `json:"shm_size,omitempty"`
|
||||||
|
CPUQuota int64 `json:"cpu_quota,omitempty"`
|
||||||
|
CPUShares int64 `json:"cpu_shares,omitempty"`
|
||||||
|
CPUSet string `json:"cpu_set,omitempty"`
|
||||||
|
OnFailure bool `json:"on_failure,omitempty"`
|
||||||
|
OnSuccess bool `json:"on_success,omitempty"`
|
||||||
|
AuthConfig Auth `json:"auth_config,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Auth defines registry authentication credentials.
|
||||||
|
Auth struct {
|
||||||
|
Username string `json:"username,omitempty"`
|
||||||
|
Password string `json:"password,omitempty"`
|
||||||
|
Email string `json:"email,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Conn defines a container network connection.
|
||||||
|
Conn struct {
|
||||||
|
Name string `json:"name"`
|
||||||
|
Aliases []string `json:"aliases"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Network defines a container network.
|
||||||
|
Network struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Driver string `json:"driver,omitempty"`
|
||||||
|
DriverOpts map[string]string `json:"driver_opts,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Volume defines a container volume.
|
||||||
|
Volume struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Driver string `json:"driver,omitempty"`
|
||||||
|
DriverOpts map[string]string `json:"driver_opts,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// State defines a container state.
|
||||||
|
State struct {
|
||||||
|
// Container exit code
|
||||||
|
ExitCode int `json:"exit_code"`
|
||||||
|
// Container exited, true or false
|
||||||
|
Exited bool `json:"exited"`
|
||||||
|
// Container is oom killed, true or false
|
||||||
|
OOMKilled bool `json:"oom_killed"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// // State defines the pipeline and process state.
|
||||||
|
// State struct {
|
||||||
|
// Pipeline struct {
|
||||||
|
// // Current pipeline step
|
||||||
|
// Step *Step `json:"step"`
|
||||||
|
// // Current pipeline error state
|
||||||
|
// Error error `json:"error"`
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// Process struct {
|
||||||
|
// // Container exit code
|
||||||
|
// ExitCode int `json:"exit_code"`
|
||||||
|
// // Container exited, true or false
|
||||||
|
// Exited bool `json:"exited"`
|
||||||
|
// // Container is oom killed, true or false
|
||||||
|
// OOMKilled bool `json:"oom_killed"`
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
)
|
38
vendor/github.com/cncd/pipeline/pipeline/error.go
generated
vendored
Normal file
38
vendor/github.com/cncd/pipeline/pipeline/error.go
generated
vendored
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrSkip is used as a return value when container execution should be
|
||||||
|
// skipped at runtime. It is not returned as an error by any function.
|
||||||
|
ErrSkip = errors.New("Skipped")
|
||||||
|
|
||||||
|
// ErrCancel is used as a return value when the container execution receives
|
||||||
|
// a cancellation signal from the context.
|
||||||
|
ErrCancel = errors.New("Cancelled")
|
||||||
|
)
|
||||||
|
|
||||||
|
// An ExitError reports an unsuccessful exit.
|
||||||
|
type ExitError struct {
|
||||||
|
Name string
|
||||||
|
Code int
|
||||||
|
}
|
||||||
|
|
||||||
|
// Error returns the error message in string format.
|
||||||
|
func (e *ExitError) Error() string {
|
||||||
|
return fmt.Sprintf("%s : exit code %d", e.Name, e.Code)
|
||||||
|
}
|
||||||
|
|
||||||
|
// An OomError reports the process received an OOMKill from the kernel.
|
||||||
|
type OomError struct {
|
||||||
|
Name string
|
||||||
|
Code int
|
||||||
|
}
|
||||||
|
|
||||||
|
// Error reteurns the error message in string format.
|
||||||
|
func (e *OomError) Error() string {
|
||||||
|
return fmt.Sprintf("%s : received oom kill", e.Name)
|
||||||
|
}
|
121
vendor/github.com/cncd/pipeline/pipeline/frontend/metadata.go
generated
vendored
Normal file
121
vendor/github.com/cncd/pipeline/pipeline/frontend/metadata.go
generated
vendored
Normal file
|
@ -0,0 +1,121 @@
|
||||||
|
package frontend
|
||||||
|
|
||||||
|
import "strconv"
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Metadata defines runtime m.
|
||||||
|
Metadata struct {
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
Repo Repo `json:"repo,omitempty"`
|
||||||
|
Curr Build `json:"curr,omitempty"`
|
||||||
|
Prev Build `json:"prev,omitempty"`
|
||||||
|
Job Job `json:"job,omitempty"`
|
||||||
|
Sys System `json:"sys,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Repo defines runtime metadata for a repository.
|
||||||
|
Repo struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Link string `json:"link,omitempty"`
|
||||||
|
Remote string `json:"remote,omitempty"`
|
||||||
|
Private bool `json:"private,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Build defines runtime metadata for a build.
|
||||||
|
Build struct {
|
||||||
|
Number int `json:"number,omitempty"`
|
||||||
|
Created int64 `json:"created,omitempty"`
|
||||||
|
Started int64 `json:"started,omitempty"`
|
||||||
|
Finished int64 `json:"finished,omitempty"`
|
||||||
|
Timeout int64 `json:"timeout,omitempty"`
|
||||||
|
Status string `json:"status,omitempty"`
|
||||||
|
Event string `json:"event,omitempty"`
|
||||||
|
Link string `json:"link,omitempty"`
|
||||||
|
Target string `json:"target,omitempty"`
|
||||||
|
Trusted bool `json:"trusted,omitempty"`
|
||||||
|
Commit Commit `json:"commit,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Commit defines runtime metadata for a commit.
|
||||||
|
Commit struct {
|
||||||
|
Sha string `json:"sha,omitempty"`
|
||||||
|
Ref string `json:"ref,omitempty"`
|
||||||
|
Refspec string `json:"refspec,omitempty"`
|
||||||
|
Branch string `json:"branch,omitempty"`
|
||||||
|
Message string `json:"message,omitempty"`
|
||||||
|
Author Author `json:"author,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Author defines runtime metadata for a commit author.
|
||||||
|
Author struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Email string `json:"email,omitempty"`
|
||||||
|
Avatar string `json:"avatar,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Job defines runtime metadata for a job.
|
||||||
|
Job struct {
|
||||||
|
Number int `json:"number,omitempty"`
|
||||||
|
Matrix map[string]string `json:"matrix,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// System defines runtime metadata for a ci/cd system.
|
||||||
|
System struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Host string `json:"host,omitempty"`
|
||||||
|
Link string `json:"link,omitempty"`
|
||||||
|
Arch string `json:"arch,omitempty"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Environ returns the metadata as a map of environment variables.
|
||||||
|
func (m *Metadata) Environ() map[string]string {
|
||||||
|
return map[string]string{
|
||||||
|
"CI_REPO": m.Repo.Name,
|
||||||
|
"CI_REPO_NAME": m.Repo.Name,
|
||||||
|
"CI_REPO_LINK": m.Repo.Link,
|
||||||
|
"CI_REPO_REMOTE": m.Repo.Remote,
|
||||||
|
"CI_REMOTE_URL": m.Repo.Remote,
|
||||||
|
"CI_REPO_PRIVATE": strconv.FormatBool(m.Repo.Private),
|
||||||
|
"CI_BUILD_NUMBER": strconv.Itoa(m.Curr.Number),
|
||||||
|
"CI_BUILD_CREATED": strconv.FormatInt(m.Curr.Created, 10),
|
||||||
|
"CI_BUILD_STARTED": strconv.FormatInt(m.Curr.Started, 10),
|
||||||
|
"CI_BUILD_FINISHED": strconv.FormatInt(m.Curr.Finished, 10),
|
||||||
|
"CI_BUILD_STATUS": m.Curr.Status,
|
||||||
|
"CI_BUILD_EVENT": m.Curr.Event,
|
||||||
|
"CI_BUILD_LINK": m.Curr.Link,
|
||||||
|
"CI_BUILD_TARGET": m.Curr.Target,
|
||||||
|
"CI_COMMIT_SHA": m.Curr.Commit.Sha,
|
||||||
|
"CI_COMMIT_REF": m.Curr.Commit.Ref,
|
||||||
|
"CI_COMMIT_REFSPEC": m.Curr.Commit.Refspec,
|
||||||
|
"CI_COMMIT_BRANCH": m.Curr.Commit.Branch,
|
||||||
|
"CI_COMMIT_MESSAGE": m.Curr.Commit.Message,
|
||||||
|
"CI_COMMIT_AUTHOR": m.Curr.Commit.Author.Name,
|
||||||
|
"CI_COMMIT_AUTHOR_NAME": m.Curr.Commit.Author.Name,
|
||||||
|
"CI_COMMIT_AUTHOR_EMAIL": m.Curr.Commit.Author.Email,
|
||||||
|
"CI_COMMIT_AUTHOR_AVATAR": m.Curr.Commit.Author.Avatar,
|
||||||
|
"CI_PREV_BUILD_NUMBER": strconv.Itoa(m.Prev.Number),
|
||||||
|
"CI_PREV_BUILD_CREATED": strconv.FormatInt(m.Prev.Created, 10),
|
||||||
|
"CI_PREV_BUILD_STARTED": strconv.FormatInt(m.Prev.Started, 10),
|
||||||
|
"CI_PREV_BUILD_FINISHED": strconv.FormatInt(m.Prev.Finished, 10),
|
||||||
|
"CI_PREV_BUILD_STATUS": m.Prev.Status,
|
||||||
|
"CI_PREV_BUILD_EVENT": m.Prev.Event,
|
||||||
|
"CI_PREV_BUILD_LINK": m.Prev.Link,
|
||||||
|
"CI_PREV_COMMIT_SHA": m.Prev.Commit.Sha,
|
||||||
|
"CI_PREV_COMMIT_REF": m.Prev.Commit.Ref,
|
||||||
|
"CI_PREV_COMMIT_REFSPEC": m.Prev.Commit.Refspec,
|
||||||
|
"CI_PREV_COMMIT_BRANCH": m.Prev.Commit.Branch,
|
||||||
|
"CI_PREV_COMMIT_MESSAGE": m.Prev.Commit.Message,
|
||||||
|
"CI_PREV_COMMIT_AUTHOR": m.Prev.Commit.Author.Name,
|
||||||
|
"CI_PREV_COMMIT_AUTHOR_NAME": m.Prev.Commit.Author.Name,
|
||||||
|
"CI_PREV_COMMIT_AUTHOR_EMAIL": m.Prev.Commit.Author.Email,
|
||||||
|
"CI_PREV_COMMIT_AUTHOR_AVATAR": m.Prev.Commit.Author.Avatar,
|
||||||
|
"CI_JOB_NUMBER": strconv.Itoa(m.Job.Number),
|
||||||
|
"CI_SYSTEM": m.Sys.Name,
|
||||||
|
"CI_SYSTEM_NAME": m.Sys.Name,
|
||||||
|
"CI_SYSTEM_LINK": m.Sys.Link,
|
||||||
|
"CI_SYSTEM_HOST": m.Sys.Host,
|
||||||
|
"CI_SYSTEM_ARCH": m.Sys.Arch,
|
||||||
|
"CI": m.Sys.Name,
|
||||||
|
}
|
||||||
|
}
|
167
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/compiler.go
generated
vendored
Normal file
167
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/compiler.go
generated
vendored
Normal file
|
@ -0,0 +1,167 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml"
|
||||||
|
// libcompose "github.com/docker/libcompose/yaml"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TODO(bradrydzewski) compiler should handle user-defined volumes from YAML
|
||||||
|
// TODO(bradrydzewski) compiler should handle user-defined networks from YAML
|
||||||
|
|
||||||
|
// Compiler compiles the yaml
|
||||||
|
type Compiler struct {
|
||||||
|
local bool
|
||||||
|
escalated []string
|
||||||
|
prefix string
|
||||||
|
volumes []string
|
||||||
|
env map[string]string
|
||||||
|
base string
|
||||||
|
path string
|
||||||
|
metadata frontend.Metadata
|
||||||
|
aliases []string
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates a new Compiler with options.
|
||||||
|
func New(opts ...Option) *Compiler {
|
||||||
|
compiler := new(Compiler)
|
||||||
|
compiler.env = map[string]string{}
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(compiler)
|
||||||
|
}
|
||||||
|
return compiler
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compile compiles the YAML configuration to the pipeline intermediate
|
||||||
|
// representation configuration format.
|
||||||
|
func (c *Compiler) Compile(conf *yaml.Config) *backend.Config {
|
||||||
|
config := new(backend.Config)
|
||||||
|
|
||||||
|
// create a default volume
|
||||||
|
config.Volumes = append(config.Volumes, &backend.Volume{
|
||||||
|
Name: fmt.Sprintf("%s_default", c.prefix),
|
||||||
|
Driver: "local",
|
||||||
|
})
|
||||||
|
|
||||||
|
// create a default network
|
||||||
|
config.Networks = append(config.Networks, &backend.Network{
|
||||||
|
Name: fmt.Sprintf("%s_default", c.prefix),
|
||||||
|
Driver: "bridge",
|
||||||
|
})
|
||||||
|
|
||||||
|
// overrides the default workspace paths when specified
|
||||||
|
// in the YAML file.
|
||||||
|
if len(conf.Workspace.Base) != 0 {
|
||||||
|
c.base = conf.Workspace.Base
|
||||||
|
}
|
||||||
|
if len(conf.Workspace.Path) != 0 {
|
||||||
|
c.path = conf.Workspace.Path
|
||||||
|
}
|
||||||
|
|
||||||
|
// add default clone step
|
||||||
|
if c.local == false && len(conf.Clone.Containers) == 0 {
|
||||||
|
container := &yaml.Container{
|
||||||
|
Image: "plugins/git:latest",
|
||||||
|
Vargs: map[string]interface{}{"depth": "0"},
|
||||||
|
}
|
||||||
|
name := fmt.Sprintf("%s_clone", c.prefix)
|
||||||
|
step := c.createProcess(name, container)
|
||||||
|
|
||||||
|
stage := new(backend.Stage)
|
||||||
|
stage.Name = name
|
||||||
|
stage.Alias = "clone"
|
||||||
|
stage.Steps = append(stage.Steps, step)
|
||||||
|
|
||||||
|
config.Stages = append(config.Stages, stage)
|
||||||
|
} else if c.local == false {
|
||||||
|
for i, container := range conf.Clone.Containers {
|
||||||
|
if !container.Constraints.Match(c.metadata) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
stage := new(backend.Stage)
|
||||||
|
stage.Name = fmt.Sprintf("%s_clone_%v", c.prefix, i)
|
||||||
|
stage.Alias = container.Name
|
||||||
|
|
||||||
|
name := fmt.Sprintf("%s_clone_%d", c.prefix, i)
|
||||||
|
step := c.createProcess(name, container)
|
||||||
|
stage.Steps = append(stage.Steps, step)
|
||||||
|
|
||||||
|
config.Stages = append(config.Stages, stage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// add services steps
|
||||||
|
if len(conf.Services.Containers) != 0 {
|
||||||
|
stage := new(backend.Stage)
|
||||||
|
stage.Name = fmt.Sprintf("%s_services", c.prefix)
|
||||||
|
stage.Alias = "services"
|
||||||
|
|
||||||
|
for _, container := range conf.Services.Containers {
|
||||||
|
c.aliases = append(c.aliases, container.Name)
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, container := range conf.Services.Containers {
|
||||||
|
name := fmt.Sprintf("%s_services_%d", c.prefix, i)
|
||||||
|
step := c.createProcess(name, container)
|
||||||
|
stage.Steps = append(stage.Steps, step)
|
||||||
|
|
||||||
|
}
|
||||||
|
config.Stages = append(config.Stages, stage)
|
||||||
|
}
|
||||||
|
|
||||||
|
// add pipeline steps. 1 pipeline step per stage, at the moment
|
||||||
|
var stage *backend.Stage
|
||||||
|
var group string
|
||||||
|
for i, container := range conf.Pipeline.Containers {
|
||||||
|
if !container.Constraints.Match(c.metadata) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if stage == nil || group != container.Group || container.Group == "" {
|
||||||
|
group = container.Group
|
||||||
|
|
||||||
|
stage = new(backend.Stage)
|
||||||
|
stage.Name = fmt.Sprintf("%s_stage_%v", c.prefix, i)
|
||||||
|
stage.Alias = container.Name
|
||||||
|
config.Stages = append(config.Stages, stage)
|
||||||
|
}
|
||||||
|
|
||||||
|
name := fmt.Sprintf("%s_step_%d", c.prefix, i)
|
||||||
|
step := c.createProcess(name, container)
|
||||||
|
stage.Steps = append(stage.Steps, step)
|
||||||
|
}
|
||||||
|
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
// func setupNetwork(step *backend.Step, network *libcompose.Network) {
|
||||||
|
// step.Networks = append(step.Networks, backend.Conn{
|
||||||
|
// Name: network.Name,
|
||||||
|
// // Aliases:
|
||||||
|
// })
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// func setupVolume(step *backend.Step, volume *libcompose.Volume) {
|
||||||
|
// step.Volumes = append(step.Volumes, volume.String())
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// var (
|
||||||
|
// // Default plugin used to clone the repository.
|
||||||
|
// defaultCloneImage = "plugins/git:latest"
|
||||||
|
//
|
||||||
|
// // Default plugin settings used to clone the repository.
|
||||||
|
// defaultCloneVargs = map[string]interface{}{
|
||||||
|
// "depth": 0,
|
||||||
|
// }
|
||||||
|
// )
|
||||||
|
//
|
||||||
|
// // defaultClone returns the default step for cloning an image.
|
||||||
|
// func defaultClone() *yaml.Container {
|
||||||
|
// return &yaml.Container{
|
||||||
|
// Image: defaultCloneImage,
|
||||||
|
// Vargs: defaultCloneVargs,
|
||||||
|
// }
|
||||||
|
// }
|
152
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/convert.go
generated
vendored
Normal file
152
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/convert.go
generated
vendored
Normal file
|
@ -0,0 +1,152 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"path"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (c *Compiler) createProcess(name string, container *yaml.Container) *backend.Step {
|
||||||
|
var (
|
||||||
|
detached bool
|
||||||
|
workingdir string
|
||||||
|
|
||||||
|
workspace = fmt.Sprintf("%s_default:%s", c.prefix, c.base)
|
||||||
|
privileged = container.Privileged
|
||||||
|
entrypoint = container.Entrypoint
|
||||||
|
command = container.Command
|
||||||
|
image = expandImage(container.Image)
|
||||||
|
// network = container.Network
|
||||||
|
)
|
||||||
|
|
||||||
|
networks := []backend.Conn{
|
||||||
|
backend.Conn{
|
||||||
|
Name: fmt.Sprintf("%s_default", c.prefix),
|
||||||
|
Aliases: c.aliases,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
volumes := []string{
|
||||||
|
workspace,
|
||||||
|
}
|
||||||
|
for _, volume := range container.Volumes.Volumes {
|
||||||
|
volumes = append(volumes, volume.String())
|
||||||
|
}
|
||||||
|
// if network == "" {
|
||||||
|
// network = fmt.Sprintf("%s_default", c.prefix)
|
||||||
|
// for _, alias := range c.aliases {
|
||||||
|
// // if alias != container.Name {
|
||||||
|
// aliases = append(aliases, alias)
|
||||||
|
// // }
|
||||||
|
// }
|
||||||
|
// } // host, bridge, none, container:<name>, overlay
|
||||||
|
|
||||||
|
// append default environment variables
|
||||||
|
environment := map[string]string{}
|
||||||
|
for k, v := range container.Environment {
|
||||||
|
environment[k] = v
|
||||||
|
}
|
||||||
|
for k, v := range c.env {
|
||||||
|
switch v {
|
||||||
|
case "", "0", "false":
|
||||||
|
continue
|
||||||
|
default:
|
||||||
|
environment[k] = v
|
||||||
|
|
||||||
|
// legacy code for drone plugins
|
||||||
|
if strings.HasPrefix(k, "CI_") {
|
||||||
|
p := strings.Replace(k, "CI_", "DRONE_", 1)
|
||||||
|
environment[p] = v
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
environment["CI_WORKSPACE"] = path.Join(c.base, c.path)
|
||||||
|
environment["DRONE_WORKSPACE"] = path.Join(c.base, c.path)
|
||||||
|
|
||||||
|
if !isService(container) {
|
||||||
|
workingdir = path.Join(c.base, c.path)
|
||||||
|
}
|
||||||
|
|
||||||
|
if isService(container) {
|
||||||
|
detached = true
|
||||||
|
}
|
||||||
|
|
||||||
|
if isPlugin(container) {
|
||||||
|
paramsToEnv(container.Vargs, environment)
|
||||||
|
|
||||||
|
if imageMatches(container.Image, c.escalated) {
|
||||||
|
privileged = true
|
||||||
|
entrypoint = []string{}
|
||||||
|
command = []string{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if isShell(container) {
|
||||||
|
entrypoint = []string{"/bin/sh", "-c"}
|
||||||
|
command = []string{"echo $CI_SCRIPT | base64 -d | /bin/sh -e"}
|
||||||
|
environment["CI_SCRIPT"] = generateScriptPosix(container.Commands)
|
||||||
|
environment["HOME"] = "/root"
|
||||||
|
environment["SHELL"] = "/bin/sh"
|
||||||
|
}
|
||||||
|
|
||||||
|
return &backend.Step{
|
||||||
|
Name: name,
|
||||||
|
Alias: container.Name,
|
||||||
|
Image: image,
|
||||||
|
Pull: container.Pull,
|
||||||
|
Detached: detached,
|
||||||
|
Privileged: privileged,
|
||||||
|
WorkingDir: workingdir,
|
||||||
|
Environment: environment,
|
||||||
|
Labels: container.Labels,
|
||||||
|
Entrypoint: entrypoint,
|
||||||
|
Command: command,
|
||||||
|
ExtraHosts: container.ExtraHosts,
|
||||||
|
Volumes: volumes,
|
||||||
|
Devices: container.Devices,
|
||||||
|
Networks: networks,
|
||||||
|
DNS: container.DNS,
|
||||||
|
DNSSearch: container.DNSSearch,
|
||||||
|
MemSwapLimit: int64(container.MemSwapLimit),
|
||||||
|
MemLimit: int64(container.MemLimit),
|
||||||
|
ShmSize: int64(container.ShmSize),
|
||||||
|
CPUQuota: int64(container.CPUQuota),
|
||||||
|
CPUShares: int64(container.CPUShares),
|
||||||
|
CPUSet: container.CPUSet,
|
||||||
|
AuthConfig: backend.Auth{
|
||||||
|
Username: container.AuthConfig.Username,
|
||||||
|
Password: container.AuthConfig.Password,
|
||||||
|
Email: container.AuthConfig.Email,
|
||||||
|
},
|
||||||
|
OnSuccess: container.Constraints.Status.Match("success"),
|
||||||
|
OnFailure: (len(container.Constraints.Status.Include)+
|
||||||
|
len(container.Constraints.Status.Exclude) != 0) &&
|
||||||
|
container.Constraints.Status.Match("failure"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func imageMatches(image string, to []string) bool {
|
||||||
|
image = trimImage(image)
|
||||||
|
for _, i := range to {
|
||||||
|
if image == i {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func isPlugin(c *yaml.Container) bool {
|
||||||
|
return len(c.Vargs) != 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func isShell(c *yaml.Container) bool {
|
||||||
|
return len(c.Commands) != 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func isService(c *yaml.Container) bool {
|
||||||
|
return c.Detached || (isPlugin(c) == false && isShell(c) == false)
|
||||||
|
}
|
36
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/image.go
generated
vendored
Normal file
36
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/image.go
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/docker/docker/reference"
|
||||||
|
)
|
||||||
|
|
||||||
|
// trimImage returns the short image name without tag.
|
||||||
|
func trimImage(name string) string {
|
||||||
|
ref, err := reference.ParseNamed(name)
|
||||||
|
if err != nil {
|
||||||
|
return name
|
||||||
|
}
|
||||||
|
return reference.TrimNamed(ref).String()
|
||||||
|
}
|
||||||
|
|
||||||
|
// expandImage returns the fully qualified image name.
|
||||||
|
func expandImage(name string) string {
|
||||||
|
ref, err := reference.ParseNamed(name)
|
||||||
|
if err != nil {
|
||||||
|
return name
|
||||||
|
}
|
||||||
|
return reference.WithDefaultTag(ref).String()
|
||||||
|
}
|
||||||
|
|
||||||
|
// matchImage returns true if the image name matches
|
||||||
|
// an image in the list. Note the image tag is not used
|
||||||
|
// in the matching logic.
|
||||||
|
func matchImage(from string, to ...string) bool {
|
||||||
|
from = trimImage(from)
|
||||||
|
for _, match := range to {
|
||||||
|
if from == match {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
145
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/option.go
generated
vendored
Normal file
145
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/option.go
generated
vendored
Normal file
|
@ -0,0 +1,145 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Option configures a compiler option.
|
||||||
|
type Option func(*Compiler)
|
||||||
|
|
||||||
|
// WithVolumes configutes the compiler with default volumes that
|
||||||
|
// are mounted to each container in the pipeline.
|
||||||
|
func WithVolumes(volumes ...string) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.volumes = volumes
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithMetadata configutes the compiler with the repostiory, build
|
||||||
|
// and system metadata. The metadata is used to remove steps from
|
||||||
|
// the compiled pipeline configuration that should be skipped. The
|
||||||
|
// metadata is also added to each container as environment variables.
|
||||||
|
func WithMetadata(metadata frontend.Metadata) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.metadata = metadata
|
||||||
|
|
||||||
|
for k, v := range metadata.Environ() {
|
||||||
|
compiler.env[k] = v
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithNetrc configures the compiler with netrc authentication
|
||||||
|
// credentials added by default to every container in the pipeline.
|
||||||
|
func WithNetrc(username, password, machine string) Option {
|
||||||
|
return WithEnviron(
|
||||||
|
map[string]string{
|
||||||
|
"CI_NETRC_USERNAME": username,
|
||||||
|
"CI_NETRC_PASSWORD": password,
|
||||||
|
"CI_NETRC_MACHINE": machine,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithWorkspace configures the compiler with the workspace base
|
||||||
|
// and path. The workspace base is a volume created at runtime and
|
||||||
|
// mounted into all containers in the pipeline. The base and path
|
||||||
|
// are joined to provide the working directory for all build and
|
||||||
|
// plugin steps in the pipeline.
|
||||||
|
func WithWorkspace(base, path string) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.base = base
|
||||||
|
compiler.path = path
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithEscalated configures the compiler to automatically execute
|
||||||
|
// images as privileged containers if the match the given list.
|
||||||
|
func WithEscalated(images ...string) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.escalated = images
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithPrefix configures the compiler with the prefix. The prefix is
|
||||||
|
// used to prefix container, volume and network names to avoid
|
||||||
|
// collision at runtime.
|
||||||
|
func WithPrefix(prefix string) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.prefix = prefix
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithLocal configures the compiler with the local flag. The local
|
||||||
|
// flag indicates the pipeline execution is running in a local development
|
||||||
|
// environment with a mounted local working directory.
|
||||||
|
func WithLocal(local bool) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
compiler.local = local
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithEnviron configures the compiler with environment variables
|
||||||
|
// added by default to every container in the pipeline.
|
||||||
|
func WithEnviron(env map[string]string) Option {
|
||||||
|
return func(compiler *Compiler) {
|
||||||
|
for k, v := range env {
|
||||||
|
compiler.env[k] = v
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithProxy configures the compiler with HTTP_PROXY, HTTPS_PROXY,
|
||||||
|
// and NO_PROXY environment variables added by default to every
|
||||||
|
// container in the pipeline.
|
||||||
|
func WithProxy() Option {
|
||||||
|
return WithEnviron(
|
||||||
|
map[string]string{
|
||||||
|
"no_proxy": noProxy,
|
||||||
|
"NO_PROXY": noProxy,
|
||||||
|
"http_proxy": httpProxy,
|
||||||
|
"HTTP_PROXY": httpProxy,
|
||||||
|
"HTTPS_PROXY": httpsProxy,
|
||||||
|
"https_proxy": httpsProxy,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO(bradrydzewski) consider an alternate approach to
|
||||||
|
// WithProxy where the proxy strings are passed directly
|
||||||
|
// to the function as named parameters.
|
||||||
|
|
||||||
|
// func WithProxy2(http, https, none string) Option {
|
||||||
|
// return WithEnviron(
|
||||||
|
// map[string]string{
|
||||||
|
// "no_proxy": none,
|
||||||
|
// "NO_PROXY": none,
|
||||||
|
// "http_proxy": http,
|
||||||
|
// "HTTP_PROXY": http,
|
||||||
|
// "HTTPS_PROXY": https,
|
||||||
|
// "https_proxy": https,
|
||||||
|
// },
|
||||||
|
// )
|
||||||
|
// }
|
||||||
|
|
||||||
|
var (
|
||||||
|
noProxy = getenv("no_proxy")
|
||||||
|
httpProxy = getenv("https_proxy")
|
||||||
|
httpsProxy = getenv("https_proxy")
|
||||||
|
)
|
||||||
|
|
||||||
|
// getenv returns the named environment variable.
|
||||||
|
func getenv(name string) (value string) {
|
||||||
|
name = strings.ToUpper(name)
|
||||||
|
if value := os.Getenv(name); value != "" {
|
||||||
|
return value
|
||||||
|
}
|
||||||
|
name = strings.ToLower(name)
|
||||||
|
if value := os.Getenv(name); value != "" {
|
||||||
|
return value
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
65
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/params.go
generated
vendored
Normal file
65
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/params.go
generated
vendored
Normal file
|
@ -0,0 +1,65 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"reflect"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
json "github.com/ghodss/yaml"
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
// paramsToEnv uses reflection to convert a map[string]interface to a list
|
||||||
|
// of environment variables.
|
||||||
|
func paramsToEnv(from map[string]interface{}, to map[string]string) error {
|
||||||
|
for k, v := range from {
|
||||||
|
if v == nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
t := reflect.TypeOf(v)
|
||||||
|
vv := reflect.ValueOf(v)
|
||||||
|
|
||||||
|
k = "PLUGIN_" + strings.ToUpper(k)
|
||||||
|
|
||||||
|
switch t.Kind() {
|
||||||
|
case reflect.Bool:
|
||||||
|
to[k] = strconv.FormatBool(vv.Bool())
|
||||||
|
|
||||||
|
case reflect.String:
|
||||||
|
to[k] = vv.String()
|
||||||
|
|
||||||
|
case reflect.Int, reflect.Int16, reflect.Int32, reflect.Int64, reflect.Int8:
|
||||||
|
to[k] = fmt.Sprintf("%v", vv.Int())
|
||||||
|
|
||||||
|
case reflect.Float32, reflect.Float64:
|
||||||
|
to[k] = fmt.Sprintf("%v", vv.Float())
|
||||||
|
|
||||||
|
case reflect.Map:
|
||||||
|
yml, _ := yaml.Marshal(vv.Interface())
|
||||||
|
out, _ := json.YAMLToJSON(yml)
|
||||||
|
to[k] = string(out)
|
||||||
|
|
||||||
|
case reflect.Slice:
|
||||||
|
out, err := yaml.Marshal(vv.Interface())
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
in := []string{}
|
||||||
|
err = yaml.Unmarshal(out, &in)
|
||||||
|
if err == nil {
|
||||||
|
to[k] = strings.Join(in, ",")
|
||||||
|
} else {
|
||||||
|
out, err = json.YAMLToJSON(out)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
to[k] = string(out)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
52
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/script_posix.go
generated
vendored
Normal file
52
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/script_posix.go
generated
vendored
Normal file
|
@ -0,0 +1,52 @@
|
||||||
|
package compiler
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"encoding/base64"
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// generateScriptPosix is a helper function that generates a build script
|
||||||
|
// for a linux container using the given
|
||||||
|
func generateScriptPosix(commands []string) string {
|
||||||
|
var buf bytes.Buffer
|
||||||
|
for _, command := range commands {
|
||||||
|
escaped := fmt.Sprintf("%q", command)
|
||||||
|
escaped = strings.Replace(escaped, "$", `\$`, -1)
|
||||||
|
buf.WriteString(fmt.Sprintf(
|
||||||
|
traceScript,
|
||||||
|
escaped,
|
||||||
|
command,
|
||||||
|
))
|
||||||
|
}
|
||||||
|
script := fmt.Sprintf(
|
||||||
|
setupScript,
|
||||||
|
buf.String(),
|
||||||
|
)
|
||||||
|
return base64.StdEncoding.EncodeToString([]byte(script))
|
||||||
|
}
|
||||||
|
|
||||||
|
// setupScript is a helper script this is added to the build to ensure
|
||||||
|
// a minimum set of environment variables are set correctly.
|
||||||
|
const setupScript = `
|
||||||
|
if [ -n "$CI_NETRC_MACHINE" ]; then
|
||||||
|
cat <<EOF > $HOME/.netrc
|
||||||
|
machine $CI_NETRC_MACHINE
|
||||||
|
login $CI_NETRC_USERNAME
|
||||||
|
password $CI_NETRC_PASSWORD
|
||||||
|
EOF
|
||||||
|
chmod 0600 $HOME/.netrc
|
||||||
|
fi
|
||||||
|
unset CI_NETRC_USERNAME
|
||||||
|
unset CI_NETRC_PASSWORD
|
||||||
|
unset CI_SCRIPT
|
||||||
|
%s
|
||||||
|
`
|
||||||
|
|
||||||
|
// traceScript is a helper script that is added to the build script
|
||||||
|
// to trace a command.
|
||||||
|
const traceScript = `
|
||||||
|
echo + %s
|
||||||
|
%s
|
||||||
|
`
|
1
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/script_win.go
generated
vendored
Normal file
1
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/script_win.go
generated
vendored
Normal file
|
@ -0,0 +1 @@
|
||||||
|
package compiler
|
68
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/config.go
generated
vendored
Normal file
68
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/config.go
generated
vendored
Normal file
|
@ -0,0 +1,68 @@
|
||||||
|
package yaml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
libcompose "github.com/docker/libcompose/yaml"
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Config defines a pipeline configuration.
|
||||||
|
Config struct {
|
||||||
|
Platform string
|
||||||
|
Branches Constraint
|
||||||
|
Workspace Workspace
|
||||||
|
Clone Containers
|
||||||
|
Pipeline Containers
|
||||||
|
Services Containers
|
||||||
|
Networks Networks
|
||||||
|
Volumes Volumes
|
||||||
|
Labels libcompose.SliceorMap
|
||||||
|
}
|
||||||
|
|
||||||
|
// Workspace defines a pipeline workspace.
|
||||||
|
Workspace struct {
|
||||||
|
Base string
|
||||||
|
Path string
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Parse parses the configuration from bytes b.
|
||||||
|
func Parse(r io.Reader) (*Config, error) {
|
||||||
|
out, err := ioutil.ReadAll(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return ParseBytes(out)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseBytes parses the configuration from bytes b.
|
||||||
|
func ParseBytes(b []byte) (*Config, error) {
|
||||||
|
out := new(Config)
|
||||||
|
err := yaml.Unmarshal(b, out)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return out, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseString parses the configuration from string s.
|
||||||
|
func ParseString(s string) (*Config, error) {
|
||||||
|
return ParseBytes(
|
||||||
|
[]byte(s),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseFile parses the configuration from path p.
|
||||||
|
func ParseFile(p string) (*Config, error) {
|
||||||
|
f, err := os.Open(p)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer f.Close()
|
||||||
|
return Parse(f)
|
||||||
|
}
|
152
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/constraint.go
generated
vendored
Normal file
152
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/constraint.go
generated
vendored
Normal file
|
@ -0,0 +1,152 @@
|
||||||
|
package yaml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"path/filepath"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend"
|
||||||
|
libcompose "github.com/docker/libcompose/yaml"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Constraints defines a set of runtime constraints.
|
||||||
|
Constraints struct {
|
||||||
|
Repo Constraint
|
||||||
|
Instance Constraint
|
||||||
|
Platform Constraint
|
||||||
|
Environment Constraint
|
||||||
|
Event Constraint
|
||||||
|
Branch Constraint
|
||||||
|
Status Constraint
|
||||||
|
Matrix ConstraintMap
|
||||||
|
}
|
||||||
|
|
||||||
|
// Constraint defines a runtime constraint.
|
||||||
|
Constraint struct {
|
||||||
|
Include []string
|
||||||
|
Exclude []string
|
||||||
|
}
|
||||||
|
|
||||||
|
// ConstraintMap defines a runtime constraint map.
|
||||||
|
ConstraintMap struct {
|
||||||
|
Include map[string]string
|
||||||
|
Exclude map[string]string
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Match returns true if all constraints match the given input. If a single
|
||||||
|
// constraint fails a false value is returned.
|
||||||
|
func (c *Constraints) Match(metadata frontend.Metadata) bool {
|
||||||
|
return c.Platform.Match(metadata.Sys.Arch) &&
|
||||||
|
c.Environment.Match(metadata.Curr.Target) &&
|
||||||
|
c.Event.Match(metadata.Curr.Event) &&
|
||||||
|
c.Branch.Match(metadata.Curr.Commit.Branch) &&
|
||||||
|
c.Repo.Match(metadata.Repo.Name) &&
|
||||||
|
c.Matrix.Match(metadata.Job.Matrix)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Match returns true if the string matches the include patterns and does not
|
||||||
|
// match any of the exclude patterns.
|
||||||
|
func (c *Constraint) Match(v string) bool {
|
||||||
|
if c.Excludes(v) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if c.Includes(v) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
if len(c.Include) == 0 {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// Includes returns true if the string matches the include patterns.
|
||||||
|
func (c *Constraint) Includes(v string) bool {
|
||||||
|
for _, pattern := range c.Include {
|
||||||
|
if ok, _ := filepath.Match(pattern, v); ok {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// Excludes returns true if the string matches the exclude patterns.
|
||||||
|
func (c *Constraint) Excludes(v string) bool {
|
||||||
|
for _, pattern := range c.Exclude {
|
||||||
|
if ok, _ := filepath.Match(pattern, v); ok {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnmarshalYAML unmarshals the constraint.
|
||||||
|
func (c *Constraint) UnmarshalYAML(unmarshal func(interface{}) error) error {
|
||||||
|
var out1 = struct {
|
||||||
|
Include libcompose.Stringorslice
|
||||||
|
Exclude libcompose.Stringorslice
|
||||||
|
}{}
|
||||||
|
|
||||||
|
var out2 libcompose.Stringorslice
|
||||||
|
|
||||||
|
unmarshal(&out1)
|
||||||
|
unmarshal(&out2)
|
||||||
|
|
||||||
|
c.Exclude = out1.Exclude
|
||||||
|
c.Include = append(
|
||||||
|
out1.Include,
|
||||||
|
out2...,
|
||||||
|
)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Match returns true if the params matches the include key values and does not
|
||||||
|
// match any of the exclude key values.
|
||||||
|
func (c *ConstraintMap) Match(params map[string]string) bool {
|
||||||
|
// when no includes or excludes automatically match
|
||||||
|
if len(c.Include) == 0 && len(c.Exclude) == 0 {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
// exclusions are processed first. So we can include everything and then
|
||||||
|
// selectively include others.
|
||||||
|
if len(c.Exclude) != 0 {
|
||||||
|
var matches int
|
||||||
|
|
||||||
|
for key, val := range c.Exclude {
|
||||||
|
if params[key] == val {
|
||||||
|
matches++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if matches == len(c.Exclude) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for key, val := range c.Include {
|
||||||
|
if params[key] != val {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnmarshalYAML unmarshals the constraint map.
|
||||||
|
func (c *ConstraintMap) UnmarshalYAML(unmarshal func(interface{}) error) error {
|
||||||
|
out1 := struct {
|
||||||
|
Include map[string]string
|
||||||
|
Exclude map[string]string
|
||||||
|
}{
|
||||||
|
Include: map[string]string{},
|
||||||
|
Exclude: map[string]string{},
|
||||||
|
}
|
||||||
|
|
||||||
|
out2 := map[string]string{}
|
||||||
|
|
||||||
|
unmarshal(&out1)
|
||||||
|
unmarshal(&out2)
|
||||||
|
|
||||||
|
c.Include = out1.Include
|
||||||
|
c.Exclude = out1.Exclude
|
||||||
|
for k, v := range out2 {
|
||||||
|
c.Include[k] = v
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
80
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/container.go
generated
vendored
Normal file
80
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/container.go
generated
vendored
Normal file
|
@ -0,0 +1,80 @@
|
||||||
|
package yaml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
libcompose "github.com/docker/libcompose/yaml"
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// AuthConfig defines registry authentication credentials.
|
||||||
|
AuthConfig struct {
|
||||||
|
Username string
|
||||||
|
Password string
|
||||||
|
Email string
|
||||||
|
}
|
||||||
|
|
||||||
|
// Containers denotes an ordered collection of containers.
|
||||||
|
Containers struct {
|
||||||
|
Containers []*Container
|
||||||
|
}
|
||||||
|
|
||||||
|
// Container defines a container.
|
||||||
|
Container struct {
|
||||||
|
AuthConfig AuthConfig `yaml:"auth_config,omitempty"`
|
||||||
|
CapAdd []string `yaml:"cap_add,omitempty"`
|
||||||
|
CapDrop []string `yaml:"cap_drop,omitempty"`
|
||||||
|
Command libcompose.Command `yaml:"command,omitempty"`
|
||||||
|
Commands libcompose.Stringorslice `yaml:"commands,omitempty"`
|
||||||
|
CPUQuota libcompose.StringorInt `yaml:"cpu_quota,omitempty"`
|
||||||
|
CPUSet string `yaml:"cpuset,omitempty"`
|
||||||
|
CPUShares libcompose.StringorInt `yaml:"cpu_shares,omitempty"`
|
||||||
|
Detached bool `yaml:"detach,omitempty"`
|
||||||
|
Devices []string `yaml:"devices,omitempty"`
|
||||||
|
DNS libcompose.Stringorslice `yaml:"dns,omitempty"`
|
||||||
|
DNSSearch libcompose.Stringorslice `yaml:"dns_search,omitempty"`
|
||||||
|
Entrypoint libcompose.Command `yaml:"entrypoint,omitempty"`
|
||||||
|
Environment libcompose.SliceorMap `yaml:"environment,omitempty"`
|
||||||
|
ExtraHosts []string `yaml:"extra_hosts,omitempty"`
|
||||||
|
Group string `yaml:"group,omitempty"`
|
||||||
|
Image string `yaml:"image,omitempty"`
|
||||||
|
Isolation string `yaml:"isolation,omitempty"`
|
||||||
|
Labels libcompose.SliceorMap `yaml:"labels,omitempty"`
|
||||||
|
MemLimit libcompose.MemStringorInt `yaml:"mem_limit,omitempty"`
|
||||||
|
MemSwapLimit libcompose.MemStringorInt `yaml:"memswap_limit,omitempty"`
|
||||||
|
MemSwappiness libcompose.MemStringorInt `yaml:"mem_swappiness,omitempty"`
|
||||||
|
Name string `yaml:"name,omitempty"`
|
||||||
|
NetworkMode string `yaml:"network_mode,omitempty"`
|
||||||
|
Networks libcompose.Networks `yaml:"networks,omitempty"`
|
||||||
|
Privileged bool `yaml:"privileged,omitempty"`
|
||||||
|
Pull bool `yaml:"pull,omitempty"`
|
||||||
|
ShmSize libcompose.MemStringorInt `yaml:"shm_size,omitempty"`
|
||||||
|
Ulimits libcompose.Ulimits `yaml:"ulimits,omitempty"`
|
||||||
|
Volumes libcompose.Volumes `yaml:"volumes,omitempty"`
|
||||||
|
Constraints Constraints `yaml:"when,omitempty"`
|
||||||
|
Vargs map[string]interface{} `yaml:",inline"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// UnmarshalYAML implements the Unmarshaller interface.
|
||||||
|
func (c *Containers) UnmarshalYAML(unmarshal func(interface{}) error) error {
|
||||||
|
slice := yaml.MapSlice{}
|
||||||
|
if err := unmarshal(&slice); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, s := range slice {
|
||||||
|
container := Container{}
|
||||||
|
out, _ := yaml.Marshal(s.Value)
|
||||||
|
|
||||||
|
if err := yaml.Unmarshal(out, &container); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if container.Name == "" {
|
||||||
|
container.Name = fmt.Sprintf("%v", s.Key)
|
||||||
|
}
|
||||||
|
c.Containers = append(c.Containers, &container)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
109
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/linter/linter.go
generated
vendored
Normal file
109
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/linter/linter.go
generated
vendored
Normal file
|
@ -0,0 +1,109 @@
|
||||||
|
package linter
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/frontend/yaml"
|
||||||
|
)
|
||||||
|
|
||||||
|
// A Linter lints a pipeline configuration.
|
||||||
|
type Linter struct {
|
||||||
|
trusted bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates a new Linter with options.
|
||||||
|
func New(opts ...Option) *Linter {
|
||||||
|
linter := new(Linter)
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(linter)
|
||||||
|
}
|
||||||
|
return linter
|
||||||
|
}
|
||||||
|
|
||||||
|
// Lint lints the configuration.
|
||||||
|
func (l *Linter) Lint(c *yaml.Config) error {
|
||||||
|
var containers []*yaml.Container
|
||||||
|
containers = append(containers, c.Pipeline.Containers...)
|
||||||
|
containers = append(containers, c.Services.Containers...)
|
||||||
|
|
||||||
|
for _, container := range containers {
|
||||||
|
if err := l.lintImage(container); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if l.trusted == false {
|
||||||
|
if err := l.lintTrusted(container); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if isService(container) == false {
|
||||||
|
if err := l.lintEntrypoint(container); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(c.Pipeline.Containers) == 0 {
|
||||||
|
return fmt.Errorf("Invalid or missing pipeline section")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *Linter) lintImage(c *yaml.Container) error {
|
||||||
|
if len(c.Image) == 0 {
|
||||||
|
return fmt.Errorf("Invalid or missing image")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *Linter) lintEntrypoint(c *yaml.Container) error {
|
||||||
|
if len(c.Entrypoint) != 0 {
|
||||||
|
return fmt.Errorf("Cannot override container entrypoint")
|
||||||
|
}
|
||||||
|
if len(c.Command) != 0 {
|
||||||
|
return fmt.Errorf("Cannot override container command")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *Linter) lintTrusted(c *yaml.Container) error {
|
||||||
|
if c.Privileged {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use privileged mode")
|
||||||
|
}
|
||||||
|
if c.ShmSize != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to override shm_size")
|
||||||
|
}
|
||||||
|
if len(c.DNS) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use custom dns")
|
||||||
|
}
|
||||||
|
if len(c.DNSSearch) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use dns_search")
|
||||||
|
}
|
||||||
|
if len(c.Devices) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use devices")
|
||||||
|
}
|
||||||
|
if len(c.ExtraHosts) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use extra_hosts")
|
||||||
|
}
|
||||||
|
if len(c.NetworkMode) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use network_mode")
|
||||||
|
}
|
||||||
|
if c.Networks.Networks != nil && len(c.Networks.Networks) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use networks")
|
||||||
|
}
|
||||||
|
if c.Volumes.Volumes != nil && len(c.Volumes.Volumes) != 0 {
|
||||||
|
return fmt.Errorf("Insufficient privileges to use volumes")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func isService(c *yaml.Container) bool {
|
||||||
|
return !isScript(c) && !isPlugin(c)
|
||||||
|
}
|
||||||
|
|
||||||
|
func isScript(c *yaml.Container) bool {
|
||||||
|
return len(c.Commands) != 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func isPlugin(c *yaml.Container) bool {
|
||||||
|
return len(c.Vargs) != 0
|
||||||
|
}
|
11
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/linter/option.go
generated
vendored
Normal file
11
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/linter/option.go
generated
vendored
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
package linter
|
||||||
|
|
||||||
|
// Option configures a linting option.
|
||||||
|
type Option func(*Linter)
|
||||||
|
|
||||||
|
// WithTrusted adds the trusted option to the linter.
|
||||||
|
func WithTrusted(trusted bool) Option {
|
||||||
|
return func(linter *Linter) {
|
||||||
|
linter.trusted = trusted
|
||||||
|
}
|
||||||
|
}
|
117
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/matrix/matrix.go
generated
vendored
Normal file
117
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/matrix/matrix.go
generated
vendored
Normal file
|
@ -0,0 +1,117 @@
|
||||||
|
package matrix
|
||||||
|
|
||||||
|
import (
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
limitTags = 10
|
||||||
|
limitAxis = 25
|
||||||
|
)
|
||||||
|
|
||||||
|
// Matrix represents the build matrix.
|
||||||
|
type Matrix map[string][]string
|
||||||
|
|
||||||
|
// Axis represents a single permutation of entries from the build matrix.
|
||||||
|
type Axis map[string]string
|
||||||
|
|
||||||
|
// String returns a string representation of an Axis as a comma-separated list
|
||||||
|
// of environment variables.
|
||||||
|
func (a Axis) String() string {
|
||||||
|
var envs []string
|
||||||
|
for k, v := range a {
|
||||||
|
envs = append(envs, k+"="+v)
|
||||||
|
}
|
||||||
|
return strings.Join(envs, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse parses the Yaml matrix definition.
|
||||||
|
func Parse(data []byte) ([]Axis, error) {
|
||||||
|
|
||||||
|
axis, err := parseList(data)
|
||||||
|
if err == nil && len(axis) != 0 {
|
||||||
|
return axis, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
matrix, err := parse(data)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// if not a matrix build return an array with just the single axis.
|
||||||
|
if len(matrix) == 0 {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return calc(matrix), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseString parses the Yaml string matrix definition.
|
||||||
|
func ParseString(data string) ([]Axis, error) {
|
||||||
|
return Parse([]byte(data))
|
||||||
|
}
|
||||||
|
|
||||||
|
func calc(matrix Matrix) []Axis {
|
||||||
|
// calculate number of permutations and extract the list of tags
|
||||||
|
// (ie go_version, redis_version, etc)
|
||||||
|
var perm int
|
||||||
|
var tags []string
|
||||||
|
for k, v := range matrix {
|
||||||
|
perm *= len(v)
|
||||||
|
if perm == 0 {
|
||||||
|
perm = len(v)
|
||||||
|
}
|
||||||
|
tags = append(tags, k)
|
||||||
|
}
|
||||||
|
|
||||||
|
// structure to hold the transformed result set
|
||||||
|
axisList := []Axis{}
|
||||||
|
|
||||||
|
// for each axis calculate the uniqe set of values that should be used.
|
||||||
|
for p := 0; p < perm; p++ {
|
||||||
|
axis := map[string]string{}
|
||||||
|
decr := perm
|
||||||
|
for i, tag := range tags {
|
||||||
|
elems := matrix[tag]
|
||||||
|
decr = decr / len(elems)
|
||||||
|
elem := p / decr % len(elems)
|
||||||
|
axis[tag] = elems[elem]
|
||||||
|
|
||||||
|
// enforce a maximum number of tags in the build matrix.
|
||||||
|
if i > limitTags {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// append to the list of axis.
|
||||||
|
axisList = append(axisList, axis)
|
||||||
|
|
||||||
|
// enforce a maximum number of axis that should be calculated.
|
||||||
|
if p > limitAxis {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return axisList
|
||||||
|
}
|
||||||
|
|
||||||
|
func parse(raw []byte) (Matrix, error) {
|
||||||
|
data := struct {
|
||||||
|
Matrix map[string][]string
|
||||||
|
}{}
|
||||||
|
err := yaml.Unmarshal(raw, &data)
|
||||||
|
return data.Matrix, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseList(raw []byte) ([]Axis, error) {
|
||||||
|
data := struct {
|
||||||
|
Matrix struct {
|
||||||
|
Include []Axis
|
||||||
|
}
|
||||||
|
}{}
|
||||||
|
|
||||||
|
err := yaml.Unmarshal(raw, &data)
|
||||||
|
return data.Matrix.Include, err
|
||||||
|
}
|
48
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/network.go
generated
vendored
Normal file
48
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/network.go
generated
vendored
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
package yaml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Networks defines a collection of networks.
|
||||||
|
Networks struct {
|
||||||
|
Networks []*Network
|
||||||
|
}
|
||||||
|
|
||||||
|
// Network defines a container network.
|
||||||
|
Network struct {
|
||||||
|
Name string `yaml:"name,omitempty"`
|
||||||
|
Driver string `yaml:"driver,omitempty"`
|
||||||
|
DriverOpts map[string]string `yaml:"driver_opts,omitempty"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// UnmarshalYAML implements the Unmarshaller interface.
|
||||||
|
func (n *Networks) UnmarshalYAML(unmarshal func(interface{}) error) error {
|
||||||
|
slice := yaml.MapSlice{}
|
||||||
|
err := unmarshal(&slice)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, s := range slice {
|
||||||
|
nn := Network{}
|
||||||
|
out, _ := yaml.Marshal(s.Value)
|
||||||
|
|
||||||
|
err = yaml.Unmarshal(out, &nn)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if nn.Name == "" {
|
||||||
|
nn.Name = fmt.Sprintf("%v", s.Key)
|
||||||
|
}
|
||||||
|
if nn.Driver == "" {
|
||||||
|
nn.Driver = "bridge"
|
||||||
|
}
|
||||||
|
n.Networks = append(n.Networks, &nn)
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
48
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/volume.go
generated
vendored
Normal file
48
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/volume.go
generated
vendored
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
package yaml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"gopkg.in/yaml.v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Volumes defines a collection of volumes.
|
||||||
|
Volumes struct {
|
||||||
|
Volumes []*Volume
|
||||||
|
}
|
||||||
|
|
||||||
|
// Volume defines a container volume.
|
||||||
|
Volume struct {
|
||||||
|
Name string `yaml:"name,omitempty"`
|
||||||
|
Driver string `yaml:"driver,omitempty"`
|
||||||
|
DriverOpts map[string]string `yaml:"driver_opts,omitempty"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// UnmarshalYAML implements the Unmarshaller interface.
|
||||||
|
func (v *Volumes) UnmarshalYAML(unmarshal func(interface{}) error) error {
|
||||||
|
slice := yaml.MapSlice{}
|
||||||
|
err := unmarshal(&slice)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, s := range slice {
|
||||||
|
vv := Volume{}
|
||||||
|
out, _ := yaml.Marshal(s.Value)
|
||||||
|
|
||||||
|
err = yaml.Unmarshal(out, &vv)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if vv.Name == "" {
|
||||||
|
vv.Name = fmt.Sprintf("%v", s.Key)
|
||||||
|
}
|
||||||
|
if vv.Driver == "" {
|
||||||
|
vv.Driver = "local"
|
||||||
|
}
|
||||||
|
v.Volumes = append(v.Volumes, &vv)
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
36
vendor/github.com/cncd/pipeline/pipeline/interrupt/interrupt.go
generated
vendored
Normal file
36
vendor/github.com/cncd/pipeline/pipeline/interrupt/interrupt.go
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
package interrupt
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"os"
|
||||||
|
"os/signal"
|
||||||
|
)
|
||||||
|
|
||||||
|
// WithContext returns a copy of parent context whose Done channel is closed
|
||||||
|
// when an os interrupt signal is received.
|
||||||
|
func WithContext(ctx context.Context) context.Context {
|
||||||
|
return WithContextFunc(ctx, func() {
|
||||||
|
println("ctrl+c received, terminating process")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithContextFunc returns a copy of parent context that is cancelled when
|
||||||
|
// an os interrupt signal is received. The callback function f is invoked
|
||||||
|
// before cancellation.
|
||||||
|
func WithContextFunc(ctx context.Context, f func()) context.Context {
|
||||||
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
|
go func() {
|
||||||
|
c := make(chan os.Signal)
|
||||||
|
signal.Notify(c, os.Interrupt)
|
||||||
|
defer signal.Stop(c)
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
case <-c:
|
||||||
|
f()
|
||||||
|
cancel()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return ctx
|
||||||
|
}
|
20
vendor/github.com/cncd/pipeline/pipeline/logger.go
generated
vendored
Normal file
20
vendor/github.com/cncd/pipeline/pipeline/logger.go
generated
vendored
Normal file
|
@ -0,0 +1,20 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/multipart"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Logger handles the process logging.
|
||||||
|
type Logger interface {
|
||||||
|
Log(*backend.Step, multipart.Reader) error
|
||||||
|
}
|
||||||
|
|
||||||
|
// LogFunc type is an adapter to allow the use of an ordinary
|
||||||
|
// function for process logging.
|
||||||
|
type LogFunc func(*backend.Step, multipart.Reader) error
|
||||||
|
|
||||||
|
// Log calls f(proc, r).
|
||||||
|
func (f LogFunc) Log(step *backend.Step, r multipart.Reader) error {
|
||||||
|
return f(step, r)
|
||||||
|
}
|
1
vendor/github.com/cncd/pipeline/pipeline/multipart/doc.go
generated
vendored
Normal file
1
vendor/github.com/cncd/pipeline/pipeline/multipart/doc.go
generated
vendored
Normal file
|
@ -0,0 +1 @@
|
||||||
|
package multipart
|
102
vendor/github.com/cncd/pipeline/pipeline/multipart/reader.go
generated
vendored
Normal file
102
vendor/github.com/cncd/pipeline/pipeline/multipart/reader.go
generated
vendored
Normal file
|
@ -0,0 +1,102 @@
|
||||||
|
package multipart
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"io"
|
||||||
|
"mime/multipart"
|
||||||
|
"net/textproto"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Reader is an iterator over parts in a multipart log stream.
|
||||||
|
type Reader interface {
|
||||||
|
// NextPart returns the next part in the multipart or
|
||||||
|
// an error. When there are no more parts, the error
|
||||||
|
// io.EOF is returned.
|
||||||
|
NextPart() (Part, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// A Part represents a single part in a multipart body.
|
||||||
|
type Part interface {
|
||||||
|
io.Reader
|
||||||
|
|
||||||
|
// Header returns the headers of the body with the
|
||||||
|
// keys canonicalized.
|
||||||
|
Header() textproto.MIMEHeader
|
||||||
|
|
||||||
|
// FileName returns the filename parameter of the
|
||||||
|
// Content-Disposition header.
|
||||||
|
FileName() string
|
||||||
|
|
||||||
|
// FormName returns the name parameter if p has a
|
||||||
|
// Content-Disposition of type form-data.
|
||||||
|
FormName() string
|
||||||
|
}
|
||||||
|
|
||||||
|
// New returns a new multipart Reader.
|
||||||
|
func New(r io.Reader) Reader {
|
||||||
|
buf := bufio.NewReader(r)
|
||||||
|
out, _ := buf.Peek(4)
|
||||||
|
|
||||||
|
if bytes.Equal(out, []byte("MIME")) {
|
||||||
|
return &multipartReader{
|
||||||
|
reader: multipart.NewReader(buf, "boundary"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return &textReader{
|
||||||
|
reader: r,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
type multipartReader struct {
|
||||||
|
reader *multipart.Reader
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *multipartReader) NextPart() (Part, error) {
|
||||||
|
next, err := r.reader.NextPart()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
part := new(part)
|
||||||
|
part.Reader = next
|
||||||
|
part.filename = next.FileName()
|
||||||
|
part.formname = next.FormName()
|
||||||
|
part.header = next.Header
|
||||||
|
return part, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
type textReader struct {
|
||||||
|
reader io.Reader
|
||||||
|
done bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *textReader) NextPart() (Part, error) {
|
||||||
|
if r.done {
|
||||||
|
return nil, io.EOF
|
||||||
|
}
|
||||||
|
r.done = true
|
||||||
|
p := new(part)
|
||||||
|
p.Reader = r.reader
|
||||||
|
p.filename = "terminal.log"
|
||||||
|
return p, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type part struct {
|
||||||
|
io.Reader
|
||||||
|
|
||||||
|
filename string
|
||||||
|
formname string
|
||||||
|
header textproto.MIMEHeader
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *part) Header() textproto.MIMEHeader { return p.header }
|
||||||
|
func (p *part) FileName() string { return p.filename }
|
||||||
|
func (p *part) FormName() string { return p.filename }
|
38
vendor/github.com/cncd/pipeline/pipeline/option.go
generated
vendored
Normal file
38
vendor/github.com/cncd/pipeline/pipeline/option.go
generated
vendored
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Option configures a runtime option.
|
||||||
|
type Option func(*Runtime)
|
||||||
|
|
||||||
|
// WithEngine returns an option configured with a runtime engine.
|
||||||
|
func WithEngine(engine backend.Engine) Option {
|
||||||
|
return func(r *Runtime) {
|
||||||
|
r.engine = engine
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithLogger returns an option configured with a runtime logger.
|
||||||
|
func WithLogger(logger Logger) Option {
|
||||||
|
return func(r *Runtime) {
|
||||||
|
r.logger = logger
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithTracer returns an option configured with a runtime tracer.
|
||||||
|
func WithTracer(tracer Tracer) Option {
|
||||||
|
return func(r *Runtime) {
|
||||||
|
r.tracer = tracer
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithContext returns an option configured with a context.
|
||||||
|
func WithContext(ctx context.Context) Option {
|
||||||
|
return func(r *Runtime) {
|
||||||
|
r.ctx = ctx
|
||||||
|
}
|
||||||
|
}
|
37
vendor/github.com/cncd/pipeline/pipeline/parse.go
generated
vendored
Normal file
37
vendor/github.com/cncd/pipeline/pipeline/parse.go
generated
vendored
Normal file
|
@ -0,0 +1,37 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"io"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Parse parses the pipeline config from an io.Reader.
|
||||||
|
func Parse(r io.Reader) (*backend.Config, error) {
|
||||||
|
cfg := new(backend.Config)
|
||||||
|
err := json.NewDecoder(r).Decode(cfg)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return cfg, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseFile parses the pipeline config from a file.
|
||||||
|
func ParseFile(path string) (*backend.Config, error) {
|
||||||
|
f, err := os.Open(path)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer f.Close()
|
||||||
|
return Parse(f)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseString parses the pipeline config from a string.
|
||||||
|
func ParseString(s string) (*backend.Config, error) {
|
||||||
|
return Parse(
|
||||||
|
strings.NewReader(s),
|
||||||
|
)
|
||||||
|
}
|
175
vendor/github.com/cncd/pipeline/pipeline/pipeline.go
generated
vendored
Normal file
175
vendor/github.com/cncd/pipeline/pipeline/pipeline.go
generated
vendored
Normal file
|
@ -0,0 +1,175 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"golang.org/x/sync/errgroup"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
"github.com/cncd/pipeline/pipeline/multipart"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// State defines the pipeline and process state.
|
||||||
|
State struct {
|
||||||
|
// Global state of the pipeline.
|
||||||
|
Pipeline struct {
|
||||||
|
// Pipeline time started
|
||||||
|
Time int64 `json:"time"`
|
||||||
|
// Current pipeline step
|
||||||
|
Step *backend.Step `json:"step"`
|
||||||
|
// Current pipeline error state
|
||||||
|
Error error `json:"error"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Current process state.
|
||||||
|
Process *backend.State
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Runtime is a configuration runtime.
|
||||||
|
type Runtime struct {
|
||||||
|
err error
|
||||||
|
spec *backend.Config
|
||||||
|
engine backend.Engine
|
||||||
|
started int64
|
||||||
|
|
||||||
|
ctx context.Context
|
||||||
|
tracer Tracer
|
||||||
|
logger Logger
|
||||||
|
}
|
||||||
|
|
||||||
|
// New returns a new runtime using the specified runtime
|
||||||
|
// configuration and runtime engine.
|
||||||
|
func New(spec *backend.Config, opts ...Option) *Runtime {
|
||||||
|
r := new(Runtime)
|
||||||
|
r.spec = spec
|
||||||
|
r.ctx = context.Background()
|
||||||
|
for _, opts := range opts {
|
||||||
|
opts(r)
|
||||||
|
}
|
||||||
|
return r
|
||||||
|
}
|
||||||
|
|
||||||
|
// Run starts the runtime and waits for it to complete.
|
||||||
|
func (r *Runtime) Run() error {
|
||||||
|
defer func() {
|
||||||
|
r.engine.Destroy(r.spec)
|
||||||
|
}()
|
||||||
|
|
||||||
|
r.started = time.Now().Unix()
|
||||||
|
if err := r.engine.Setup(r.spec); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, stage := range r.spec.Stages {
|
||||||
|
select {
|
||||||
|
case <-r.ctx.Done():
|
||||||
|
return ErrCancel
|
||||||
|
case err := <-r.execAll(stage.Steps):
|
||||||
|
if err != nil {
|
||||||
|
r.err = err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return r.err
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
func (r *Runtime) execAll(procs []*backend.Step) <-chan error {
|
||||||
|
var g errgroup.Group
|
||||||
|
done := make(chan error)
|
||||||
|
|
||||||
|
for _, proc := range procs {
|
||||||
|
proc := proc
|
||||||
|
g.Go(func() error {
|
||||||
|
return r.exec(proc)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
done <- g.Wait()
|
||||||
|
close(done)
|
||||||
|
}()
|
||||||
|
return done
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
func (r *Runtime) exec(proc *backend.Step) error {
|
||||||
|
switch {
|
||||||
|
case r.err != nil && proc.OnFailure == false:
|
||||||
|
return nil
|
||||||
|
case r.err == nil && proc.OnSuccess == false:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if r.tracer != nil {
|
||||||
|
state := new(State)
|
||||||
|
state.Pipeline.Time = r.started
|
||||||
|
state.Pipeline.Error = r.err
|
||||||
|
state.Pipeline.Step = proc
|
||||||
|
state.Process = new(backend.State) // empty
|
||||||
|
if err := r.tracer.Trace(state); err == ErrSkip {
|
||||||
|
return nil
|
||||||
|
} else if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := r.engine.Exec(proc); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if r.logger != nil {
|
||||||
|
rc, err := r.engine.Tail(proc)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
r.logger.Log(proc, multipart.New(rc))
|
||||||
|
rc.Close()
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
if proc.Detached {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
wait, err := r.engine.Wait(proc)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if r.tracer != nil {
|
||||||
|
state := new(State)
|
||||||
|
state.Pipeline.Time = r.started
|
||||||
|
state.Pipeline.Error = r.err
|
||||||
|
state.Pipeline.Step = proc
|
||||||
|
state.Process = wait
|
||||||
|
if err := r.tracer.Trace(state); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if wait.OOMKilled {
|
||||||
|
return &OomError{
|
||||||
|
Name: proc.Name,
|
||||||
|
Code: wait.ExitCode,
|
||||||
|
}
|
||||||
|
} else if wait.ExitCode != 0 {
|
||||||
|
return &ExitError{
|
||||||
|
Name: proc.Name,
|
||||||
|
Code: wait.ExitCode,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
180
vendor/github.com/cncd/pipeline/pipeline/rpc/client.go
generated
vendored
Normal file
180
vendor/github.com/cncd/pipeline/pipeline/rpc/client.go
generated
vendored
Normal file
|
@ -0,0 +1,180 @@
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"log"
|
||||||
|
"math"
|
||||||
|
"net/http"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/gorilla/websocket"
|
||||||
|
"github.com/sourcegraph/jsonrpc2"
|
||||||
|
websocketrpc "github.com/sourcegraph/jsonrpc2/websocket"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
methodNext = "next"
|
||||||
|
methodNotify = "notify"
|
||||||
|
methodExtend = "extend"
|
||||||
|
methodUpdate = "update"
|
||||||
|
methodLog = "log"
|
||||||
|
methodSave = "save"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
saveReq struct {
|
||||||
|
ID string `json:"id"`
|
||||||
|
Mime string `json:"mime"`
|
||||||
|
Data []byte `json:"data"`
|
||||||
|
}
|
||||||
|
|
||||||
|
updateReq struct {
|
||||||
|
ID string `json:"id"`
|
||||||
|
State State `json:"state"`
|
||||||
|
}
|
||||||
|
|
||||||
|
logReq struct {
|
||||||
|
ID string `json:"id"`
|
||||||
|
Line *Line `json:"line"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
defaultRetryClount = math.MaxInt32
|
||||||
|
defaultBackoff = 10 * time.Second
|
||||||
|
)
|
||||||
|
|
||||||
|
// Client represents an rpc client.
|
||||||
|
type Client struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
conn *jsonrpc2.Conn
|
||||||
|
done bool
|
||||||
|
retry int
|
||||||
|
backoff time.Duration
|
||||||
|
endpoint string
|
||||||
|
token string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewClient returns a new Client.
|
||||||
|
func NewClient(endpoint string, opts ...Option) (*Client, error) {
|
||||||
|
cli := &Client{
|
||||||
|
endpoint: endpoint,
|
||||||
|
retry: defaultRetryClount,
|
||||||
|
backoff: defaultBackoff,
|
||||||
|
}
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(cli)
|
||||||
|
}
|
||||||
|
err := cli.openRetry()
|
||||||
|
return cli, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Next returns the next pipeline in the queue.
|
||||||
|
func (t *Client) Next(c context.Context) (*Pipeline, error) {
|
||||||
|
res := new(Pipeline)
|
||||||
|
err := t.call(c, methodNext, nil, res)
|
||||||
|
return res, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Notify returns true if the pipeline should be cancelled.
|
||||||
|
func (t *Client) Notify(c context.Context, id string) (bool, error) {
|
||||||
|
out := false
|
||||||
|
err := t.call(c, methodNotify, id, &out)
|
||||||
|
return out, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Extend extends the pipeline deadline.
|
||||||
|
func (t *Client) Extend(c context.Context, id string) error {
|
||||||
|
return t.call(c, methodExtend, id, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update updates the pipeline state.
|
||||||
|
func (t *Client) Update(c context.Context, id string, state State) error {
|
||||||
|
params := updateReq{id, state}
|
||||||
|
return t.call(c, methodUpdate, ¶ms, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Log writes the pipeline log entry.
|
||||||
|
func (t *Client) Log(c context.Context, id string, line *Line) error {
|
||||||
|
params := logReq{id, line}
|
||||||
|
return t.call(c, methodLog, ¶ms, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Save saves the pipeline artifact.
|
||||||
|
func (t *Client) Save(c context.Context, id, mime string, file io.Reader) error {
|
||||||
|
data, err := ioutil.ReadAll(file)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
params := saveReq{id, mime, data}
|
||||||
|
return t.call(c, methodSave, params, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close closes the client connection.
|
||||||
|
func (t *Client) Close() error {
|
||||||
|
t.Lock()
|
||||||
|
t.done = true
|
||||||
|
t.Unlock()
|
||||||
|
return t.conn.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
// call makes the remote prodedure call. If the call fails due to connectivity
|
||||||
|
// issues the connection is re-establish and call re-attempted.
|
||||||
|
func (t *Client) call(ctx context.Context, name string, req, res interface{}) error {
|
||||||
|
if err := t.conn.Call(ctx, name, req, res); err == nil {
|
||||||
|
return nil
|
||||||
|
} else if err != jsonrpc2.ErrClosed && err != io.ErrUnexpectedEOF {
|
||||||
|
log.Printf("rpc: error making call: %s", err)
|
||||||
|
return err
|
||||||
|
} else {
|
||||||
|
log.Printf("rpc: error making call: connection closed: %s", err)
|
||||||
|
}
|
||||||
|
if err := t.openRetry(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return t.conn.Call(ctx, name, req, res)
|
||||||
|
}
|
||||||
|
|
||||||
|
// openRetry opens the connection and will retry on failure until
|
||||||
|
// the connection is successfully open, or the maximum retry count
|
||||||
|
// is exceeded.
|
||||||
|
func (t *Client) openRetry() error {
|
||||||
|
for i := 0; i < t.retry; i++ {
|
||||||
|
err := t.open()
|
||||||
|
if err == nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err == io.EOF {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Printf("rpc: error re-connecting: %s", err)
|
||||||
|
<-time.After(t.backoff)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// open creates a websocket connection to a peer and establishes a json
|
||||||
|
// rpc communication stream.
|
||||||
|
func (t *Client) open() error {
|
||||||
|
t.Lock()
|
||||||
|
defer t.Unlock()
|
||||||
|
if t.done {
|
||||||
|
return io.EOF
|
||||||
|
}
|
||||||
|
header := map[string][]string{
|
||||||
|
"Content-Type": {"application/json-rpc"},
|
||||||
|
"Authorization": {"Bearer " + t.token},
|
||||||
|
}
|
||||||
|
conn, _, err := websocket.DefaultDialer.Dial(t.endpoint, http.Header(header))
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
stream := websocketrpc.NewObjectStream(conn)
|
||||||
|
t.conn = jsonrpc2.NewConn(context.Background(), stream, nil)
|
||||||
|
return nil
|
||||||
|
}
|
70
vendor/github.com/cncd/pipeline/pipeline/rpc/line.go
generated
vendored
Normal file
70
vendor/github.com/cncd/pipeline/pipeline/rpc/line.go
generated
vendored
Normal file
|
@ -0,0 +1,70 @@
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Identifies the type of line in the logs.
|
||||||
|
const (
|
||||||
|
LineStdout int = iota
|
||||||
|
LineStderr
|
||||||
|
LineExitCode
|
||||||
|
LineMetadata
|
||||||
|
LineProgress
|
||||||
|
)
|
||||||
|
|
||||||
|
// Line is a line of console output.
|
||||||
|
type Line struct {
|
||||||
|
Proc string `json:"proc,omitempty"`
|
||||||
|
Time int64 `json:"time,omitempty"`
|
||||||
|
Type int `json:"type,omitempty"`
|
||||||
|
Pos int `json:"pos,omityempty"`
|
||||||
|
Out string `json:"out,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *Line) String() string {
|
||||||
|
switch l.Type {
|
||||||
|
case LineExitCode:
|
||||||
|
return fmt.Sprintf("[%s] exit code %s", l.Proc, l.Out)
|
||||||
|
default:
|
||||||
|
return fmt.Sprintf("[%s:L%v:%vs] %s", l.Proc, l.Pos, l.Time, l.Out)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// LineWriter sends logs to the client.
|
||||||
|
type LineWriter struct {
|
||||||
|
peer Peer
|
||||||
|
id string
|
||||||
|
name string
|
||||||
|
num int
|
||||||
|
now time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewLineWriter returns a new line reader.
|
||||||
|
func NewLineWriter(peer Peer, id, name string) *LineWriter {
|
||||||
|
w := new(LineWriter)
|
||||||
|
w.peer = peer
|
||||||
|
w.id = id
|
||||||
|
w.name = name
|
||||||
|
w.num = 0
|
||||||
|
w.now = time.Now().UTC()
|
||||||
|
return w
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *LineWriter) Write(p []byte) (n int, err error) {
|
||||||
|
for _, part := range bytes.Split(p, []byte{'\n'}) {
|
||||||
|
line := &Line{
|
||||||
|
Out: string(part),
|
||||||
|
Proc: w.name,
|
||||||
|
Pos: w.num,
|
||||||
|
Time: int64(time.Since(w.now).Seconds()),
|
||||||
|
Type: LineStdout,
|
||||||
|
}
|
||||||
|
w.peer.Log(context.Background(), w.id, line)
|
||||||
|
w.num++
|
||||||
|
}
|
||||||
|
return len(p), nil
|
||||||
|
}
|
29
vendor/github.com/cncd/pipeline/pipeline/rpc/option.go
generated
vendored
Normal file
29
vendor/github.com/cncd/pipeline/pipeline/rpc/option.go
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import "time"
|
||||||
|
|
||||||
|
// Option configures a client option.
|
||||||
|
type Option func(*Client)
|
||||||
|
|
||||||
|
// WithBackoff configures the backoff duration when attempting
|
||||||
|
// to re-connect to a server.
|
||||||
|
func WithBackoff(d time.Duration) Option {
|
||||||
|
return func(c *Client) {
|
||||||
|
c.backoff = d
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithRetryLimit configures the maximum number of retries when
|
||||||
|
// connecting to the server.
|
||||||
|
func WithRetryLimit(i int) Option {
|
||||||
|
return func(c *Client) {
|
||||||
|
c.retry = i
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithToken configures the client authorization token.
|
||||||
|
func WithToken(t string) Option {
|
||||||
|
return func(c *Client) {
|
||||||
|
c.token = t
|
||||||
|
}
|
||||||
|
}
|
55
vendor/github.com/cncd/pipeline/pipeline/rpc/peer.go
generated
vendored
Normal file
55
vendor/github.com/cncd/pipeline/pipeline/rpc/peer.go
generated
vendored
Normal file
|
@ -0,0 +1,55 @@
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/cncd/pipeline/pipeline/backend"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// Filter defines filters for fetching items from the queue.
|
||||||
|
Filter struct {
|
||||||
|
Platform string `json:"platform"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// State defines the pipeline state.
|
||||||
|
State struct {
|
||||||
|
Proc string `json:"proc"`
|
||||||
|
Exited bool `json:"exited"`
|
||||||
|
ExitCode int `json:"exit_code"`
|
||||||
|
Started int64 `json:"started"`
|
||||||
|
Finished int64 `json:"finished"`
|
||||||
|
Error string `json:"error"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pipeline defines the pipeline execution details.
|
||||||
|
Pipeline struct {
|
||||||
|
ID string `json:"id"`
|
||||||
|
Config *backend.Config `json:"config"`
|
||||||
|
Timeout int64 `json:"timeout"`
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Peer defines a peer-to-peer connection.
|
||||||
|
type Peer interface {
|
||||||
|
// Next returns the next pipeline in the queue.
|
||||||
|
Next(c context.Context) (*Pipeline, error)
|
||||||
|
|
||||||
|
// Notify returns true if the pipeline should be cancelled.
|
||||||
|
// TODO: rename to Done, Wait?
|
||||||
|
Notify(c context.Context, id string) (bool, error)
|
||||||
|
|
||||||
|
// Extend extends the pipeline deadline
|
||||||
|
Extend(c context.Context, id string) error
|
||||||
|
|
||||||
|
// Update updates the pipeline state.
|
||||||
|
Update(c context.Context, id string, state State) error
|
||||||
|
|
||||||
|
// Save saves the pipeline artifact.
|
||||||
|
// TODO rename to Upload
|
||||||
|
Save(c context.Context, id, mime string, file io.Reader) error
|
||||||
|
|
||||||
|
// Log writes the pipeline log entry.
|
||||||
|
Log(c context.Context, id string, line *Line) error
|
||||||
|
}
|
126
vendor/github.com/cncd/pipeline/pipeline/rpc/server.go
generated
vendored
Normal file
126
vendor/github.com/cncd/pipeline/pipeline/rpc/server.go
generated
vendored
Normal file
|
@ -0,0 +1,126 @@
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"net/http"
|
||||||
|
|
||||||
|
"github.com/gorilla/websocket"
|
||||||
|
"github.com/sourcegraph/jsonrpc2"
|
||||||
|
websocketrpc "github.com/sourcegraph/jsonrpc2/websocket"
|
||||||
|
)
|
||||||
|
|
||||||
|
// errNoSuchMethod is returned when the name rpc method does not exist.
|
||||||
|
var errNoSuchMethod = errors.New("No such rpc method")
|
||||||
|
|
||||||
|
// noContext is an empty context used when no context is required.
|
||||||
|
var noContext = context.Background()
|
||||||
|
|
||||||
|
// Server represents an rpc server.
|
||||||
|
type Server struct {
|
||||||
|
peer Peer
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewServer returns an rpc Server.
|
||||||
|
func NewServer(peer Peer) *Server {
|
||||||
|
return &Server{peer}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ServeHTTP implements an http.Handler that answers rpc requests.
|
||||||
|
func (s *Server) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||||
|
upgrader := websocket.Upgrader{}
|
||||||
|
c, err := upgrader.Upgrade(w, r, nil)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
conn := jsonrpc2.NewConn(ctx,
|
||||||
|
websocketrpc.NewObjectStream(c),
|
||||||
|
jsonrpc2.HandlerWithError(s.router),
|
||||||
|
)
|
||||||
|
defer func() {
|
||||||
|
cancel()
|
||||||
|
conn.Close()
|
||||||
|
}()
|
||||||
|
<-conn.DisconnectNotify()
|
||||||
|
}
|
||||||
|
|
||||||
|
// router implements an jsonrpc2.Handler that answers RPC requests.
|
||||||
|
func (s *Server) router(ctx context.Context, conn *jsonrpc2.Conn, req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
switch req.Method {
|
||||||
|
case methodNext:
|
||||||
|
return s.next(ctx, req)
|
||||||
|
case methodNotify:
|
||||||
|
return s.notify(ctx, req)
|
||||||
|
case methodExtend:
|
||||||
|
return s.extend(ctx, req)
|
||||||
|
case methodUpdate:
|
||||||
|
return s.update(req)
|
||||||
|
case methodLog:
|
||||||
|
return s.log(req)
|
||||||
|
case methodSave:
|
||||||
|
return s.save(req)
|
||||||
|
default:
|
||||||
|
return nil, errNoSuchMethod
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// next unmarshals the rpc request parameters and invokes the peer.Next
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) next(ctx context.Context, req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
return s.peer.Next(ctx)
|
||||||
|
}
|
||||||
|
|
||||||
|
// notify unmarshals the rpc request parameters and invokes the peer.Notify
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) notify(ctx context.Context, req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
var id string
|
||||||
|
err := json.Unmarshal([]byte(*req.Params), &id)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return s.peer.Notify(ctx, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
// extend unmarshals the rpc request parameters and invokes the peer.Extend
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) extend(ctx context.Context, req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
var id string
|
||||||
|
err := json.Unmarshal([]byte(*req.Params), &id)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, s.peer.Extend(ctx, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
// update unmarshals the rpc request parameters and invokes the peer.Update
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) update(req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
in := new(updateReq)
|
||||||
|
if err := json.Unmarshal([]byte(*req.Params), in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, s.peer.Update(noContext, in.ID, in.State)
|
||||||
|
}
|
||||||
|
|
||||||
|
// log unmarshals the rpc request parameters and invokes the peer.Log
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) log(req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
in := new(logReq)
|
||||||
|
if err := json.Unmarshal([]byte(*req.Params), in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, s.peer.Log(noContext, in.ID, in.Line)
|
||||||
|
}
|
||||||
|
|
||||||
|
// save unmarshals the rpc request parameters and invokes the peer.Save
|
||||||
|
// procedure. The results are retuned and written to the rpc response.
|
||||||
|
func (s *Server) save(req *jsonrpc2.Request) (interface{}, error) {
|
||||||
|
in := new(saveReq)
|
||||||
|
if err := json.Unmarshal([]byte(*req.Params), in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return nil, s.peer.Save(noContext, in.ID, in.Mime, bytes.NewBuffer(in.Data))
|
||||||
|
}
|
45
vendor/github.com/cncd/pipeline/pipeline/tracer.go
generated
vendored
Normal file
45
vendor/github.com/cncd/pipeline/pipeline/tracer.go
generated
vendored
Normal file
|
@ -0,0 +1,45 @@
|
||||||
|
package pipeline
|
||||||
|
|
||||||
|
import (
|
||||||
|
"strconv"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Tracer handles process tracing.
|
||||||
|
type Tracer interface {
|
||||||
|
Trace(*State) error
|
||||||
|
}
|
||||||
|
|
||||||
|
// TraceFunc type is an adapter to allow the use of ordinary
|
||||||
|
// functions as a Tracer.
|
||||||
|
type TraceFunc func(*State) error
|
||||||
|
|
||||||
|
// Trace calls f(proc, state).
|
||||||
|
func (f TraceFunc) Trace(state *State) error {
|
||||||
|
return f(state)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DefaultTracer provides a tracer that updates the CI_ enviornment
|
||||||
|
// variables to include the correct timestamp and status.
|
||||||
|
// TODO(bradrydzewski) find either a new home or better name for this.
|
||||||
|
var DefaultTracer = TraceFunc(func(state *State) error {
|
||||||
|
if state.Process.Exited {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if state.Pipeline.Step.Environment == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "success"
|
||||||
|
state.Pipeline.Step.Environment["CI_BUILD_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10)
|
||||||
|
state.Pipeline.Step.Environment["CI_BUILD_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10)
|
||||||
|
|
||||||
|
state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "success"
|
||||||
|
state.Pipeline.Step.Environment["CI_JOB_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10)
|
||||||
|
state.Pipeline.Step.Environment["CI_JOB_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10)
|
||||||
|
|
||||||
|
if state.Pipeline.Error != nil {
|
||||||
|
state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "failure"
|
||||||
|
state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "failure"
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
29
vendor/github.com/cncd/pubsub/LICENSE
generated
vendored
Normal file
29
vendor/github.com/cncd/pubsub/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
BSD 3-Clause License
|
||||||
|
|
||||||
|
Copyright (c) 2017, Brad Rydzewski
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this
|
||||||
|
list of conditions and the following disclaimer.
|
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
|
||||||
|
* Neither the name of the copyright holder nor the names of its
|
||||||
|
contributors may be used to endorse or promote products derived from
|
||||||
|
this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
6
vendor/github.com/cncd/pubsub/README
generated
vendored
Normal file
6
vendor/github.com/cncd/pubsub/README
generated
vendored
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
Go package provides a common interface for publish-subscriber messaging.
|
||||||
|
|
||||||
|
Documentation:
|
||||||
|
|
||||||
|
http://godoc.org/github.com/cncd/pubsub
|
||||||
|
http://godoc.org/github.com/cncd/pubsub/gcp
|
75
vendor/github.com/cncd/pubsub/pub.go
generated
vendored
Normal file
75
vendor/github.com/cncd/pubsub/pub.go
generated
vendored
Normal file
|
@ -0,0 +1,75 @@
|
||||||
|
package pubsub
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"sync"
|
||||||
|
)
|
||||||
|
|
||||||
|
type subscriber struct {
|
||||||
|
receiver Receiver
|
||||||
|
}
|
||||||
|
|
||||||
|
type publisher struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
topics map[string]*topic
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates an in-memory publisher.
|
||||||
|
func New() Publisher {
|
||||||
|
return &publisher{
|
||||||
|
topics: make(map[string]*topic),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *publisher) Create(c context.Context, dest string) error {
|
||||||
|
p.Lock()
|
||||||
|
t, ok := p.topics[dest]
|
||||||
|
if !ok {
|
||||||
|
t = newTopic(dest)
|
||||||
|
p.topics[dest] = t
|
||||||
|
}
|
||||||
|
p.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *publisher) Publish(c context.Context, dest string, message Message) error {
|
||||||
|
p.Lock()
|
||||||
|
t, ok := p.topics[dest]
|
||||||
|
p.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
t.publish(message)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *publisher) Subscribe(c context.Context, dest string, receiver Receiver) error {
|
||||||
|
p.Lock()
|
||||||
|
t, ok := p.topics[dest]
|
||||||
|
p.Unlock()
|
||||||
|
if !ok {
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
s := &subscriber{
|
||||||
|
receiver: receiver,
|
||||||
|
}
|
||||||
|
t.subscribe(s)
|
||||||
|
select {
|
||||||
|
case <-c.Done():
|
||||||
|
case <-t.done:
|
||||||
|
}
|
||||||
|
t.unsubscribe(s)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *publisher) Remove(c context.Context, dest string) error {
|
||||||
|
p.Lock()
|
||||||
|
t, ok := p.topics[dest]
|
||||||
|
if ok {
|
||||||
|
delete(p.topics, dest)
|
||||||
|
t.close()
|
||||||
|
}
|
||||||
|
p.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
71
vendor/github.com/cncd/pubsub/pubsub.go
generated
vendored
Normal file
71
vendor/github.com/cncd/pubsub/pubsub.go
generated
vendored
Normal file
|
@ -0,0 +1,71 @@
|
||||||
|
// Package pubsub implements a publish-subscriber messaging system.
|
||||||
|
package pubsub
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"errors"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ErrNotFound is returned when the named topic does not exist.
|
||||||
|
var ErrNotFound = errors.New("topic not found")
|
||||||
|
|
||||||
|
// Message defines a published message.
|
||||||
|
type Message struct {
|
||||||
|
// ID identifies this message.
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
|
||||||
|
// Data is the actual data in the entry.
|
||||||
|
Data []byte `json:"data"`
|
||||||
|
|
||||||
|
// Labels represents the key-value pairs the entry is lebeled with.
|
||||||
|
Labels map[string]string `json:"labels,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Receiver receives published messages.
|
||||||
|
type Receiver func(Message)
|
||||||
|
|
||||||
|
// Publisher defines a mechanism for communicating messages from a group
|
||||||
|
// of senders, called producers, to a group of consumers.
|
||||||
|
type Publisher interface {
|
||||||
|
// Create creates the named topic.
|
||||||
|
Create(c context.Context, topic string) error
|
||||||
|
|
||||||
|
// Publish publishes the message.
|
||||||
|
Publish(c context.Context, topic string, message Message) error
|
||||||
|
|
||||||
|
// Subscribe subscribes to the topic. The Receiver function is a callback
|
||||||
|
// function that receives published messages.
|
||||||
|
Subscribe(c context.Context, topic string, receiver Receiver) error
|
||||||
|
|
||||||
|
// Remove removes the named topic.
|
||||||
|
Remove(c context.Context, topic string) error
|
||||||
|
}
|
||||||
|
|
||||||
|
// // global instance of the queue.
|
||||||
|
// var global = New()
|
||||||
|
//
|
||||||
|
// // Set sets the global queue.
|
||||||
|
// func Set(p Publisher) {
|
||||||
|
// global = p
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Create creates the named topic.
|
||||||
|
// func Create(c context.Context, topic string) error {
|
||||||
|
// return global.Create(c, topic)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Publish publishes the message.
|
||||||
|
// func Publish(c context.Context, topic string, message Message) error {
|
||||||
|
// return global.Publish(c, topic, message)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Subscribe subscribes to the topic. The Receiver function is a callback
|
||||||
|
// // function that receives published messages.
|
||||||
|
// func Subscribe(c context.Context, topic string, receiver Receiver) error {
|
||||||
|
// return global.Subscribe(c, topic, receiver)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Remove removes the topic.
|
||||||
|
// func Remove(c context.Context, topic string) error {
|
||||||
|
// return global.Remove(c, topic)
|
||||||
|
// }
|
45
vendor/github.com/cncd/pubsub/topic.go
generated
vendored
Normal file
45
vendor/github.com/cncd/pubsub/topic.go
generated
vendored
Normal file
|
@ -0,0 +1,45 @@
|
||||||
|
package pubsub
|
||||||
|
|
||||||
|
import "sync"
|
||||||
|
|
||||||
|
type topic struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
name string
|
||||||
|
done chan bool
|
||||||
|
subs map[*subscriber]struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
func newTopic(dest string) *topic {
|
||||||
|
return &topic{
|
||||||
|
name: dest,
|
||||||
|
done: make(chan bool),
|
||||||
|
subs: make(map[*subscriber]struct{}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *topic) subscribe(s *subscriber) {
|
||||||
|
t.Lock()
|
||||||
|
t.subs[s] = struct{}{}
|
||||||
|
t.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *topic) unsubscribe(s *subscriber) {
|
||||||
|
t.Lock()
|
||||||
|
delete(t.subs, s)
|
||||||
|
t.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *topic) publish(m Message) {
|
||||||
|
t.Lock()
|
||||||
|
for s := range t.subs {
|
||||||
|
go s.receiver(m)
|
||||||
|
}
|
||||||
|
t.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *topic) close() {
|
||||||
|
t.Lock()
|
||||||
|
close(t.done)
|
||||||
|
t.Unlock()
|
||||||
|
}
|
29
vendor/github.com/cncd/queue/LICENSE
generated
vendored
Normal file
29
vendor/github.com/cncd/queue/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
BSD 3-Clause License
|
||||||
|
|
||||||
|
Copyright (c) 2017, Brad Rydzewski
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this
|
||||||
|
list of conditions and the following disclaimer.
|
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
|
||||||
|
* Neither the name of the copyright holder nor the names of its
|
||||||
|
contributors may be used to endorse or promote products derived from
|
||||||
|
this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
6
vendor/github.com/cncd/queue/README
generated
vendored
Normal file
6
vendor/github.com/cncd/queue/README
generated
vendored
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
Go package provides a common interface for working with task queues.
|
||||||
|
|
||||||
|
Documentation:
|
||||||
|
|
||||||
|
http://godoc.org/github.com/cncd/queue
|
||||||
|
http://godoc.org/github.com/cncd/queue/gcp
|
190
vendor/github.com/cncd/queue/fifo.go
generated
vendored
Normal file
190
vendor/github.com/cncd/queue/fifo.go
generated
vendored
Normal file
|
@ -0,0 +1,190 @@
|
||||||
|
package queue
|
||||||
|
|
||||||
|
import (
|
||||||
|
"container/list"
|
||||||
|
"context"
|
||||||
|
"log"
|
||||||
|
"runtime"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
type entry struct {
|
||||||
|
item *Task
|
||||||
|
done chan bool
|
||||||
|
retry int
|
||||||
|
error error
|
||||||
|
deadline time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
type worker struct {
|
||||||
|
filter Filter
|
||||||
|
channel chan *Task
|
||||||
|
}
|
||||||
|
|
||||||
|
type fifo struct {
|
||||||
|
sync.Mutex
|
||||||
|
|
||||||
|
workers map[*worker]struct{}
|
||||||
|
running map[string]*entry
|
||||||
|
pending *list.List
|
||||||
|
extension time.Duration
|
||||||
|
}
|
||||||
|
|
||||||
|
// New returns a new fifo queue.
|
||||||
|
func New() Queue {
|
||||||
|
return &fifo{
|
||||||
|
workers: map[*worker]struct{}{},
|
||||||
|
running: map[string]*entry{},
|
||||||
|
pending: list.New(),
|
||||||
|
extension: time.Minute * 10,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push pushes an item to the tail of this queue.
|
||||||
|
func (q *fifo) Push(c context.Context, task *Task) error {
|
||||||
|
q.Lock()
|
||||||
|
q.pending.PushBack(task)
|
||||||
|
q.Unlock()
|
||||||
|
go q.process()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Poll retrieves and removes the head of this queue.
|
||||||
|
func (q *fifo) Poll(c context.Context, f Filter) (*Task, error) {
|
||||||
|
q.Lock()
|
||||||
|
w := &worker{
|
||||||
|
channel: make(chan *Task, 1),
|
||||||
|
filter: f,
|
||||||
|
}
|
||||||
|
q.workers[w] = struct{}{}
|
||||||
|
q.Unlock()
|
||||||
|
go q.process()
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-c.Done():
|
||||||
|
q.Lock()
|
||||||
|
delete(q.workers, w)
|
||||||
|
q.Unlock()
|
||||||
|
return nil, nil
|
||||||
|
case t := <-w.channel:
|
||||||
|
return t, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Done signals that the item is done executing.
|
||||||
|
func (q *fifo) Done(c context.Context, id string) error {
|
||||||
|
return q.Error(c, id, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Error signals that the item is done executing with error.
|
||||||
|
func (q *fifo) Error(c context.Context, id string, err error) error {
|
||||||
|
q.Lock()
|
||||||
|
state, ok := q.running[id]
|
||||||
|
if ok {
|
||||||
|
state.error = err
|
||||||
|
close(state.done)
|
||||||
|
delete(q.running, id)
|
||||||
|
}
|
||||||
|
q.Unlock()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Wait waits until the item is done executing.
|
||||||
|
func (q *fifo) Wait(c context.Context, id string) error {
|
||||||
|
q.Lock()
|
||||||
|
state := q.running[id]
|
||||||
|
q.Unlock()
|
||||||
|
if state != nil {
|
||||||
|
select {
|
||||||
|
case <-c.Done():
|
||||||
|
case <-state.done:
|
||||||
|
return state.error
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Extend extends the task execution deadline.
|
||||||
|
func (q *fifo) Extend(c context.Context, id string) error {
|
||||||
|
q.Lock()
|
||||||
|
defer q.Unlock()
|
||||||
|
|
||||||
|
state, ok := q.running[id]
|
||||||
|
if ok {
|
||||||
|
state.deadline = time.Now().Add(q.extension)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return ErrNotFound
|
||||||
|
}
|
||||||
|
|
||||||
|
// Info returns internal queue information.
|
||||||
|
func (q *fifo) Info(c context.Context) InfoT {
|
||||||
|
q.Lock()
|
||||||
|
stats := InfoT{}
|
||||||
|
stats.Stats.Workers = len(q.workers)
|
||||||
|
stats.Stats.Pending = q.pending.Len()
|
||||||
|
stats.Stats.Running = len(q.running)
|
||||||
|
|
||||||
|
for e := q.pending.Front(); e != nil; e = e.Next() {
|
||||||
|
stats.Pending = append(stats.Pending, e.Value.(*Task))
|
||||||
|
}
|
||||||
|
for _, entry := range q.running {
|
||||||
|
stats.Running = append(stats.Running, entry.item)
|
||||||
|
}
|
||||||
|
|
||||||
|
q.Unlock()
|
||||||
|
return stats
|
||||||
|
}
|
||||||
|
|
||||||
|
// helper function that loops through the queue and attempts to
|
||||||
|
// match the item to a single subscriber.
|
||||||
|
func (q *fifo) process() {
|
||||||
|
defer func() {
|
||||||
|
// the risk of panic is low. This code can probably be removed
|
||||||
|
// once the code has been used in real world installs without issue.
|
||||||
|
if err := recover(); err != nil {
|
||||||
|
const size = 64 << 10
|
||||||
|
buf := make([]byte, size)
|
||||||
|
buf = buf[:runtime.Stack(buf, false)]
|
||||||
|
log.Printf("queue: unexpected panic: %v\n%s", err, buf)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
q.Lock()
|
||||||
|
defer q.Unlock()
|
||||||
|
|
||||||
|
// TODO(bradrydzewski) move this to a helper function
|
||||||
|
// push items to the front of the queue if the item expires.
|
||||||
|
for id, state := range q.running {
|
||||||
|
if time.Now().After(state.deadline) {
|
||||||
|
q.pending.PushFront(state.item)
|
||||||
|
delete(q.running, id)
|
||||||
|
close(state.done)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var next *list.Element
|
||||||
|
loop:
|
||||||
|
for e := q.pending.Front(); e != nil; e = next {
|
||||||
|
next = e.Next()
|
||||||
|
item := e.Value.(*Task)
|
||||||
|
for w := range q.workers {
|
||||||
|
if w.filter(item) {
|
||||||
|
delete(q.workers, w)
|
||||||
|
q.pending.Remove(e)
|
||||||
|
|
||||||
|
q.running[item.ID] = &entry{
|
||||||
|
item: item,
|
||||||
|
done: make(chan bool),
|
||||||
|
deadline: time.Now().Add(q.extension),
|
||||||
|
}
|
||||||
|
|
||||||
|
w.channel <- item
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
110
vendor/github.com/cncd/queue/queue.go
generated
vendored
Normal file
110
vendor/github.com/cncd/queue/queue.go
generated
vendored
Normal file
|
@ -0,0 +1,110 @@
|
||||||
|
package queue
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"errors"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrCancel indicates the task was cancelled.
|
||||||
|
ErrCancel = errors.New("queue: task cancelled")
|
||||||
|
|
||||||
|
// ErrNotFound indicates the task was not found in the queue.
|
||||||
|
ErrNotFound = errors.New("queue: task not found")
|
||||||
|
)
|
||||||
|
|
||||||
|
// Task defines a unit of work in the queue.
|
||||||
|
type Task struct {
|
||||||
|
// ID identifies this task.
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
|
||||||
|
// Data is the actual data in the entry.
|
||||||
|
Data []byte `json:"data"`
|
||||||
|
|
||||||
|
// Labels represents the key-value pairs the entry is lebeled with.
|
||||||
|
Labels map[string]string `json:"labels,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// InfoT provides runtime information.
|
||||||
|
type InfoT struct {
|
||||||
|
Pending []*Task `json:"pending"`
|
||||||
|
Running []*Task `json:"running"`
|
||||||
|
Stats struct {
|
||||||
|
Workers int `json:"worker_count"`
|
||||||
|
Pending int `json:"pending_count"`
|
||||||
|
Running int `json:"running_count"`
|
||||||
|
Complete int `json:"completed_count"`
|
||||||
|
} `json:"stats"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Filter filters tasks in the queue. If the Filter returns false,
|
||||||
|
// the Task is skipped and not returned to the subscriber.
|
||||||
|
type Filter func(*Task) bool
|
||||||
|
|
||||||
|
// Queue defines a task queue for scheduling tasks among
|
||||||
|
// a pool of workers.
|
||||||
|
type Queue interface {
|
||||||
|
// Push pushes an task to the tail of this queue.
|
||||||
|
Push(c context.Context, task *Task) error
|
||||||
|
|
||||||
|
// Poll retrieves and removes a task head of this queue.
|
||||||
|
Poll(c context.Context, f Filter) (*Task, error)
|
||||||
|
|
||||||
|
// Extend extends the deadline for a task.
|
||||||
|
Extend(c context.Context, id string) error
|
||||||
|
|
||||||
|
// Done signals the task is complete.
|
||||||
|
Done(c context.Context, id string) error
|
||||||
|
|
||||||
|
// Error signals the task is complete with errors.
|
||||||
|
Error(c context.Context, id string, err error) error
|
||||||
|
|
||||||
|
// Wait waits until the task is complete.
|
||||||
|
Wait(c context.Context, id string) error
|
||||||
|
|
||||||
|
// Info returns internal queue information.
|
||||||
|
Info(c context.Context) InfoT
|
||||||
|
}
|
||||||
|
|
||||||
|
// // global instance of the queue.
|
||||||
|
// var global = New()
|
||||||
|
//
|
||||||
|
// // Set sets the global queue.
|
||||||
|
// func Set(queue Queue) {
|
||||||
|
// global = queue
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Push pushes an task to the tail of the global queue.
|
||||||
|
// func Push(c context.Context, task *Task) error {
|
||||||
|
// return global.Push(c, task)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Poll retrieves and removes a task head of the global queue.
|
||||||
|
// func Poll(c context.Context, f Filter) (*Task, error) {
|
||||||
|
// return global.Poll(c, f)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Extend extends the deadline for a task.
|
||||||
|
// func Extend(c context.Context, id string) error {
|
||||||
|
// return global.Extend(c, id)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Done signals the task is complete.
|
||||||
|
// func Done(c context.Context, id string) error {
|
||||||
|
// return global.Done(c, id)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Error signals the task is complete with errors.
|
||||||
|
// func Error(c context.Context, id string, err error) {
|
||||||
|
// global.Error(c, id, err)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Wait waits until the task is complete.
|
||||||
|
// func Wait(c context.Context, id string) error {
|
||||||
|
// return global.Wait(c, id)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Info returns internal queue information.
|
||||||
|
// func Info(c context.Context) InfoT {
|
||||||
|
// return global.Info(c)
|
||||||
|
// }
|
1
vendor/github.com/cncd/queue/worker.go
generated
vendored
Normal file
1
vendor/github.com/cncd/queue/worker.go
generated
vendored
Normal file
|
@ -0,0 +1 @@
|
||||||
|
package queue
|
86
vendor/vendor.json
vendored
86
vendor/vendor.json
vendored
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"comment": "",
|
"comment": "",
|
||||||
"ignore": "test",
|
"ignore": "test github.com/drone/mq/ github.com/tidwall/redlog/ google.golang.org/appengine/ github.com/syndtr/goleveldb/ github.com/drone/drone-ui/",
|
||||||
"package": [
|
"package": [
|
||||||
{
|
{
|
||||||
"path": "code.google.com/p/go.crypto/ssh",
|
"path": "code.google.com/p/go.crypto/ssh",
|
||||||
|
@ -17,6 +17,90 @@
|
||||||
"revision": "5cef21e2e4f0fd147973b558d4db7395176bcd95",
|
"revision": "5cef21e2e4f0fd147973b558d4db7395176bcd95",
|
||||||
"revisionTime": "2016-03-22T13:50:45-07:00"
|
"revisionTime": "2016-03-22T13:50:45-07:00"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "HhHyclewPAZ7sZQCsl2nMSe1T5s=",
|
||||||
|
"path": "github.com/cncd/logging",
|
||||||
|
"revision": "03b6463409fecbd23f04587adf9bc71da13796a2",
|
||||||
|
"revisionTime": "2017-03-05T07:05:34Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "W3AuK8ocqHwlUajGmQLFvnRhTZE=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "PSzh0ix/rlMrS/Cl3aH6GHGrJuo=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/backend",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "0CGXRaYwZhJxGIrGhn8WGpkFqPo=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/backend/docker",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "uUagpzha5ah/a3RO6IImvzHYFlY=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/frontend",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "O0sulBQAHJeNLg3lO38Cq5uf/eg=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "e1lZWQdObXCKWqZOGlOeaeERQMc=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/compiler",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "Q0GkNUFamVYIA1Fd8r0A5M6Gx54=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/linter",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "kx2sPUIMozPC/g6E4w48h3FfH3k=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/matrix",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "2/3f3oNmxXy5kcrRLCFa24Oc9O4=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/interrupt",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "8eTwXZPM/Kp9uE/mnhpWDTiX7nY=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/multipart",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "5axmtZsHaQ5uE/tuNQZygquNx8U=",
|
||||||
|
"path": "github.com/cncd/pipeline/pipeline/rpc",
|
||||||
|
"revision": "b5457789534af0da2a05bd9fd0c0afe6fea391c6",
|
||||||
|
"revisionTime": "2017-03-04T04:47:59Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "7Qj1DK0ceAXkYztW0l3+L6sn+V8=",
|
||||||
|
"path": "github.com/cncd/pubsub",
|
||||||
|
"revision": "0691529ab100a0f3c4e2087407d2788a0787ad9c",
|
||||||
|
"revisionTime": "2017-03-03T07:06:35Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "AG4M07wOZNTnSFHJIfdXT2ymnts=",
|
||||||
|
"path": "github.com/cncd/queue",
|
||||||
|
"revision": "1ce1ada7160f1eda015a16c1b7f9ea497fa36873",
|
||||||
|
"revisionTime": "2017-03-03T07:04:55Z"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"path": "github.com/codegangsta/cli",
|
"path": "github.com/codegangsta/cli",
|
||||||
"revision": "70e3fa51ebed95df8c0fbe1519c1c1f9bc16bb13",
|
"revision": "70e3fa51ebed95df8c0fbe1519c1c1f9bc16bb13",
|
||||||
|
|
Loading…
Reference in a new issue