mirror of
https://github.com/woodpecker-ci/woodpecker.git
synced 2024-11-30 05:41:12 +00:00
enable restart
This commit is contained in:
parent
39bb69a27d
commit
395f0d52f9
6 changed files with 886 additions and 236 deletions
|
@ -113,7 +113,7 @@ func run(ctx context.Context, client rpc.Peer, filter rpc.Filter) error {
|
||||||
|
|
||||||
cancelled := abool.New()
|
cancelled := abool.New()
|
||||||
go func() {
|
go func() {
|
||||||
if werr := client.Wait(ctx, work.ID); err != nil {
|
if werr := client.Wait(ctx, work.ID); werr != nil {
|
||||||
cancelled.SetTo(true)
|
cancelled.SetTo(true)
|
||||||
log.Printf("pipeline: cancel signal received: %s: %s", work.ID, werr)
|
log.Printf("pipeline: cancel signal received: %s: %s", work.ID, werr)
|
||||||
cancel()
|
cancel()
|
||||||
|
|
221
server/build.go
221
server/build.go
|
@ -3,6 +3,7 @@ package server
|
||||||
import (
|
import (
|
||||||
"bufio"
|
"bufio"
|
||||||
"context"
|
"context"
|
||||||
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
@ -11,6 +12,8 @@ import (
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
log "github.com/Sirupsen/logrus"
|
log "github.com/Sirupsen/logrus"
|
||||||
|
"github.com/cncd/pipeline/pipeline/rpc"
|
||||||
|
"github.com/cncd/pubsub"
|
||||||
"github.com/cncd/queue"
|
"github.com/cncd/queue"
|
||||||
"github.com/drone/drone/remote"
|
"github.com/drone/drone/remote"
|
||||||
"github.com/drone/drone/shared/httputil"
|
"github.com/drone/drone/shared/httputil"
|
||||||
|
@ -169,6 +172,11 @@ func DeleteBuild(c *gin.Context) {
|
||||||
|
|
||||||
func PostBuild(c *gin.Context) {
|
func PostBuild(c *gin.Context) {
|
||||||
|
|
||||||
|
if os.Getenv("DRONE_CANARY") == "true" {
|
||||||
|
PostBuild2(c)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
remote_ := remote.FromContext(c)
|
remote_ := remote.FromContext(c)
|
||||||
repo := session.Repo(c)
|
repo := session.Repo(c)
|
||||||
fork := c.DefaultQuery("fork", "false")
|
fork := c.DefaultQuery("fork", "false")
|
||||||
|
@ -282,6 +290,7 @@ func PostBuild(c *gin.Context) {
|
||||||
build.Started = 0
|
build.Started = 0
|
||||||
build.Finished = 0
|
build.Finished = 0
|
||||||
build.Enqueued = time.Now().UTC().Unix()
|
build.Enqueued = time.Now().UTC().Unix()
|
||||||
|
build.Error = ""
|
||||||
for _, job := range jobs {
|
for _, job := range jobs {
|
||||||
for k, v := range buildParams {
|
for k, v := range buildParams {
|
||||||
job.Environment[k] = v
|
job.Environment[k] = v
|
||||||
|
@ -395,3 +404,215 @@ func copyLogs(dest io.Writer, src io.Reader) error {
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
func PostBuild2(c *gin.Context) {
|
||||||
|
|
||||||
|
remote_ := remote.FromContext(c)
|
||||||
|
repo := session.Repo(c)
|
||||||
|
fork := c.DefaultQuery("fork", "false")
|
||||||
|
|
||||||
|
num, err := strconv.Atoi(c.Param("number"))
|
||||||
|
if err != nil {
|
||||||
|
c.AbortWithError(http.StatusBadRequest, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
user, err := store.GetUser(c, repo.UserID)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to find repo owner %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(500, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
build, err := store.GetBuildNumber(c, repo, num)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to get build %d. %s", num, err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// if the remote has a refresh token, the current access token
|
||||||
|
// may be stale. Therefore, we should refresh prior to dispatching
|
||||||
|
// the job.
|
||||||
|
if refresher, ok := remote_.(remote.Refresher); ok {
|
||||||
|
ok, _ := refresher.Refresh(user)
|
||||||
|
if ok {
|
||||||
|
store.UpdateUser(c, user)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// fetch the .drone.yml file from the database
|
||||||
|
cfg := ToConfig(c)
|
||||||
|
raw, err := remote_.File(user, repo, build, cfg.Yaml)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to get build config for %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
netrc, err := remote_.Netrc(user, repo)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to generate netrc for %s. %s", repo.FullName, err)
|
||||||
|
c.AbortWithError(500, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
jobs, err := store.GetJobList(c, build)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("failure to get build %d jobs. %s", build.Number, err)
|
||||||
|
c.AbortWithError(404, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// must not restart a running build
|
||||||
|
if build.Status == model.StatusPending || build.Status == model.StatusRunning {
|
||||||
|
c.String(409, "Cannot re-start a started build")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// forking the build creates a duplicate of the build
|
||||||
|
// and then executes. This retains prior build history.
|
||||||
|
if forkit, _ := strconv.ParseBool(fork); forkit {
|
||||||
|
build.ID = 0
|
||||||
|
build.Number = 0
|
||||||
|
build.Parent = num
|
||||||
|
for _, job := range jobs {
|
||||||
|
job.ID = 0
|
||||||
|
job.NodeID = 0
|
||||||
|
}
|
||||||
|
err := store.CreateBuild(c, build, jobs...)
|
||||||
|
if err != nil {
|
||||||
|
c.String(500, err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
event := c.DefaultQuery("event", build.Event)
|
||||||
|
if event == model.EventPush ||
|
||||||
|
event == model.EventPull ||
|
||||||
|
event == model.EventTag ||
|
||||||
|
event == model.EventDeploy {
|
||||||
|
build.Event = event
|
||||||
|
}
|
||||||
|
build.Deploy = c.DefaultQuery("deploy_to", build.Deploy)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read query string parameters into buildParams, exclude reserved params
|
||||||
|
var buildParams = map[string]string{}
|
||||||
|
for key, val := range c.Request.URL.Query() {
|
||||||
|
switch key {
|
||||||
|
case "fork", "event", "deploy_to":
|
||||||
|
default:
|
||||||
|
// We only accept string literals, because build parameters will be
|
||||||
|
// injected as environment variables
|
||||||
|
buildParams[key] = val[0]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// todo move this to database tier
|
||||||
|
// and wrap inside a transaction
|
||||||
|
build.Status = model.StatusPending
|
||||||
|
build.Started = 0
|
||||||
|
build.Finished = 0
|
||||||
|
build.Enqueued = time.Now().UTC().Unix()
|
||||||
|
build.Error = ""
|
||||||
|
for _, job := range jobs {
|
||||||
|
for k, v := range buildParams {
|
||||||
|
job.Environment[k] = v
|
||||||
|
}
|
||||||
|
job.Error = ""
|
||||||
|
job.Status = model.StatusPending
|
||||||
|
job.Started = 0
|
||||||
|
job.Finished = 0
|
||||||
|
job.ExitCode = 0
|
||||||
|
job.NodeID = 0
|
||||||
|
job.Enqueued = build.Enqueued
|
||||||
|
store.UpdateJob(c, job)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = store.UpdateBuild(c, build)
|
||||||
|
if err != nil {
|
||||||
|
c.AbortWithStatus(500)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
c.JSON(202, build)
|
||||||
|
|
||||||
|
// get the previous build so that we can send
|
||||||
|
// on status change notifications
|
||||||
|
last, _ := store.GetBuildLastBefore(c, repo, build.Branch, build.ID)
|
||||||
|
secs, err := store.GetMergedSecretList(c, repo)
|
||||||
|
if err != nil {
|
||||||
|
log.Debugf("Error getting secrets for %s#%d. %s", repo.FullName, build.Number, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
b := builder{
|
||||||
|
Repo: repo,
|
||||||
|
Curr: build,
|
||||||
|
Last: last,
|
||||||
|
Netrc: netrc,
|
||||||
|
Secs: secs,
|
||||||
|
Link: httputil.GetURL(c.Request),
|
||||||
|
Yaml: string(raw),
|
||||||
|
}
|
||||||
|
items, err := b.Build()
|
||||||
|
if err != nil {
|
||||||
|
build.Status = model.StatusError
|
||||||
|
build.Started = time.Now().Unix()
|
||||||
|
build.Finished = build.Started
|
||||||
|
build.Error = err.Error()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, item := range items {
|
||||||
|
// TODO prevent possible index out of bounds
|
||||||
|
item.Job.ID = jobs[i].ID
|
||||||
|
build.Jobs = append(build.Jobs, item.Job)
|
||||||
|
store.UpdateJob(c, item.Job)
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
// publish topic
|
||||||
|
//
|
||||||
|
message := pubsub.Message{
|
||||||
|
Labels: map[string]string{
|
||||||
|
"repo": repo.FullName,
|
||||||
|
"private": strconv.FormatBool(repo.IsPrivate),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
message.Data, _ = json.Marshal(model.Event{
|
||||||
|
Type: model.Enqueued,
|
||||||
|
Repo: *repo,
|
||||||
|
Build: *build,
|
||||||
|
})
|
||||||
|
// TODO remove global reference
|
||||||
|
config.pubsub.Publish(c, "topic/events", message)
|
||||||
|
//
|
||||||
|
// end publish topic
|
||||||
|
//
|
||||||
|
|
||||||
|
for _, item := range items {
|
||||||
|
task := new(queue.Task)
|
||||||
|
task.ID = fmt.Sprint(item.Job.ID)
|
||||||
|
task.Labels = map[string]string{}
|
||||||
|
task.Labels["platform"] = item.Platform
|
||||||
|
for k, v := range item.Labels {
|
||||||
|
task.Labels[k] = v
|
||||||
|
}
|
||||||
|
|
||||||
|
task.Data, _ = json.Marshal(rpc.Pipeline{
|
||||||
|
ID: fmt.Sprint(item.Job.ID),
|
||||||
|
Config: item.Config,
|
||||||
|
Timeout: b.Repo.Timeout,
|
||||||
|
})
|
||||||
|
|
||||||
|
config.logger.Open(context.Background(), task.ID)
|
||||||
|
config.queue.Push(context.Background(), task)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
849
server/hook2.go
849
server/hook2.go
|
@ -45,6 +45,600 @@ func GetQueueInfo(c *gin.Context) {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// func PostHookOld(c *gin.Context) {
|
||||||
|
// remote_ := remote.FromContext(c)
|
||||||
|
//
|
||||||
|
// tmprepo, build, err := remote_.Hook(c.Request)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to parse hook. %s", err)
|
||||||
|
// c.AbortWithError(400, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// if build == nil {
|
||||||
|
// c.Writer.WriteHeader(200)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// if tmprepo == nil {
|
||||||
|
// logrus.Errorf("failure to ascertain repo from hook.")
|
||||||
|
// c.Writer.WriteHeader(400)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // skip the build if any case-insensitive combination of the words "skip" and "ci"
|
||||||
|
// // wrapped in square brackets appear in the commit message
|
||||||
|
// skipMatch := skipRe.FindString(build.Message)
|
||||||
|
// if len(skipMatch) > 0 {
|
||||||
|
// logrus.Infof("ignoring hook. %s found in %s", skipMatch, build.Commit)
|
||||||
|
// c.Writer.WriteHeader(204)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// repo, err := store.GetRepoOwnerName(c, tmprepo.Owner, tmprepo.Name)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to find repo %s/%s from hook. %s", tmprepo.Owner, tmprepo.Name, err)
|
||||||
|
// c.AbortWithError(404, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // get the token and verify the hook is authorized
|
||||||
|
// parsed, err := token.ParseRequest(c.Request, func(t *token.Token) (string, error) {
|
||||||
|
// return repo.Hash, nil
|
||||||
|
// })
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to parse token from hook for %s. %s", repo.FullName, err)
|
||||||
|
// c.AbortWithError(400, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// if parsed.Text != repo.FullName {
|
||||||
|
// logrus.Errorf("failure to verify token from hook. Expected %s, got %s", repo.FullName, parsed.Text)
|
||||||
|
// c.AbortWithStatus(403)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// if repo.UserID == 0 {
|
||||||
|
// logrus.Warnf("ignoring hook. repo %s has no owner.", repo.FullName)
|
||||||
|
// c.Writer.WriteHeader(204)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// var skipped = true
|
||||||
|
// if (build.Event == model.EventPush && repo.AllowPush) ||
|
||||||
|
// (build.Event == model.EventPull && repo.AllowPull) ||
|
||||||
|
// (build.Event == model.EventDeploy && repo.AllowDeploy) ||
|
||||||
|
// (build.Event == model.EventTag && repo.AllowTag) {
|
||||||
|
// skipped = false
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// if skipped {
|
||||||
|
// logrus.Infof("ignoring hook. repo %s is disabled for %s events.", repo.FullName, build.Event)
|
||||||
|
// c.Writer.WriteHeader(204)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// user, err := store.GetUser(c, repo.UserID)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to find repo owner %s. %s", repo.FullName, err)
|
||||||
|
// c.AbortWithError(500, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // if the remote has a refresh token, the current access token
|
||||||
|
// // may be stale. Therefore, we should refresh prior to dispatching
|
||||||
|
// // the job.
|
||||||
|
// if refresher, ok := remote_.(remote.Refresher); ok {
|
||||||
|
// ok, _ := refresher.Refresh(user)
|
||||||
|
// if ok {
|
||||||
|
// store.UpdateUser(c, user)
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // fetch the build file from the database
|
||||||
|
// cfg := ToConfig(c)
|
||||||
|
// raw, err := remote_.File(user, repo, build, cfg.Yaml)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to get build config for %s. %s", repo.FullName, err)
|
||||||
|
// c.AbortWithError(404, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// sec, err := remote_.File(user, repo, build, cfg.Shasum)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Debugf("cannot find yaml signature for %s. %s", repo.FullName, err)
|
||||||
|
// // NOTE we don't exit on failure. The sec file is optional
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// axes, err := matrix.Parse(raw)
|
||||||
|
// if err != nil {
|
||||||
|
// c.String(500, "Failed to parse yaml file or calculate matrix. %s", err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// if len(axes) == 0 {
|
||||||
|
// axes = append(axes, matrix.Axis{})
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// netrc, err := remote_.Netrc(user, repo)
|
||||||
|
// if err != nil {
|
||||||
|
// c.String(500, "Failed to generate netrc file. %s", err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // verify the branches can be built vs skipped
|
||||||
|
// branches, err := yaml.ParseBytes(raw)
|
||||||
|
// if err != nil {
|
||||||
|
// c.String(500, "Failed to parse yaml file. %s", err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
// if !branches.Branches.Match(build.Branch) && build.Event != model.EventTag && build.Event != model.EventDeploy {
|
||||||
|
// c.String(200, "Branch does not match restrictions defined in yaml")
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// signature, err := jose.ParseSigned(string(sec))
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Debugf("cannot parse .drone.yml.sig file. %s", err)
|
||||||
|
// } else if len(sec) == 0 {
|
||||||
|
// logrus.Debugf("cannot parse .drone.yml.sig file. empty file")
|
||||||
|
// } else {
|
||||||
|
// build.Signed = true
|
||||||
|
// output, verr := signature.Verify([]byte(repo.Hash))
|
||||||
|
// if verr != nil {
|
||||||
|
// logrus.Debugf("cannot verify .drone.yml.sig file. %s", verr)
|
||||||
|
// } else if string(output) != string(raw) {
|
||||||
|
// logrus.Debugf("cannot verify .drone.yml.sig file. no match")
|
||||||
|
// } else {
|
||||||
|
// build.Verified = true
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // update some build fields
|
||||||
|
// build.Status = model.StatusPending
|
||||||
|
// build.RepoID = repo.ID
|
||||||
|
//
|
||||||
|
// // and use a transaction
|
||||||
|
// var jobs []*model.Job
|
||||||
|
// for num, axis := range axes {
|
||||||
|
// jobs = append(jobs, &model.Job{
|
||||||
|
// BuildID: build.ID,
|
||||||
|
// Number: num + 1,
|
||||||
|
// Status: model.StatusPending,
|
||||||
|
// Environment: axis,
|
||||||
|
// })
|
||||||
|
// }
|
||||||
|
// err = store.CreateBuild(c, build, jobs...)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("failure to save commit for %s. %s", repo.FullName, err)
|
||||||
|
// c.AbortWithError(500, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// c.JSON(200, build)
|
||||||
|
//
|
||||||
|
// uri := fmt.Sprintf("%s/%s/%d", httputil.GetURL(c.Request), repo.FullName, build.Number)
|
||||||
|
// err = remote_.Status(user, repo, build, uri)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Errorf("error setting commit status for %s/%d", repo.FullName, build.Number)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // get the previous build so that we can send
|
||||||
|
// // on status change notifications
|
||||||
|
// last, _ := store.GetBuildLastBefore(c, repo, build.Branch, build.ID)
|
||||||
|
// secs, err := store.GetMergedSecretList(c, repo)
|
||||||
|
// if err != nil {
|
||||||
|
// logrus.Debugf("Error getting secrets for %s#%d. %s", repo.FullName, build.Number, err)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// //
|
||||||
|
// // BELOW: NEW
|
||||||
|
// //
|
||||||
|
//
|
||||||
|
// b := builder{
|
||||||
|
// Repo: repo,
|
||||||
|
// Curr: build,
|
||||||
|
// Last: last,
|
||||||
|
// Netrc: netrc,
|
||||||
|
// Secs: secs,
|
||||||
|
// Link: httputil.GetURL(c.Request),
|
||||||
|
// Yaml: string(raw),
|
||||||
|
// }
|
||||||
|
// items, err := b.Build()
|
||||||
|
// if err != nil {
|
||||||
|
// build.Status = model.StatusError
|
||||||
|
// build.Started = time.Now().Unix()
|
||||||
|
// build.Finished = build.Started
|
||||||
|
// build.Error = err.Error()
|
||||||
|
// store.CreateBuild(c, build, build.Jobs...)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// for _, item := range items {
|
||||||
|
// build.Jobs = append(build.Jobs, item.Job)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// if err := store.CreateBuild(c, build, build.Jobs...); err != nil {
|
||||||
|
// logrus.Errorf("failure to save commit for %s. %s", repo.FullName, err)
|
||||||
|
// c.AbortWithError(500, err)
|
||||||
|
// return
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// for _, item := range items {
|
||||||
|
//
|
||||||
|
// task := new(queue.Task)
|
||||||
|
// task.ID = fmt.Sprint(item.Job.ID)
|
||||||
|
// task.Labels = map[string]string{}
|
||||||
|
// task.Labels["platform"] = item.Platform
|
||||||
|
// for k, v := range item.Labels {
|
||||||
|
// task.Labels[k] = v
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// task.Data, _ = json.Marshal(rpc.Pipeline{
|
||||||
|
// ID: fmt.Sprint(item.Job.ID),
|
||||||
|
// Config: item.Config,
|
||||||
|
// Timeout: b.Repo.Timeout,
|
||||||
|
// })
|
||||||
|
//
|
||||||
|
// config.logger.Open(context.Background(), task.ID)
|
||||||
|
// config.queue.Push(context.Background(), task)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// //
|
||||||
|
// // new code here
|
||||||
|
// //
|
||||||
|
//
|
||||||
|
// message := pubsub.Message{
|
||||||
|
// Labels: map[string]string{
|
||||||
|
// "repo": repo.FullName,
|
||||||
|
// "private": strconv.FormatBool(repo.IsPrivate),
|
||||||
|
// },
|
||||||
|
// }
|
||||||
|
// message.Data, _ = json.Marshal(model.Event{
|
||||||
|
// Type: model.Enqueued,
|
||||||
|
// Repo: *repo,
|
||||||
|
// Build: *build,
|
||||||
|
// })
|
||||||
|
// // TODO remove global reference
|
||||||
|
// config.pubsub.Publish(c, "topic/events", message)
|
||||||
|
//
|
||||||
|
// //
|
||||||
|
// // workspace
|
||||||
|
// //
|
||||||
|
//
|
||||||
|
// for _, job := range jobs {
|
||||||
|
//
|
||||||
|
// metadata := metadataFromStruct(repo, build, last, job, httputil.GetURL(c.Request))
|
||||||
|
// environ := metadata.Environ()
|
||||||
|
//
|
||||||
|
// secrets := map[string]string{}
|
||||||
|
// for _, sec := range secs {
|
||||||
|
// if !sec.MatchEvent(build.Event) {
|
||||||
|
// continue
|
||||||
|
// }
|
||||||
|
// if build.Verified || sec.SkipVerify {
|
||||||
|
// secrets[sec.Name] = sec.Value
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// sub := func(name string) string {
|
||||||
|
// if v, ok := environ[name]; ok {
|
||||||
|
// return v
|
||||||
|
// }
|
||||||
|
// return secrets[name]
|
||||||
|
// }
|
||||||
|
// if s, err := envsubst.Eval(string(raw), sub); err != nil {
|
||||||
|
// raw = []byte(s)
|
||||||
|
// }
|
||||||
|
// parsed, err := yaml.ParseBytes(raw)
|
||||||
|
// if err != nil {
|
||||||
|
// job.ExitCode = 255
|
||||||
|
// job.Enqueued = time.Now().Unix()
|
||||||
|
// job.Started = time.Now().Unix()
|
||||||
|
// job.Finished = time.Now().Unix()
|
||||||
|
// job.Error = err.Error()
|
||||||
|
// store.UpdateBuildJob(c, build, job)
|
||||||
|
// continue
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// lerr := linter.New(
|
||||||
|
// linter.WithTrusted(repo.IsTrusted),
|
||||||
|
// ).Lint(parsed)
|
||||||
|
// if lerr != nil {
|
||||||
|
// job.ExitCode = 255
|
||||||
|
// job.Enqueued = time.Now().Unix()
|
||||||
|
// job.Started = time.Now().Unix()
|
||||||
|
// job.Finished = time.Now().Unix()
|
||||||
|
// job.Error = lerr.Error()
|
||||||
|
// store.UpdateBuildJob(c, build, job)
|
||||||
|
// continue
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// ir := compiler.New(
|
||||||
|
// compiler.WithEnviron(environ),
|
||||||
|
// // TODO ability to customize the escalated plugins
|
||||||
|
// compiler.WithEscalated("plugins/docker", "plugins/gcr", "plugins/ecr"),
|
||||||
|
// compiler.WithLocal(false),
|
||||||
|
// compiler.WithNetrc(netrc.Login, netrc.Password, netrc.Machine),
|
||||||
|
// compiler.WithPrefix(
|
||||||
|
// fmt.Sprintf(
|
||||||
|
// "%d_%d",
|
||||||
|
// job.ID,
|
||||||
|
// time.Now().Unix(),
|
||||||
|
// ),
|
||||||
|
// ),
|
||||||
|
// compiler.WithEnviron(job.Environment),
|
||||||
|
// compiler.WithProxy(),
|
||||||
|
// // TODO ability to set global volumes for things like certs
|
||||||
|
// compiler.WithVolumes(),
|
||||||
|
// compiler.WithWorkspaceFromURL("/drone", repo.Link),
|
||||||
|
// ).Compile(parsed)
|
||||||
|
//
|
||||||
|
// // TODO there is a chicken and egg problem here because
|
||||||
|
// // the compiled yaml has a platform environment variable
|
||||||
|
// // that is not correctly set, because we are just about
|
||||||
|
// // to set it ....
|
||||||
|
// // TODO maybe we remove platform from metadata and let
|
||||||
|
// // the compiler set the value from the yaml itself.
|
||||||
|
// if parsed.Platform == "" {
|
||||||
|
// parsed.Platform = "linux/amd64"
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// for _, sec := range secs {
|
||||||
|
// if !sec.MatchEvent(build.Event) {
|
||||||
|
// continue
|
||||||
|
// }
|
||||||
|
// if build.Verified || sec.SkipVerify {
|
||||||
|
// ir.Secrets = append(ir.Secrets, &backend.Secret{
|
||||||
|
// Mask: sec.Conceal,
|
||||||
|
// Name: sec.Name,
|
||||||
|
// Value: sec.Value,
|
||||||
|
// })
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// task := new(queue.Task)
|
||||||
|
// task.ID = fmt.Sprint(job.ID)
|
||||||
|
// task.Labels = map[string]string{}
|
||||||
|
// task.Labels["platform"] = parsed.Platform
|
||||||
|
// if parsed.Labels != nil {
|
||||||
|
// for k, v := range parsed.Labels {
|
||||||
|
// task.Labels[k] = v
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// task.Data, _ = json.Marshal(rpc.Pipeline{
|
||||||
|
// ID: fmt.Sprint(job.ID),
|
||||||
|
// Config: ir,
|
||||||
|
// Timeout: repo.Timeout,
|
||||||
|
// })
|
||||||
|
//
|
||||||
|
// config.logger.Open(context.Background(), task.ID)
|
||||||
|
// config.queue.Push(context.Background(), task)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// }
|
||||||
|
|
||||||
|
// return the metadata from the cli context.
|
||||||
|
func metadataFromStruct(repo *model.Repo, build, last *model.Build, job *model.Job, link string) frontend.Metadata {
|
||||||
|
return frontend.Metadata{
|
||||||
|
Repo: frontend.Repo{
|
||||||
|
Name: repo.Name,
|
||||||
|
Link: repo.Link,
|
||||||
|
Remote: repo.Clone,
|
||||||
|
Private: repo.IsPrivate,
|
||||||
|
},
|
||||||
|
Curr: frontend.Build{
|
||||||
|
Number: build.Number,
|
||||||
|
Created: build.Created,
|
||||||
|
Started: build.Started,
|
||||||
|
Finished: build.Finished,
|
||||||
|
Status: build.Status,
|
||||||
|
Event: build.Event,
|
||||||
|
Link: build.Link,
|
||||||
|
Target: build.Deploy,
|
||||||
|
Commit: frontend.Commit{
|
||||||
|
Sha: build.Commit,
|
||||||
|
Ref: build.Ref,
|
||||||
|
Refspec: build.Refspec,
|
||||||
|
Branch: build.Branch,
|
||||||
|
Message: build.Message,
|
||||||
|
Author: frontend.Author{
|
||||||
|
Name: build.Author,
|
||||||
|
Email: build.Email,
|
||||||
|
Avatar: build.Avatar,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Prev: frontend.Build{
|
||||||
|
Number: last.Number,
|
||||||
|
Created: last.Created,
|
||||||
|
Started: last.Started,
|
||||||
|
Finished: last.Finished,
|
||||||
|
Status: last.Status,
|
||||||
|
Event: last.Event,
|
||||||
|
Link: last.Link,
|
||||||
|
Target: last.Deploy,
|
||||||
|
Commit: frontend.Commit{
|
||||||
|
Sha: last.Commit,
|
||||||
|
Ref: last.Ref,
|
||||||
|
Refspec: last.Refspec,
|
||||||
|
Branch: last.Branch,
|
||||||
|
Message: last.Message,
|
||||||
|
Author: frontend.Author{
|
||||||
|
Name: last.Author,
|
||||||
|
Email: last.Email,
|
||||||
|
Avatar: last.Avatar,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Job: frontend.Job{
|
||||||
|
Number: job.Number,
|
||||||
|
Matrix: job.Environment,
|
||||||
|
},
|
||||||
|
Sys: frontend.System{
|
||||||
|
Name: "drone",
|
||||||
|
Link: link,
|
||||||
|
Arch: "linux/amd64",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// use helper funciton to return ([]backend.Config, error)
|
||||||
|
|
||||||
|
// 1. fetch everything from github
|
||||||
|
// 2. create and persist the build object
|
||||||
|
//
|
||||||
|
// 3. generate the build jobs [Launcher?]
|
||||||
|
// a. parse yaml
|
||||||
|
// b. lint yaml
|
||||||
|
// c. compile yaml
|
||||||
|
//
|
||||||
|
// 4. persist the build jobs (... what if I already have jobs, via re-start)
|
||||||
|
// 5. update github status
|
||||||
|
// 6. send to queue
|
||||||
|
// 7. trigger pubsub
|
||||||
|
|
||||||
|
type builder struct {
|
||||||
|
Repo *model.Repo
|
||||||
|
Curr *model.Build
|
||||||
|
Last *model.Build
|
||||||
|
Netrc *model.Netrc
|
||||||
|
Secs []*model.Secret
|
||||||
|
Link string
|
||||||
|
Yaml string
|
||||||
|
}
|
||||||
|
|
||||||
|
type buildItem struct {
|
||||||
|
Job *model.Job
|
||||||
|
Platform string
|
||||||
|
Labels map[string]string
|
||||||
|
Config *backend.Config
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *builder) Build() ([]*buildItem, error) {
|
||||||
|
|
||||||
|
axes, err := matrix.ParseString(b.Yaml)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if len(axes) == 0 {
|
||||||
|
axes = append(axes, matrix.Axis{})
|
||||||
|
}
|
||||||
|
|
||||||
|
var items []*buildItem
|
||||||
|
for i, axis := range axes {
|
||||||
|
job := &model.Job{
|
||||||
|
BuildID: b.Curr.ID,
|
||||||
|
Number: i + 1,
|
||||||
|
Status: model.StatusPending,
|
||||||
|
Environment: axis,
|
||||||
|
Enqueued: b.Curr.Created,
|
||||||
|
}
|
||||||
|
|
||||||
|
metadata := metadataFromStruct(b.Repo, b.Curr, b.Last, job, b.Link)
|
||||||
|
environ := metadata.Environ()
|
||||||
|
for k, v := range metadata.EnvironDrone() {
|
||||||
|
environ[k] = v
|
||||||
|
}
|
||||||
|
|
||||||
|
secrets := map[string]string{}
|
||||||
|
for _, sec := range b.Secs {
|
||||||
|
if !sec.MatchEvent(b.Curr.Event) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if b.Curr.Verified || sec.SkipVerify {
|
||||||
|
secrets[sec.Name] = sec.Value
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sub := func(name string) string {
|
||||||
|
if v, ok := environ[name]; ok {
|
||||||
|
return v
|
||||||
|
}
|
||||||
|
return secrets[name]
|
||||||
|
}
|
||||||
|
|
||||||
|
y := b.Yaml
|
||||||
|
if s, err := envsubst.Eval(y, sub); err != nil {
|
||||||
|
y = s
|
||||||
|
}
|
||||||
|
|
||||||
|
parsed, err := yaml.ParseString(y)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
metadata.Sys.Arch = parsed.Platform
|
||||||
|
if metadata.Sys.Arch == "" {
|
||||||
|
metadata.Sys.Arch = "linux/amd64"
|
||||||
|
}
|
||||||
|
|
||||||
|
lerr := linter.New(
|
||||||
|
linter.WithTrusted(b.Repo.IsTrusted),
|
||||||
|
).Lint(parsed)
|
||||||
|
if lerr != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
ir := compiler.New(
|
||||||
|
compiler.WithEnviron(environ),
|
||||||
|
// TODO ability to customize the escalated plugins
|
||||||
|
compiler.WithEscalated("plugins/docker", "plugins/gcr", "plugins/ecr"),
|
||||||
|
compiler.WithLocal(false),
|
||||||
|
compiler.WithNetrc(b.Netrc.Login, b.Netrc.Password, b.Netrc.Machine),
|
||||||
|
compiler.WithPrefix(
|
||||||
|
fmt.Sprintf(
|
||||||
|
"%d_%d",
|
||||||
|
job.ID,
|
||||||
|
time.Now().Unix(),
|
||||||
|
),
|
||||||
|
),
|
||||||
|
compiler.WithEnviron(job.Environment),
|
||||||
|
compiler.WithProxy(),
|
||||||
|
// TODO ability to set global volumes for things like certs
|
||||||
|
compiler.WithVolumes(),
|
||||||
|
compiler.WithWorkspaceFromURL("/drone", b.Curr.Link),
|
||||||
|
).Compile(parsed)
|
||||||
|
|
||||||
|
for _, sec := range b.Secs {
|
||||||
|
if !sec.MatchEvent(b.Curr.Event) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if b.Curr.Verified || sec.SkipVerify {
|
||||||
|
ir.Secrets = append(ir.Secrets, &backend.Secret{
|
||||||
|
Mask: sec.Conceal,
|
||||||
|
Name: sec.Name,
|
||||||
|
Value: sec.Value,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
item := &buildItem{
|
||||||
|
Job: job,
|
||||||
|
Config: ir,
|
||||||
|
Labels: parsed.Labels,
|
||||||
|
Platform: metadata.Sys.Arch,
|
||||||
|
}
|
||||||
|
if item.Labels == nil {
|
||||||
|
item.Labels = map[string]string{}
|
||||||
|
}
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
|
||||||
|
return items, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
func PostHook2(c *gin.Context) {
|
func PostHook2(c *gin.Context) {
|
||||||
remote_ := remote.FromContext(c)
|
remote_ := remote.FromContext(c)
|
||||||
|
|
||||||
|
@ -142,16 +736,6 @@ func PostHook2(c *gin.Context) {
|
||||||
sec, err := remote_.File(user, repo, build, cfg.Shasum)
|
sec, err := remote_.File(user, repo, build, cfg.Shasum)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logrus.Debugf("cannot find yaml signature for %s. %s", repo.FullName, err)
|
logrus.Debugf("cannot find yaml signature for %s. %s", repo.FullName, err)
|
||||||
// NOTE we don't exit on failure. The sec file is optional
|
|
||||||
}
|
|
||||||
|
|
||||||
axes, err := matrix.Parse(raw)
|
|
||||||
if err != nil {
|
|
||||||
c.String(500, "Failed to parse yaml file or calculate matrix. %s", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if len(axes) == 0 {
|
|
||||||
axes = append(axes, matrix.Axis{})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
netrc, err := remote_.Netrc(user, repo)
|
netrc, err := remote_.Netrc(user, repo)
|
||||||
|
@ -192,18 +776,7 @@ func PostHook2(c *gin.Context) {
|
||||||
build.Status = model.StatusPending
|
build.Status = model.StatusPending
|
||||||
build.RepoID = repo.ID
|
build.RepoID = repo.ID
|
||||||
|
|
||||||
// and use a transaction
|
if err := store.CreateBuild(c, build, build.Jobs...); err != nil {
|
||||||
var jobs []*model.Job
|
|
||||||
for num, axis := range axes {
|
|
||||||
jobs = append(jobs, &model.Job{
|
|
||||||
BuildID: build.ID,
|
|
||||||
Number: num + 1,
|
|
||||||
Status: model.StatusPending,
|
|
||||||
Environment: axis,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
err = store.CreateBuild(c, build, jobs...)
|
|
||||||
if err != nil {
|
|
||||||
logrus.Errorf("failure to save commit for %s. %s", repo.FullName, err)
|
logrus.Errorf("failure to save commit for %s. %s", repo.FullName, err)
|
||||||
c.AbortWithError(500, err)
|
c.AbortWithError(500, err)
|
||||||
return
|
return
|
||||||
|
@ -211,12 +784,6 @@ func PostHook2(c *gin.Context) {
|
||||||
|
|
||||||
c.JSON(200, build)
|
c.JSON(200, build)
|
||||||
|
|
||||||
uri := fmt.Sprintf("%s/%s/%d", httputil.GetURL(c.Request), repo.FullName, build.Number)
|
|
||||||
err = remote_.Status(user, repo, build, uri)
|
|
||||||
if err != nil {
|
|
||||||
logrus.Errorf("error setting commit status for %s/%d", repo.FullName, build.Number)
|
|
||||||
}
|
|
||||||
|
|
||||||
// get the previous build so that we can send
|
// get the previous build so that we can send
|
||||||
// on status change notifications
|
// on status change notifications
|
||||||
last, _ := store.GetBuildLastBefore(c, repo, build.Branch, build.ID)
|
last, _ := store.GetBuildLastBefore(c, repo, build.Branch, build.ID)
|
||||||
|
@ -226,9 +793,44 @@ func PostHook2(c *gin.Context) {
|
||||||
}
|
}
|
||||||
|
|
||||||
//
|
//
|
||||||
// new code here
|
// BELOW: NEW
|
||||||
//
|
//
|
||||||
|
|
||||||
|
defer func() {
|
||||||
|
uri := fmt.Sprintf("%s/%s/%d", httputil.GetURL(c.Request), repo.FullName, build.Number)
|
||||||
|
err = remote_.Status(user, repo, build, uri)
|
||||||
|
if err != nil {
|
||||||
|
logrus.Errorf("error setting commit status for %s/%d", repo.FullName, build.Number)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
b := builder{
|
||||||
|
Repo: repo,
|
||||||
|
Curr: build,
|
||||||
|
Last: last,
|
||||||
|
Netrc: netrc,
|
||||||
|
Secs: secs,
|
||||||
|
Link: httputil.GetURL(c.Request),
|
||||||
|
Yaml: string(raw),
|
||||||
|
}
|
||||||
|
items, err := b.Build()
|
||||||
|
if err != nil {
|
||||||
|
build.Status = model.StatusError
|
||||||
|
build.Started = time.Now().Unix()
|
||||||
|
build.Finished = build.Started
|
||||||
|
build.Error = err.Error()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, item := range items {
|
||||||
|
build.Jobs = append(build.Jobs, item.Job)
|
||||||
|
store.CreateJob(c, item.Job)
|
||||||
|
// TODO err
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
// publish topic
|
||||||
|
//
|
||||||
message := pubsub.Message{
|
message := pubsub.Message{
|
||||||
Labels: map[string]string{
|
Labels: map[string]string{
|
||||||
"repo": repo.FullName,
|
"repo": repo.FullName,
|
||||||
|
@ -242,199 +844,26 @@ func PostHook2(c *gin.Context) {
|
||||||
})
|
})
|
||||||
// TODO remove global reference
|
// TODO remove global reference
|
||||||
config.pubsub.Publish(c, "topic/events", message)
|
config.pubsub.Publish(c, "topic/events", message)
|
||||||
|
|
||||||
//
|
//
|
||||||
// workspace
|
// end publish topic
|
||||||
//
|
//
|
||||||
|
|
||||||
for _, job := range jobs {
|
for _, item := range items {
|
||||||
|
|
||||||
metadata := metadataFromStruct(repo, build, last, job, httputil.GetURL(c.Request))
|
|
||||||
environ := metadata.Environ()
|
|
||||||
|
|
||||||
secrets := map[string]string{}
|
|
||||||
for _, sec := range secs {
|
|
||||||
if !sec.MatchEvent(build.Event) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if build.Verified || sec.SkipVerify {
|
|
||||||
secrets[sec.Name] = sec.Value
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sub := func(name string) string {
|
|
||||||
if v, ok := environ[name]; ok {
|
|
||||||
return v
|
|
||||||
}
|
|
||||||
return secrets[name]
|
|
||||||
}
|
|
||||||
if s, err := envsubst.Eval(string(raw), sub); err != nil {
|
|
||||||
raw = []byte(s)
|
|
||||||
}
|
|
||||||
parsed, err := yaml.ParseBytes(raw)
|
|
||||||
if err != nil {
|
|
||||||
job.ExitCode = 255
|
|
||||||
job.Enqueued = time.Now().Unix()
|
|
||||||
job.Started = time.Now().Unix()
|
|
||||||
job.Finished = time.Now().Unix()
|
|
||||||
job.Error = err.Error()
|
|
||||||
store.UpdateBuildJob(c, build, job)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
lerr := linter.New(
|
|
||||||
linter.WithTrusted(repo.IsTrusted),
|
|
||||||
).Lint(parsed)
|
|
||||||
if lerr != nil {
|
|
||||||
job.ExitCode = 255
|
|
||||||
job.Enqueued = time.Now().Unix()
|
|
||||||
job.Started = time.Now().Unix()
|
|
||||||
job.Finished = time.Now().Unix()
|
|
||||||
job.Error = lerr.Error()
|
|
||||||
store.UpdateBuildJob(c, build, job)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
ir := compiler.New(
|
|
||||||
compiler.WithEnviron(environ),
|
|
||||||
// TODO ability to customize the escalated plugins
|
|
||||||
compiler.WithEscalated("plugins/docker", "plugins/gcr", "plugins/ecr"),
|
|
||||||
compiler.WithLocal(false),
|
|
||||||
compiler.WithNetrc(netrc.Login, netrc.Password, netrc.Machine),
|
|
||||||
compiler.WithPrefix(
|
|
||||||
fmt.Sprintf(
|
|
||||||
"%d_%d",
|
|
||||||
job.ID,
|
|
||||||
time.Now().Unix(),
|
|
||||||
),
|
|
||||||
),
|
|
||||||
compiler.WithEnviron(job.Environment),
|
|
||||||
compiler.WithProxy(),
|
|
||||||
// TODO ability to set global volumes for things like certs
|
|
||||||
compiler.WithVolumes(),
|
|
||||||
compiler.WithWorkspaceFromURL("/drone", repo.Link),
|
|
||||||
).Compile(parsed)
|
|
||||||
|
|
||||||
// TODO there is a chicken and egg problem here because
|
|
||||||
// the compiled yaml has a platform environment variable
|
|
||||||
// that is not correctly set, because we are just about
|
|
||||||
// to set it ....
|
|
||||||
// TODO maybe we remove platform from metadata and let
|
|
||||||
// the compiler set the value from the yaml itself.
|
|
||||||
if parsed.Platform == "" {
|
|
||||||
parsed.Platform = "linux/amd64"
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, sec := range secs {
|
|
||||||
if !sec.MatchEvent(build.Event) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if build.Verified || sec.SkipVerify {
|
|
||||||
ir.Secrets = append(ir.Secrets, &backend.Secret{
|
|
||||||
Mask: sec.Conceal,
|
|
||||||
Name: sec.Name,
|
|
||||||
Value: sec.Value,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
task := new(queue.Task)
|
task := new(queue.Task)
|
||||||
task.ID = fmt.Sprint(job.ID)
|
task.ID = fmt.Sprint(item.Job.ID)
|
||||||
task.Labels = map[string]string{}
|
task.Labels = map[string]string{}
|
||||||
task.Labels["platform"] = parsed.Platform
|
task.Labels["platform"] = item.Platform
|
||||||
if parsed.Labels != nil {
|
for k, v := range item.Labels {
|
||||||
for k, v := range parsed.Labels {
|
|
||||||
task.Labels[k] = v
|
task.Labels[k] = v
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
task.Data, _ = json.Marshal(rpc.Pipeline{
|
task.Data, _ = json.Marshal(rpc.Pipeline{
|
||||||
ID: fmt.Sprint(job.ID),
|
ID: fmt.Sprint(item.Job.ID),
|
||||||
Config: ir,
|
Config: item.Config,
|
||||||
Timeout: repo.Timeout,
|
Timeout: b.Repo.Timeout,
|
||||||
})
|
})
|
||||||
|
|
||||||
config.logger.Open(context.Background(), task.ID)
|
config.logger.Open(context.Background(), task.ID)
|
||||||
config.queue.Push(context.Background(), task)
|
config.queue.Push(context.Background(), task)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// return the metadata from the cli context.
|
|
||||||
func metadataFromStruct(repo *model.Repo, build, last *model.Build, job *model.Job, link string) frontend.Metadata {
|
|
||||||
return frontend.Metadata{
|
|
||||||
Repo: frontend.Repo{
|
|
||||||
Name: repo.Name,
|
|
||||||
Link: repo.Link,
|
|
||||||
Remote: repo.Clone,
|
|
||||||
Private: repo.IsPrivate,
|
|
||||||
},
|
|
||||||
Curr: frontend.Build{
|
|
||||||
Number: build.Number,
|
|
||||||
Created: build.Created,
|
|
||||||
Started: build.Started,
|
|
||||||
Finished: build.Finished,
|
|
||||||
Status: build.Status,
|
|
||||||
Event: build.Event,
|
|
||||||
Link: build.Link,
|
|
||||||
Target: build.Deploy,
|
|
||||||
Commit: frontend.Commit{
|
|
||||||
Sha: build.Commit,
|
|
||||||
Ref: build.Ref,
|
|
||||||
Refspec: build.Refspec,
|
|
||||||
Branch: build.Branch,
|
|
||||||
Message: build.Message,
|
|
||||||
Author: frontend.Author{
|
|
||||||
Name: build.Author,
|
|
||||||
Email: build.Email,
|
|
||||||
Avatar: build.Avatar,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Prev: frontend.Build{
|
|
||||||
Number: last.Number,
|
|
||||||
Created: last.Created,
|
|
||||||
Started: last.Started,
|
|
||||||
Finished: last.Finished,
|
|
||||||
Status: last.Status,
|
|
||||||
Event: last.Event,
|
|
||||||
Link: last.Link,
|
|
||||||
Target: last.Deploy,
|
|
||||||
Commit: frontend.Commit{
|
|
||||||
Sha: last.Commit,
|
|
||||||
Ref: last.Ref,
|
|
||||||
Refspec: last.Refspec,
|
|
||||||
Branch: last.Branch,
|
|
||||||
Message: last.Message,
|
|
||||||
Author: frontend.Author{
|
|
||||||
Name: last.Author,
|
|
||||||
Email: last.Email,
|
|
||||||
Avatar: last.Avatar,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Job: frontend.Job{
|
|
||||||
Number: job.Number,
|
|
||||||
Matrix: job.Environment,
|
|
||||||
},
|
|
||||||
Sys: frontend.System{
|
|
||||||
Name: "drone",
|
|
||||||
Link: link,
|
|
||||||
Arch: "linux/amd64",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// use helper funciton to return ([]backend.Config, error)
|
|
||||||
|
|
||||||
type builder struct {
|
|
||||||
secs []*model.Secret
|
|
||||||
repo *model.Repo
|
|
||||||
build *model.Build
|
|
||||||
last *model.Build
|
|
||||||
jobs []*model.Job
|
|
||||||
link string
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *builder) Build() ([]*backend.Config, error) {
|
|
||||||
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
|
|
|
@ -51,7 +51,6 @@ func init() {
|
||||||
|
|
||||||
func RPCHandler(c *gin.Context) {
|
func RPCHandler(c *gin.Context) {
|
||||||
|
|
||||||
fmt.Println(c.Request.Header.Write(os.Stdout))
|
|
||||||
if secret := c.Request.Header.Get("Authorization"); secret != "Bearer "+config.secret {
|
if secret := c.Request.Header.Get("Authorization"); secret != "Bearer "+config.secret {
|
||||||
log.Printf("Unable to connect agent. Invalid authorization token %q does not match %q", secret, config.secret)
|
log.Printf("Unable to connect agent. Invalid authorization token %q does not match %q", secret, config.secret)
|
||||||
c.String(401, "Unable to connect agent. Invalid authorization token")
|
c.String(401, "Unable to connect agent. Invalid authorization token")
|
||||||
|
|
1
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/compiler.go
generated
vendored
1
vendor/github.com/cncd/pipeline/pipeline/frontend/yaml/compiler/compiler.go
generated
vendored
|
@ -64,6 +64,7 @@ func (c *Compiler) Compile(conf *yaml.Config) *backend.Config {
|
||||||
// add default clone step
|
// add default clone step
|
||||||
if c.local == false && len(conf.Clone.Containers) == 0 {
|
if c.local == false && len(conf.Clone.Containers) == 0 {
|
||||||
container := &yaml.Container{
|
container := &yaml.Container{
|
||||||
|
Name: "clone",
|
||||||
Image: "plugins/git:latest",
|
Image: "plugins/git:latest",
|
||||||
Vargs: map[string]interface{}{"depth": "0"},
|
Vargs: map[string]interface{}{"depth": "0"},
|
||||||
}
|
}
|
||||||
|
|
46
vendor/vendor.json
vendored
46
vendor/vendor.json
vendored
|
@ -33,68 +33,68 @@
|
||||||
{
|
{
|
||||||
"checksumSHA1": "W3AuK8ocqHwlUajGmQLFvnRhTZE=",
|
"checksumSHA1": "W3AuK8ocqHwlUajGmQLFvnRhTZE=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline",
|
"path": "github.com/cncd/pipeline/pipeline",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "Qu2FreqaMr8Yx2bW9O0cxAGgjr0=",
|
"checksumSHA1": "Qu2FreqaMr8Yx2bW9O0cxAGgjr0=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/backend",
|
"path": "github.com/cncd/pipeline/pipeline/backend",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "0CGXRaYwZhJxGIrGhn8WGpkFqPo=",
|
"checksumSHA1": "0CGXRaYwZhJxGIrGhn8WGpkFqPo=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/backend/docker",
|
"path": "github.com/cncd/pipeline/pipeline/backend/docker",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "/8wE+cVb7T4PQZgpLNu0DHzKGuE=",
|
"checksumSHA1": "/8wE+cVb7T4PQZgpLNu0DHzKGuE=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/frontend",
|
"path": "github.com/cncd/pipeline/pipeline/frontend",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "O0sulBQAHJeNLg3lO38Cq5uf/eg=",
|
"checksumSHA1": "O0sulBQAHJeNLg3lO38Cq5uf/eg=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml",
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "Iu+QmUqkN9ZsBdmVlCclVKthJbM=",
|
"checksumSHA1": "srNvtlujHSHC8YXrnFPdy5V7qoQ=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/compiler",
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/compiler",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "Q0GkNUFamVYIA1Fd8r0A5M6Gx54=",
|
"checksumSHA1": "Q0GkNUFamVYIA1Fd8r0A5M6Gx54=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/linter",
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/linter",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "kx2sPUIMozPC/g6E4w48h3FfH3k=",
|
"checksumSHA1": "kx2sPUIMozPC/g6E4w48h3FfH3k=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/matrix",
|
"path": "github.com/cncd/pipeline/pipeline/frontend/yaml/matrix",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "2/3f3oNmxXy5kcrRLCFa24Oc9O4=",
|
"checksumSHA1": "2/3f3oNmxXy5kcrRLCFa24Oc9O4=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/interrupt",
|
"path": "github.com/cncd/pipeline/pipeline/interrupt",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "uOjTfke7Qxosrivgz/nVTHeIP5g=",
|
"checksumSHA1": "uOjTfke7Qxosrivgz/nVTHeIP5g=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/multipart",
|
"path": "github.com/cncd/pipeline/pipeline/multipart",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "MratmNKJ78/IhWvDsZphN01CtmE=",
|
"checksumSHA1": "MratmNKJ78/IhWvDsZphN01CtmE=",
|
||||||
"path": "github.com/cncd/pipeline/pipeline/rpc",
|
"path": "github.com/cncd/pipeline/pipeline/rpc",
|
||||||
"revision": "a4badab355e3c2ecc03f2048b0b97f0180df12b2",
|
"revision": "3614f41e777232289a3cd479c2301274556e3346",
|
||||||
"revisionTime": "2017-03-13T03:05:09Z"
|
"revisionTime": "2017-03-14T15:17:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "7Qj1DK0ceAXkYztW0l3+L6sn+V8=",
|
"checksumSHA1": "7Qj1DK0ceAXkYztW0l3+L6sn+V8=",
|
||||||
|
|
Loading…
Reference in a new issue