Compare commits
45 Commits
4a4b6fd185
..
v0.2.0
| Author | SHA1 | Date | |
|---|---|---|---|
| e1382e50ea | |||
| 7f44e5ed41 | |||
| bcc53dfbe0 | |||
| bbf96498aa | |||
| 954966db58 | |||
| ed7df18f83 | |||
| a8e9a68f0e | |||
| 20c664f0ed | |||
| 0000ea2a13 | |||
| fe53a17160 | |||
| f190274bce | |||
| d9ba14550e | |||
| a2acb99689 | |||
| 191b73fe41 | |||
| 3ca1481632 | |||
| c0e33fa52a | |||
| 63529b7174 | |||
| 7e7c49c2e7 | |||
| 712a7b1429 | |||
| 3ae27bffc5 | |||
| 5373a37bee | |||
| 6fee5aa268 | |||
| b475631df6 | |||
| fbf918d627 | |||
| 4069e1b0e1 | |||
| 708fbca91a | |||
| 77a5514578 | |||
| 77a8d0840a | |||
| 62b4e8f17e | |||
| 620c20f717 | |||
| 0979a2379e | |||
| 85ebd856eb | |||
| 6b103d074e | |||
| 664fe8fd09 | |||
| edafd5108a | |||
| c0f6186eac | |||
| bfd05b6a8a | |||
| 6d2936393b | |||
| e4043ae3be | |||
| 7a665aa348 | |||
| d870335d25 | |||
| 4bda3c7a3b | |||
| 724757b23c | |||
| 1e0526a599 | |||
| c1d4f3cc16 |
+573
-19
@@ -8,9 +8,143 @@ import (
|
|||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
"github.com/graphql-go/graphql"
|
"github.com/graphql-go/graphql"
|
||||||
"github.com/graphql-go/handler"
|
"github.com/graphql-go/handler"
|
||||||
|
"github.com/op/go-logging"
|
||||||
)
|
)
|
||||||
|
|
||||||
func createSchema(db database.Database) (graphql.Schema, error) {
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
|
func createSchema(db database.Database, pollChan chan uuid.UUID) (graphql.Schema, error) {
|
||||||
|
runnerType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
|
Name: "Runner",
|
||||||
|
Description: "A runner available for use inside of a pipeline.",
|
||||||
|
Fields: graphql.Fields{
|
||||||
|
"id": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The id of the runner.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if runner, ok := p.Source.(database.Runner); ok {
|
||||||
|
return runner.Id, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"name": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The name of the runner.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if runner, ok := p.Source.(database.Runner); ok {
|
||||||
|
return runner.Name, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"token": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The token.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if runner, ok := p.Source.(database.Runner); ok {
|
||||||
|
return runner.Token, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
secretType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
|
Name: "Secret",
|
||||||
|
Description: "A secret available for use inside of a pipeline.",
|
||||||
|
Fields: graphql.Fields{
|
||||||
|
"id": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The id of the secret.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if secret, ok := p.Source.(database.Secret); ok {
|
||||||
|
return secret.Id, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"name": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The name of the secret.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if secret, ok := p.Source.(database.Secret); ok {
|
||||||
|
return secret.Name, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"secret": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The secret.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if secret, ok := p.Source.(database.Secret); ok {
|
||||||
|
return secret.Secret, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
cloneCredentialType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
|
Name: "CloneCredential",
|
||||||
|
Description: "A credential for authenticating with the pipeline source host.",
|
||||||
|
Fields: graphql.Fields{
|
||||||
|
"id": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The id of the credential.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if credential, ok := p.Source.(database.CloneCredential); ok {
|
||||||
|
return credential.Id, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"name": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The name of the credential.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if credential, ok := p.Source.(database.CloneCredential); ok {
|
||||||
|
return credential.Name, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"type": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The credential type.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if credential, ok := p.Source.(database.CloneCredential); ok {
|
||||||
|
return credential.Type, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"username": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The username to user with the credential.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if credential, ok := p.Source.(database.CloneCredential); ok {
|
||||||
|
return credential.Username, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"secret": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The secret for the credential.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if credential, ok := p.Source.(database.CloneCredential); ok {
|
||||||
|
return credential.Secret, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
webhookType := graphql.NewObject(graphql.ObjectConfig{
|
webhookType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
Name: "Webhook",
|
Name: "Webhook",
|
||||||
Description: "A webhook for triggering pipelines",
|
Description: "A webhook for triggering pipelines",
|
||||||
@@ -48,6 +182,74 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
|
runType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
|
Name: "Run",
|
||||||
|
Description: "A pipeline run",
|
||||||
|
Fields: graphql.Fields{
|
||||||
|
"id": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The id of the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return run.Id, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"inProgress": &graphql.Field{
|
||||||
|
Type: graphql.Boolean,
|
||||||
|
Description: "The progress status of the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return run.InProgress, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"result": &graphql.Field{
|
||||||
|
// TODO: handle bigint properly here
|
||||||
|
Type: graphql.Float,
|
||||||
|
Description: "The result of the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return run.Result, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"buildOutput": &graphql.Field{
|
||||||
|
Type: graphql.String,
|
||||||
|
Description: "Logs of the top level container build for the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return string(run.BuildOutput), nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"stdout": &graphql.Field{
|
||||||
|
Type: graphql.String,
|
||||||
|
Description: "The stdout used to validate the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return string(run.Stdout), nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"stderr": &graphql.Field{
|
||||||
|
Type: graphql.String,
|
||||||
|
Description: "The stderr used to validate the run.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if run, ok := p.Source.(database.Run); ok {
|
||||||
|
return string(run.Stderr), nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
pipelineType := graphql.NewObject(graphql.ObjectConfig{
|
pipelineType := graphql.NewObject(graphql.ObjectConfig{
|
||||||
Name: "Pipeline",
|
Name: "Pipeline",
|
||||||
Description: "A pipeline for running ci jobs",
|
Description: "A pipeline for running ci jobs",
|
||||||
@@ -92,23 +294,26 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
return nil, nil
|
return nil, nil
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
"createWebhook": &graphql.Field{
|
"cloneCredential": &graphql.Field{
|
||||||
Type: webhookType,
|
Type: cloneCredentialType,
|
||||||
Description: "Create a new webhook",
|
Description: "The configured credential for cloning the pipeline source.",
|
||||||
Args: graphql.FieldConfigArgument{
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
"type": &graphql.ArgumentConfig{
|
if pipeline, ok := p.Source.(database.Pipeline); ok {
|
||||||
Type: graphql.NewNonNull(graphql.String),
|
if pipeline.CloneCredential != nil {
|
||||||
},
|
return db.GetCloneCredentialById(*pipeline.CloneCredential)
|
||||||
},
|
|
||||||
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
|
||||||
webhook, err := db.CreateWebhook(
|
|
||||||
database.WebhookSender(params.Args["type"].(string)),
|
|
||||||
params.Source.(database.Pipeline).Id,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
}
|
||||||
return webhook, nil
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"secrets": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(graphql.NewNonNull(secretType))),
|
||||||
|
Description: "The list of secrets for the pipeline.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if pipeline, ok := p.Source.(database.Pipeline); ok {
|
||||||
|
return db.GetSecretsForPipeline(pipeline.Id)
|
||||||
|
}
|
||||||
|
return []database.Secret{}, nil
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
"webhooks": &graphql.Field{
|
"webhooks": &graphql.Field{
|
||||||
@@ -121,6 +326,16 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
return []database.Webhook{}, nil
|
return []database.Webhook{}, nil
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
"runs": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(graphql.NewNonNull(runType))),
|
||||||
|
Description: "The list of runs for the pipeline.",
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
if pipeline, ok := p.Source.(database.Pipeline); ok {
|
||||||
|
return db.GetRunsForPipeline(pipeline.Id)
|
||||||
|
}
|
||||||
|
return []database.Webhook{}, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
@@ -143,6 +358,50 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
return db.GetPipelineById(id)
|
return db.GetPipelineById(id)
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
"Pipelines": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(pipelineType)),
|
||||||
|
Args: graphql.FieldConfigArgument{},
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
return db.GetPipelines()
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"CloneCredential": &graphql.Field{
|
||||||
|
Type: cloneCredentialType,
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"id": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
Description: "The id of the requested credential.",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
id, err := uuid.Parse(p.Args["id"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return db.GetCloneCredentialById(id)
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"CloneCredentials": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(cloneCredentialType)),
|
||||||
|
Args: graphql.FieldConfigArgument{},
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
return db.GetCredentials()
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"Secrets": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(secretType)),
|
||||||
|
Args: graphql.FieldConfigArgument{},
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
return db.GetSecrets()
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"Runners": &graphql.Field{
|
||||||
|
Type: graphql.NewNonNull(graphql.NewList(runnerType)),
|
||||||
|
Args: graphql.FieldConfigArgument{},
|
||||||
|
Resolve: func(p graphql.ResolveParams) (interface{}, error) {
|
||||||
|
return db.GetRunners()
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
@@ -162,6 +421,9 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
"pollInterval": &graphql.ArgumentConfig{
|
"pollInterval": &graphql.ArgumentConfig{
|
||||||
Type: graphql.Int,
|
Type: graphql.Int,
|
||||||
},
|
},
|
||||||
|
"cloneCredentialId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
var interval int
|
var interval int
|
||||||
@@ -170,14 +432,306 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
} else {
|
} else {
|
||||||
interval = 0
|
interval = 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var credential *uuid.UUID
|
||||||
|
if credentialVal, ok := params.Args["cloneCredentialId"]; ok {
|
||||||
|
id, err := uuid.Parse(credentialVal.(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
credential = &id
|
||||||
|
} else {
|
||||||
|
credential = nil
|
||||||
|
}
|
||||||
|
|
||||||
pipeline, err := db.CreatePipeline(
|
pipeline, err := db.CreatePipeline(
|
||||||
params.Args["name"].(string),
|
params.Args["name"].(string),
|
||||||
params.Args["url"].(string),
|
params.Args["url"].(string),
|
||||||
interval,
|
interval,
|
||||||
|
credential,
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pollChan <- pipeline.Id
|
||||||
|
|
||||||
|
return pipeline, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"createWebhook": &graphql.Field{
|
||||||
|
Type: webhookType,
|
||||||
|
Description: "Create a new webhook",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"type": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"pipelineId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
id, err := uuid.Parse(params.Args["id"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
webhook, err := db.CreateWebhook(
|
||||||
|
database.WebhookSender(params.Args["type"].(string)),
|
||||||
|
id,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return webhook, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"createCloneCredential": &graphql.Field{
|
||||||
|
Type: cloneCredentialType,
|
||||||
|
Description: "Create a new CloneCredential",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"name": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"type": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"username": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"secret": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
credential, err := db.CreateCredential(
|
||||||
|
params.Args["name"].(string),
|
||||||
|
database.CloneCredentialType(params.Args["type"].(string)),
|
||||||
|
params.Args["username"].(string),
|
||||||
|
params.Args["secret"].(string),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return credential, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"createSecret": &graphql.Field{
|
||||||
|
Type: secretType,
|
||||||
|
Description: "Create a new secret",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"name": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"secret": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
secret, err := db.CreateSecret(
|
||||||
|
params.Args["name"].(string),
|
||||||
|
params.Args["secret"].(string),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return secret, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"createRunner": &graphql.Field{
|
||||||
|
Type: runnerType,
|
||||||
|
Description: "Create a new runner",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"name": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
runner, err := db.CreateRunner(
|
||||||
|
params.Args["name"].(string),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return runner, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"updatePipeline": &graphql.Field{
|
||||||
|
Type: pipelineType,
|
||||||
|
Description: "Create a new pipeline",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"pipelineId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
"name": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
"url": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
"pollInterval": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.Int,
|
||||||
|
},
|
||||||
|
"cloneCredentialId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
pipelineId, err := uuid.Parse(params.Args["pipelineId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var name *string
|
||||||
|
var url *string
|
||||||
|
var interval *int
|
||||||
|
|
||||||
|
if nameVal, ok := params.Args["name"]; ok {
|
||||||
|
nameVal := nameVal.(string)
|
||||||
|
name = &nameVal
|
||||||
|
} else {
|
||||||
|
name = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if urlVal, ok := params.Args["url"]; ok {
|
||||||
|
urlVal := urlVal.(string)
|
||||||
|
url = &urlVal
|
||||||
|
} else {
|
||||||
|
url = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if intervalVal, ok := params.Args["pollInterval"]; ok {
|
||||||
|
intervalVal := intervalVal.(int)
|
||||||
|
interval = &intervalVal
|
||||||
|
} else {
|
||||||
|
interval = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline, err := db.UpdatePipeline(
|
||||||
|
pipelineId,
|
||||||
|
name,
|
||||||
|
url,
|
||||||
|
interval,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pollChan <- pipeline.Id
|
||||||
|
|
||||||
|
return pipeline, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"setPipelineCloneCredential": &graphql.Field{
|
||||||
|
Type: pipelineType,
|
||||||
|
Description: "Set the CloneCredential used by a pipeline to clone the source repo",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"cloneCredentialId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
"pipelineId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
pipelineId, err := uuid.Parse(params.Args["pipelineId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if cloneCredentialIdVal, ok := params.Args["cloneCredentialId"]; ok {
|
||||||
|
cloneCredentialId, err := uuid.Parse(cloneCredentialIdVal.(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline, err := db.SetPipelineCloneCredential(pipelineId, &cloneCredentialId)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return pipeline, nil
|
||||||
|
} else {
|
||||||
|
pipeline, err := db.SetPipelineCloneCredential(pipelineId, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return pipeline, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"addSecretToPipeline": &graphql.Field{
|
||||||
|
Type: pipelineType,
|
||||||
|
Description: "Allow a secret to be accessed by a pipeline.",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"secretId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
"pipelineId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
secretId, err := uuid.Parse(params.Args["secretId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipelineId, err := uuid.Parse(params.Args["pipelineId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = db.AssignSecretToPipeline(pipelineId, secretId)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline, err := db.GetPipelineById(pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return pipeline, nil
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"removeSecretFromPipeline": &graphql.Field{
|
||||||
|
Type: pipelineType,
|
||||||
|
Description: "Remove a pipeline's access to a secret.",
|
||||||
|
Args: graphql.FieldConfigArgument{
|
||||||
|
"secretId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.String,
|
||||||
|
},
|
||||||
|
"pipelineId": &graphql.ArgumentConfig{
|
||||||
|
Type: graphql.NewNonNull(graphql.String),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Resolve: func(params graphql.ResolveParams) (interface{}, error) {
|
||||||
|
|
||||||
|
secretId, err := uuid.Parse(params.Args["secretId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipelineId, err := uuid.Parse(params.Args["pipelineId"].(string))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = db.RemoveSecretFromPipeline(pipelineId, secretId)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline, err := db.GetPipelineById(pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
return pipeline, nil
|
return pipeline, nil
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@@ -195,9 +749,9 @@ func createSchema(db database.Database) (graphql.Schema, error) {
|
|||||||
return schema, nil
|
return schema, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func CreateHandler(db database.Database, mux *http.ServeMux) error {
|
func CreateHandler(db database.Database, pollChan chan uuid.UUID, mux *http.ServeMux) error {
|
||||||
|
|
||||||
schema, err := createSchema(db)
|
schema, err := createSchema(db, pollChan)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -6,24 +6,6 @@ import (
|
|||||||
"git.ohea.xyz/golang/config"
|
"git.ohea.xyz/golang/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
type WebhookSender string
|
|
||||||
|
|
||||||
const (
|
|
||||||
Gitea WebhookSender = "gitea"
|
|
||||||
)
|
|
||||||
|
|
||||||
type Webhook struct {
|
|
||||||
Sender WebhookSender
|
|
||||||
Secret string
|
|
||||||
}
|
|
||||||
|
|
||||||
type Job struct {
|
|
||||||
URL string
|
|
||||||
Webhook *Webhook
|
|
||||||
Cron *string
|
|
||||||
PollInterval uint64
|
|
||||||
}
|
|
||||||
|
|
||||||
type Runner struct {
|
type Runner struct {
|
||||||
Secret string
|
Secret string
|
||||||
}
|
}
|
||||||
@@ -60,7 +42,6 @@ type Config struct {
|
|||||||
Port int
|
Port int
|
||||||
DBConfig DBConfig
|
DBConfig DBConfig
|
||||||
PipelineConf PipelineConf
|
PipelineConf PipelineConf
|
||||||
Jobs map[string]Job
|
|
||||||
Runners map[string]Runner
|
Runners map[string]Runner
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -88,7 +69,6 @@ func GetConfig() (config.Config[Config], error) {
|
|||||||
Source: "/opt/cursorius/working",
|
Source: "/opt/cursorius/working",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
Jobs: make(map[string]Job),
|
|
||||||
Runners: make(map[string]Runner),
|
Runners: make(map[string]Runner),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|||||||
+65
-13
@@ -8,14 +8,14 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"github.com/jackc/pgx/v5"
|
"github.com/jackc/pgx/v5/pgxpool"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
)
|
)
|
||||||
|
|
||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
type Database struct {
|
type Database struct {
|
||||||
Conn *pgx.Conn
|
Conn *pgxpool.Pool
|
||||||
}
|
}
|
||||||
|
|
||||||
func LaunchDB(conf config.DBConfig) (Database, error) {
|
func LaunchDB(conf config.DBConfig) (Database, error) {
|
||||||
@@ -40,9 +40,15 @@ func LaunchDB(conf config.DBConfig) (Database, error) {
|
|||||||
db := Database{}
|
db := Database{}
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
|
log.Infof("Connecting to database with URL \"%v\"", dbURLNoPasswd)
|
||||||
|
db.Conn, err = pgxpool.New(context.Background(), dbURL)
|
||||||
|
if err != nil {
|
||||||
|
return db, fmt.Errorf("could not create database pool: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// sleep until we can sucessfully acquire a connection
|
||||||
for i := 0; i < 10; i++ {
|
for i := 0; i < 10; i++ {
|
||||||
log.Infof("Connecting to database with URL \"%v\" (attempt %v)", dbURLNoPasswd, i)
|
_, err = db.Conn.Acquire(context.Background())
|
||||||
db.Conn, err = pgx.Connect(context.Background(), dbURL)
|
|
||||||
if err == nil {
|
if err == nil {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
@@ -80,19 +86,52 @@ SELECT EXISTS (
|
|||||||
return db, nil
|
return db, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func initDB(conn *pgx.Conn) error {
|
func initDB(conn *pgxpool.Pool) error {
|
||||||
createTablesQuery := `
|
createTablesQuery := `
|
||||||
CREATE EXTENSION IF NOT EXISTS "uuid-ossp";
|
CREATE EXTENSION IF NOT EXISTS "uuid-ossp";
|
||||||
|
|
||||||
CREATE TABLE version (
|
CREATE TABLE version (
|
||||||
version INT NOT NULL
|
version INT NOT NULL
|
||||||
|
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE clone_credentials (
|
||||||
|
id UUID PRIMARY KEY,
|
||||||
|
name TEXT NOT NULL,
|
||||||
|
type TEXT NOT NULL,
|
||||||
|
username TEXT NOT NULL,
|
||||||
|
secret TEXT NOT NULL
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE TABLE pipelines (
|
CREATE TABLE pipelines (
|
||||||
id UUID PRIMARY KEY,
|
id UUID PRIMARY KEY,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
url TEXT NOT NULL,
|
url TEXT NOT NULL,
|
||||||
poll_interval INTEGER
|
poll_interval INTEGER,
|
||||||
|
clone_credential UUID DEFAULT NULL,
|
||||||
|
|
||||||
|
CONSTRAINT fk_clone_credential
|
||||||
|
FOREIGN KEY(clone_credential)
|
||||||
|
REFERENCES clone_credentials(id)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE secrets (
|
||||||
|
id UUID PRIMARY KEY,
|
||||||
|
name TEXT NOT NULL UNIQUE,
|
||||||
|
secret TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE pipeline_secret_mappings (
|
||||||
|
pipeline UUID NOT NULL,
|
||||||
|
secret UUID NOT NULL,
|
||||||
|
|
||||||
|
CONSTRAINT fk_pipeline
|
||||||
|
FOREIGN KEY(pipeline)
|
||||||
|
REFERENCES pipelines(id),
|
||||||
|
|
||||||
|
CONSTRAINT fk_secret
|
||||||
|
FOREIGN KEY(secret)
|
||||||
|
REFERENCES secrets(id)
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE TABLE webhooks (
|
CREATE TABLE webhooks (
|
||||||
@@ -106,16 +145,14 @@ CREATE TABLE webhooks (
|
|||||||
REFERENCES pipelines(id)
|
REFERENCES pipelines(id)
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE TABLE runners (
|
|
||||||
id UUID PRIMARY KEY,
|
|
||||||
name TEXT,
|
|
||||||
secret TEXT
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE runs (
|
CREATE TABLE runs (
|
||||||
id UUID PRIMARY KEY,
|
id UUID PRIMARY KEY,
|
||||||
pipeline UUID,
|
pipeline UUID,
|
||||||
result BOOLEAN NOT NULL,
|
in_progress BOOLEAN DEFAULT NULL,
|
||||||
|
build_output TEXT DEFAULT NULL,
|
||||||
|
result BIGINT DEFAULT NULL,
|
||||||
|
stdout TEXT DEFAULT NULL,
|
||||||
|
stderr TEXT DEFAULT NULL,
|
||||||
|
|
||||||
CONSTRAINT fk_pipeline
|
CONSTRAINT fk_pipeline
|
||||||
FOREIGN KEY(pipeline)
|
FOREIGN KEY(pipeline)
|
||||||
@@ -137,6 +174,21 @@ CREATE TABLE command_executions (
|
|||||||
REFERENCES runs(id)
|
REFERENCES runs(id)
|
||||||
);
|
);
|
||||||
|
|
||||||
|
CREATE TABLE runners (
|
||||||
|
id UUID PRIMARY KEY,
|
||||||
|
name TEXT NOT NULL UNIQUE,
|
||||||
|
token TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE pipeline_refs (
|
||||||
|
name TEXT PRIMARY KEY NOT NULL,
|
||||||
|
pipeline_id UUID NOT NULL,
|
||||||
|
hash TEXT NOT NULL,
|
||||||
|
|
||||||
|
CONSTRAINT fk_pipeline_id
|
||||||
|
FOREIGN KEY(pipeline_id)
|
||||||
|
REFERENCES pipelines(id)
|
||||||
|
);
|
||||||
`
|
`
|
||||||
|
|
||||||
_, err := conn.Exec(context.Background(), createTablesQuery)
|
_, err := conn.Exec(context.Background(), createTablesQuery)
|
||||||
|
|||||||
+566
-18
@@ -3,11 +3,42 @@ package database
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"regexp"
|
||||||
|
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (d *Database) GetPipelineById(id uuid.UUID) (Pipeline, error) {
|
func (db *Database) GetPipelines() ([]Pipeline, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, name, url, poll_interval, clone_credential
|
||||||
|
FROM pipelines;`
|
||||||
|
|
||||||
|
pipelines := make([]Pipeline, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query)
|
||||||
|
if err != nil {
|
||||||
|
return pipelines, fmt.Errorf("Could not query database for pipelines: %w", err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var pipeline Pipeline
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(&idStr, &pipeline.Name, &pipeline.Url, &pipeline.PollInterval, &pipeline.CloneCredential); err != nil {
|
||||||
|
return pipelines, err
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return pipelines, err
|
||||||
|
}
|
||||||
|
pipelines = append(pipelines, pipeline)
|
||||||
|
}
|
||||||
|
|
||||||
|
return pipelines, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetPipelineById(id uuid.UUID) (Pipeline, error) {
|
||||||
query := `
|
query := `
|
||||||
SELECT name, url, poll_interval
|
SELECT name, url, poll_interval
|
||||||
FROM pipelines
|
FROM pipelines
|
||||||
@@ -17,7 +48,7 @@ WHERE id=$1;`
|
|||||||
Id: id,
|
Id: id,
|
||||||
}
|
}
|
||||||
|
|
||||||
err := d.Conn.QueryRow(context.Background(), query, id).Scan(&pipeline.Name, &pipeline.Url, &pipeline.PollInterval)
|
err := db.Conn.QueryRow(context.Background(), query, id).Scan(&pipeline.Name, &pipeline.Url, &pipeline.PollInterval)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return pipeline, fmt.Errorf("Could not query database for pipeline with id %v: %w", id.String(), err)
|
return pipeline, fmt.Errorf("Could not query database for pipeline with id %v: %w", id.String(), err)
|
||||||
}
|
}
|
||||||
@@ -25,29 +56,114 @@ WHERE id=$1;`
|
|||||||
return pipeline, nil
|
return pipeline, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Database) CreatePipeline(name string, url string, pollInterval int) (Pipeline, error) {
|
func (db *Database) CreatePipeline(name string, url string, pollInterval int, credential *uuid.UUID) (Pipeline, error) {
|
||||||
query := `
|
query := `
|
||||||
INSERT INTO pipelines (id, name, url, poll_interval)
|
INSERT INTO pipelines (id, name, url, poll_interval, clone_credential)
|
||||||
VALUES (uuid_generate_v4(), $1, $2, $3)
|
VALUES (uuid_generate_v4(), $1, $2, $3, $4)
|
||||||
RETURNING id, name, url, poll_interval;`
|
RETURNING id, name, url, poll_interval;`
|
||||||
|
|
||||||
pipeline := Pipeline{}
|
pipeline := Pipeline{}
|
||||||
var idStr string
|
var idStr string
|
||||||
err := d.Conn.QueryRow(context.Background(), query, name, url, pollInterval).Scan(&idStr, &pipeline.Name, &pipeline.Url, &pipeline.PollInterval)
|
err := db.Conn.QueryRow(context.Background(), query, name, url, pollInterval, credential).Scan(&idStr, &pipeline.Name, &pipeline.Url, &pipeline.PollInterval)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return pipeline, err
|
return pipeline, fmt.Errorf("Could not create pipeline: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
id, err := uuid.Parse(idStr)
|
pipeline.Id, err = uuid.Parse(idStr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return pipeline, err
|
return pipeline, fmt.Errorf("Could not parse UUID generated by DB: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
pipeline.Id = id
|
|
||||||
|
|
||||||
return pipeline, nil
|
return pipeline, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (db *Database) UpdatePipeline(pipelineId uuid.UUID, name *string, url *string, pollInterval *int) (Pipeline, error) {
|
||||||
|
query := `
|
||||||
|
UPDATE pipelines
|
||||||
|
SET name=$1, url=$2, poll_interval=$3
|
||||||
|
WHERE id=$4
|
||||||
|
RETURNING name, url, poll_interval, clone_credential;`
|
||||||
|
|
||||||
|
pipeline, err := db.GetPipelineById(pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return pipeline, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var nameNew string
|
||||||
|
var urlNew string
|
||||||
|
var pollIntervalNew int
|
||||||
|
|
||||||
|
if name != nil {
|
||||||
|
nameNew = *name
|
||||||
|
} else {
|
||||||
|
nameNew = pipeline.Name
|
||||||
|
}
|
||||||
|
if url != nil {
|
||||||
|
urlNew = *url
|
||||||
|
} else {
|
||||||
|
urlNew = pipeline.Url
|
||||||
|
}
|
||||||
|
if pollInterval != nil {
|
||||||
|
pollIntervalNew = *pollInterval
|
||||||
|
} else {
|
||||||
|
pollIntervalNew = pipeline.PollInterval
|
||||||
|
}
|
||||||
|
|
||||||
|
err = db.Conn.QueryRow(context.Background(),
|
||||||
|
query, nameNew, urlNew, pollIntervalNew, pipelineId).Scan(
|
||||||
|
&pipeline.Name, &pipeline.Url, &pipeline.PollInterval, &pipeline.CloneCredential,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return pipeline, fmt.Errorf("Could not add credential to pipeline: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return pipeline, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) SetPipelineCloneCredential(pipelineId uuid.UUID, credentialId *uuid.UUID) (Pipeline, error) {
|
||||||
|
query := `
|
||||||
|
UPDATE pipelines
|
||||||
|
SET clone_credential=$1
|
||||||
|
WHERE id=$2
|
||||||
|
RETURNING name, url, poll_interval, clone_credential;`
|
||||||
|
|
||||||
|
pipeline := Pipeline{
|
||||||
|
Id: pipelineId,
|
||||||
|
}
|
||||||
|
|
||||||
|
err := db.Conn.QueryRow(context.Background(),
|
||||||
|
query, credentialId, pipelineId).Scan(
|
||||||
|
&pipeline.Name, &pipeline.Url, &pipeline.PollInterval, &pipeline.CloneCredential,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return pipeline, fmt.Errorf("Could not add credential to pipeline: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return pipeline, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) RemovePipelineCredential(pipelineId uuid.UUID) (Pipeline, error) {
|
||||||
|
query := `
|
||||||
|
UPDATE pipelines
|
||||||
|
SET credential=null
|
||||||
|
WHERE id=$1
|
||||||
|
RETURNING name, url, poll_interval, clone_credential;`
|
||||||
|
|
||||||
|
pipeline := Pipeline{
|
||||||
|
Id: pipelineId,
|
||||||
|
}
|
||||||
|
|
||||||
|
err := db.Conn.QueryRow(context.Background(),
|
||||||
|
query, pipelineId).Scan(
|
||||||
|
&pipeline.Name, &pipeline.Url, &pipeline.PollInterval, &pipeline.CloneCredential,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return pipeline, fmt.Errorf("Could not add credential to pipeline: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return pipeline, err
|
||||||
|
}
|
||||||
|
|
||||||
func (db *Database) GetWebhooksForPipeline(id uuid.UUID) ([]Webhook, error) {
|
func (db *Database) GetWebhooksForPipeline(id uuid.UUID) ([]Webhook, error) {
|
||||||
query := `
|
query := `
|
||||||
SELECT id, server_type, secret
|
SELECT id, server_type, secret
|
||||||
@@ -58,7 +174,7 @@ WHERE pipeline=$1;`
|
|||||||
|
|
||||||
rows, err := db.Conn.Query(context.Background(), query, id)
|
rows, err := db.Conn.Query(context.Background(), query, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal(err)
|
return webhooks, fmt.Errorf("Could not get webhooks for pipeline with id \"%v\": %w", id, err)
|
||||||
}
|
}
|
||||||
defer rows.Close()
|
defer rows.Close()
|
||||||
|
|
||||||
@@ -79,7 +195,7 @@ WHERE pipeline=$1;`
|
|||||||
return webhooks, nil
|
return webhooks, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Database) GetWebhookById(id uuid.UUID) (Webhook, error) {
|
func (db *Database) GetWebhookById(id uuid.UUID) (Webhook, error) {
|
||||||
query := `
|
query := `
|
||||||
SELECT server_type, secret, pipeline
|
SELECT server_type, secret, pipeline
|
||||||
FROM webhooks
|
FROM webhooks
|
||||||
@@ -89,7 +205,7 @@ WHERE id=$1;`
|
|||||||
Id: id,
|
Id: id,
|
||||||
}
|
}
|
||||||
|
|
||||||
err := d.Conn.QueryRow(context.Background(), query, id).Scan(&webhook.ServerType, &webhook.Secret, &webhook.Pipeline)
|
err := db.Conn.QueryRow(context.Background(), query, id).Scan(&webhook.ServerType, &webhook.Secret, &webhook.Pipeline)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return webhook, fmt.Errorf("Could not query database for webhook with id %v: %w", id.String(), err)
|
return webhook, fmt.Errorf("Could not query database for webhook with id %v: %w", id.String(), err)
|
||||||
}
|
}
|
||||||
@@ -97,9 +213,7 @@ WHERE id=$1;`
|
|||||||
return webhook, nil
|
return webhook, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Database) CreateWebhook(serverType WebhookSender, pipelineId uuid.UUID) (Webhook, error) {
|
func (db *Database) CreateWebhook(serverType WebhookSender, pipelineId uuid.UUID) (Webhook, error) {
|
||||||
|
|
||||||
//WITH secret_val as (select substr(md5(random()::text), 0, 50)),
|
|
||||||
|
|
||||||
query := `
|
query := `
|
||||||
INSERT INTO webhooks (id, server_type, secret, pipeline)
|
INSERT INTO webhooks (id, server_type, secret, pipeline)
|
||||||
@@ -108,7 +222,7 @@ RETURNING id, server_type, secret, pipeline;`
|
|||||||
|
|
||||||
webhook := Webhook{}
|
webhook := Webhook{}
|
||||||
var idStr string
|
var idStr string
|
||||||
err := d.Conn.QueryRow(context.Background(), query, string(serverType), pipelineId).Scan(&idStr, &webhook.ServerType, &webhook.Secret, &webhook.Pipeline)
|
err := db.Conn.QueryRow(context.Background(), query, string(serverType), pipelineId).Scan(&idStr, &webhook.ServerType, &webhook.Secret, &webhook.Pipeline)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return webhook, err
|
return webhook, err
|
||||||
}
|
}
|
||||||
@@ -122,3 +236,437 @@ RETURNING id, server_type, secret, pipeline;`
|
|||||||
|
|
||||||
return webhook, nil
|
return webhook, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (db *Database) CreateCredential(name string, credentialtype CloneCredentialType, username string, secret string) (CloneCredential, error) {
|
||||||
|
query := `
|
||||||
|
INSERT INTO clone_credentials (id, name, type, username, secret)
|
||||||
|
VALUES(uuid_generate_v4(), $1, $2, $3, $4)
|
||||||
|
RETURNING id, name, type, username, secret;`
|
||||||
|
|
||||||
|
credential := CloneCredential{}
|
||||||
|
var idStr string
|
||||||
|
err := db.Conn.QueryRow(
|
||||||
|
context.Background(),
|
||||||
|
query,
|
||||||
|
name,
|
||||||
|
string(credentialtype),
|
||||||
|
username,
|
||||||
|
secret,
|
||||||
|
).Scan(&idStr, &credential.Name, &credential.Type, &credential.Username, &credential.Secret)
|
||||||
|
if err != nil {
|
||||||
|
return credential, err
|
||||||
|
}
|
||||||
|
|
||||||
|
id, err := uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return credential, err
|
||||||
|
}
|
||||||
|
|
||||||
|
credential.Id = id
|
||||||
|
|
||||||
|
return credential, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetCloneCredentialById(id uuid.UUID) (CloneCredential, error) {
|
||||||
|
query := `
|
||||||
|
SELECT name, type, username, secret
|
||||||
|
FROM clone_credentials
|
||||||
|
WHERE id=$1;`
|
||||||
|
|
||||||
|
log.Debugf("requested credential with id %v", id)
|
||||||
|
|
||||||
|
credential := CloneCredential{
|
||||||
|
Id: id,
|
||||||
|
}
|
||||||
|
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, id).Scan(&credential.Name, &credential.Type, &credential.Username, &credential.Secret)
|
||||||
|
if err != nil {
|
||||||
|
return credential, fmt.Errorf("Could not query database for credential with id %v: %w", id.String(), err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return credential, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetCredentials() ([]CloneCredential, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, name, type, username, secret
|
||||||
|
FROM clone_credentials;`
|
||||||
|
|
||||||
|
credentials := make([]CloneCredential, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query)
|
||||||
|
if err != nil {
|
||||||
|
return credentials, fmt.Errorf("Could not query database for credentials: %w", err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var credential CloneCredential
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(&idStr, &credential.Name, &credential.Type, &credential.Username, &credential.Secret); err != nil {
|
||||||
|
return credentials, err
|
||||||
|
}
|
||||||
|
|
||||||
|
credential.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return credentials, err
|
||||||
|
}
|
||||||
|
credentials = append(credentials, credential)
|
||||||
|
}
|
||||||
|
|
||||||
|
return credentials, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) CreateRun(pipelineId uuid.UUID) (Run, error) {
|
||||||
|
query := `
|
||||||
|
INSERT INTO runs (id, pipeline, in_progress)
|
||||||
|
VALUES(uuid_generate_v4(), $1, true)
|
||||||
|
RETURNING id, pipeline, in_progress;`
|
||||||
|
|
||||||
|
run := Run{}
|
||||||
|
var idStr string
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, pipelineId).Scan(&idStr, &run.Pipeline, &run.InProgress)
|
||||||
|
if err != nil {
|
||||||
|
return run, err
|
||||||
|
}
|
||||||
|
|
||||||
|
run.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return run, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return run, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) UpdateRunBuildOutput(runId uuid.UUID, buildResult string) error {
|
||||||
|
query := `
|
||||||
|
UPDATE runs
|
||||||
|
SET build_output=$1
|
||||||
|
WHERE id=$2;`
|
||||||
|
|
||||||
|
_, err := db.Conn.Exec(context.Background(),
|
||||||
|
query, buildResult, runId)
|
||||||
|
|
||||||
|
return err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) UpdateRunResult(r Run) error {
|
||||||
|
query := `
|
||||||
|
UPDATE runs
|
||||||
|
SET in_progress=$1, result=$2, stdout=$3, stderr=$4
|
||||||
|
WHERE id=$5;`
|
||||||
|
|
||||||
|
// TODO: does r.Result need a pointer derefrence?
|
||||||
|
_, err := db.Conn.Exec(context.Background(),
|
||||||
|
query, r.InProgress, r.Result, r.Stdout, r.Stderr, r.Id)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetRunsForPipeline(pipelineId uuid.UUID) ([]Run, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, in_progress, result, build_output, stdout, stderr
|
||||||
|
FROM runs
|
||||||
|
WHERE pipeline=$1;`
|
||||||
|
|
||||||
|
runs := make([]Run, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query, pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return runs, fmt.Errorf("Could not get runs for pipeline with id \"%v\": %w", pipelineId, err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var run Run
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(
|
||||||
|
&idStr,
|
||||||
|
&run.InProgress,
|
||||||
|
&run.Result,
|
||||||
|
&run.BuildOutput,
|
||||||
|
&run.Stdout,
|
||||||
|
&run.Stderr,
|
||||||
|
); err != nil {
|
||||||
|
return runs, err
|
||||||
|
}
|
||||||
|
|
||||||
|
run.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return runs, err
|
||||||
|
}
|
||||||
|
runs = append(runs, run)
|
||||||
|
}
|
||||||
|
|
||||||
|
return runs, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetPipelineRefs(pipelineId uuid.UUID) (map[string]string, error) {
|
||||||
|
query := `
|
||||||
|
SELECT name, hash
|
||||||
|
FROM pipeline_refs
|
||||||
|
WHERE pipeline_id=$1;`
|
||||||
|
|
||||||
|
refsMap := make(map[string]string)
|
||||||
|
|
||||||
|
refs, err := db.Conn.Query(context.Background(), query, pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return refsMap, fmt.Errorf("Could not get pipeline refs for pipeline with id \"%v\": %w", pipelineId, err)
|
||||||
|
}
|
||||||
|
defer refs.Close()
|
||||||
|
|
||||||
|
for refs.Next() {
|
||||||
|
var name string
|
||||||
|
var hash string
|
||||||
|
if err := refs.Scan(
|
||||||
|
&name,
|
||||||
|
&hash,
|
||||||
|
); err != nil {
|
||||||
|
return refsMap, err
|
||||||
|
}
|
||||||
|
|
||||||
|
refsMap[name] = hash
|
||||||
|
}
|
||||||
|
|
||||||
|
return refsMap, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) UpdatePipelineRefs(pipelineId uuid.UUID, refsMap map[string]string) error {
|
||||||
|
|
||||||
|
query := `
|
||||||
|
INSERT INTO pipeline_refs(name, pipeline_id, hash)
|
||||||
|
VALUES($1, $2, $3)
|
||||||
|
ON CONFLICT (name)
|
||||||
|
DO
|
||||||
|
UPDATE SET hash=$3;`
|
||||||
|
|
||||||
|
for name, hash := range refsMap {
|
||||||
|
_, err := db.Conn.Exec(context.Background(), query, name, pipelineId, hash)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetSecrets() ([]Secret, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, name, secret
|
||||||
|
FROM secrets;`
|
||||||
|
|
||||||
|
secrets := make([]Secret, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query)
|
||||||
|
if err != nil {
|
||||||
|
return secrets, fmt.Errorf("Could not query database for secrets: %w", err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var secret Secret
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(&idStr, &secret.Name, &secret.Secret); err != nil {
|
||||||
|
return secrets, err
|
||||||
|
}
|
||||||
|
|
||||||
|
secret.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return secrets, err
|
||||||
|
}
|
||||||
|
secrets = append(secrets, secret)
|
||||||
|
}
|
||||||
|
|
||||||
|
return secrets, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetSecretById(id uuid.UUID) (Secret, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, name, secret
|
||||||
|
FROM secrets
|
||||||
|
WHERE id=$1;`
|
||||||
|
|
||||||
|
secret := Secret{
|
||||||
|
Id: id,
|
||||||
|
}
|
||||||
|
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, id).Scan(&secret.Name, &secret.Secret)
|
||||||
|
if err != nil {
|
||||||
|
return secret, fmt.Errorf("Could not query database for secret with id %v: %w", id.String(), err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return secret, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) CreateSecret(name string, secret string) (Secret, error) {
|
||||||
|
s := Secret{}
|
||||||
|
|
||||||
|
// validate that the secret is only A-Z or underscores and less than 256 characters
|
||||||
|
if len(name) > 256 {
|
||||||
|
return s, fmt.Errorf("secret name must be 256 characters or less")
|
||||||
|
}
|
||||||
|
|
||||||
|
validName := regexp.MustCompile(`[A-Z0-9_]+$`)
|
||||||
|
if !validName.MatchString(name) {
|
||||||
|
return s, fmt.Errorf("secren name must be made up of only uppercase letters, numbers, and underscores")
|
||||||
|
}
|
||||||
|
|
||||||
|
query := `
|
||||||
|
INSERT INTO secrets (id, name, secret)
|
||||||
|
VALUES (uuid_generate_v4(), $1, $2)
|
||||||
|
RETURNING id, name, secret;`
|
||||||
|
|
||||||
|
var idStr string
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, name, secret).Scan(&idStr, &s.Name, &s.Secret)
|
||||||
|
if err != nil {
|
||||||
|
return s, fmt.Errorf("Could not create secret: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return s, fmt.Errorf("Could not parse UUID generated by DB: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return s, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) AssignSecretToPipeline(pipelineId uuid.UUID, secretId uuid.UUID) error {
|
||||||
|
query := `
|
||||||
|
INSERT INTO pipeline_secret_mappings (pipeline, secret)
|
||||||
|
VALUES ($1, $2);`
|
||||||
|
|
||||||
|
_, err := db.Conn.Exec(context.Background(), query, pipelineId, secretId)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) RemoveSecretFromPipeline(pipelineId uuid.UUID, secretId uuid.UUID) error {
|
||||||
|
// TODO: implement this
|
||||||
|
return fmt.Errorf("Not implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetSecretsForPipeline(pipelineId uuid.UUID) ([]Secret, error) {
|
||||||
|
query := `
|
||||||
|
SELECT
|
||||||
|
secrets.id, secrets.name, secrets.secret
|
||||||
|
FROM
|
||||||
|
secrets INNER JOIN pipeline_secret_mappings
|
||||||
|
ON secrets.id = pipeline_secret_mappings.secret
|
||||||
|
WHERE
|
||||||
|
pipeline_secret_mappings.pipeline=$1
|
||||||
|
;`
|
||||||
|
|
||||||
|
secrets := make([]Secret, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query, pipelineId)
|
||||||
|
if err != nil {
|
||||||
|
return secrets, fmt.Errorf("Could not get secrets for pipeline with id \"%v\": %w", pipelineId, err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var secret Secret
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(
|
||||||
|
&idStr,
|
||||||
|
&secret.Name,
|
||||||
|
&secret.Secret,
|
||||||
|
); err != nil {
|
||||||
|
return secrets, err
|
||||||
|
}
|
||||||
|
|
||||||
|
secret.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return secrets, err
|
||||||
|
}
|
||||||
|
secrets = append(secrets, secret)
|
||||||
|
}
|
||||||
|
|
||||||
|
return secrets, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetRunners() ([]Runner, error) {
|
||||||
|
query := `
|
||||||
|
SELECT id, name, token
|
||||||
|
FROM runners;`
|
||||||
|
|
||||||
|
runners := make([]Runner, 0)
|
||||||
|
|
||||||
|
rows, err := db.Conn.Query(context.Background(), query)
|
||||||
|
if err != nil {
|
||||||
|
return runners, fmt.Errorf("Could not query database for runners: %w", err)
|
||||||
|
}
|
||||||
|
defer rows.Close()
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var runner Runner
|
||||||
|
var idStr string
|
||||||
|
if err := rows.Scan(&idStr, &runner.Name, &runner.Token); err != nil {
|
||||||
|
return runners, err
|
||||||
|
}
|
||||||
|
|
||||||
|
runner.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return runners, err
|
||||||
|
}
|
||||||
|
runners = append(runners, runner)
|
||||||
|
}
|
||||||
|
|
||||||
|
return runners, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) GetRunnerById(id uuid.UUID) (Runner, error) {
|
||||||
|
query := `
|
||||||
|
SELECT name, token
|
||||||
|
FROM runners
|
||||||
|
WHERE id=$1;`
|
||||||
|
|
||||||
|
runner := Runner{
|
||||||
|
Id: id,
|
||||||
|
}
|
||||||
|
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, id).Scan(&runner.Name, &runner.Token)
|
||||||
|
if err != nil {
|
||||||
|
return runner, fmt.Errorf("Could not query database for runner with id %v: %w", id.String(), err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return runner, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (db *Database) CreateRunner(name string) (Runner, error) {
|
||||||
|
s := Runner{}
|
||||||
|
|
||||||
|
// validate that the runner name is only A-Z or underscores and less than 256 characters
|
||||||
|
if len(name) > 256 {
|
||||||
|
return s, fmt.Errorf("runner name must be 256 characters or less")
|
||||||
|
}
|
||||||
|
|
||||||
|
validName := regexp.MustCompile(`[A-Z0-9_]+$`)
|
||||||
|
if !validName.MatchString(name) {
|
||||||
|
return s, fmt.Errorf("runner name must be made up of only uppercase letters, numbers, and underscores")
|
||||||
|
}
|
||||||
|
|
||||||
|
query := `
|
||||||
|
INSERT INTO runners (id, name, token)
|
||||||
|
VALUES
|
||||||
|
(
|
||||||
|
uuid_generate_v4(),
|
||||||
|
$1,
|
||||||
|
(
|
||||||
|
SELECT md5(random()::text)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
RETURNING id, name, token;`
|
||||||
|
|
||||||
|
var idStr string
|
||||||
|
err := db.Conn.QueryRow(context.Background(), query, name).Scan(&idStr, &s.Name, &s.Token)
|
||||||
|
if err != nil {
|
||||||
|
return s, fmt.Errorf("Could not create runner: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.Id, err = uuid.Parse(idStr)
|
||||||
|
if err != nil {
|
||||||
|
return s, fmt.Errorf("Could not parse UUID generated by DB: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return s, nil
|
||||||
|
}
|
||||||
|
|||||||
+39
-8
@@ -6,11 +6,38 @@ import (
|
|||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
type CloneCredentialType string
|
||||||
|
|
||||||
|
const (
|
||||||
|
USER_PASS CloneCredentialType = "USER_PASS"
|
||||||
|
SSH_KEY CloneCredentialType = "SSH_KEY"
|
||||||
|
)
|
||||||
|
|
||||||
|
type CloneCredential struct {
|
||||||
|
Id uuid.UUID
|
||||||
|
Name string
|
||||||
|
Type CloneCredentialType
|
||||||
|
Username string
|
||||||
|
Secret string
|
||||||
|
}
|
||||||
|
|
||||||
type Pipeline struct {
|
type Pipeline struct {
|
||||||
Id uuid.UUID
|
Id uuid.UUID
|
||||||
Name string
|
Name string
|
||||||
Url string
|
Url string
|
||||||
PollInterval int
|
PollInterval int
|
||||||
|
CloneCredential *uuid.UUID
|
||||||
|
}
|
||||||
|
|
||||||
|
type Secret struct {
|
||||||
|
Id uuid.UUID
|
||||||
|
Name string
|
||||||
|
Secret string
|
||||||
|
}
|
||||||
|
|
||||||
|
type PipelineSecretMapping struct {
|
||||||
|
Pipeline uuid.UUID
|
||||||
|
Secret uuid.UUID
|
||||||
}
|
}
|
||||||
|
|
||||||
type WebhookSender string
|
type WebhookSender string
|
||||||
@@ -26,25 +53,29 @@ type Webhook struct {
|
|||||||
Pipeline uuid.UUID
|
Pipeline uuid.UUID
|
||||||
}
|
}
|
||||||
|
|
||||||
type Runner struct {
|
|
||||||
Id uuid.UUID
|
|
||||||
Name string
|
|
||||||
Secret string
|
|
||||||
}
|
|
||||||
|
|
||||||
type Run struct {
|
type Run struct {
|
||||||
Id uuid.UUID
|
Id uuid.UUID
|
||||||
Pipeline uuid.UUID
|
Pipeline uuid.UUID
|
||||||
Result bool
|
InProgress bool
|
||||||
|
Result *int64
|
||||||
|
BuildOutput []byte
|
||||||
|
Stdout []byte
|
||||||
|
Stderr []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
type CommandExecution struct {
|
type CommandExecution struct {
|
||||||
Id uuid.UUID
|
Id uuid.UUID
|
||||||
RunId uuid.UUID
|
RunId uuid.UUID
|
||||||
Command string
|
Command []string
|
||||||
ReturnCode int
|
ReturnCode int
|
||||||
Stdout string
|
Stdout string
|
||||||
Stderr string
|
Stderr string
|
||||||
StartTime time.Time
|
StartTime time.Time
|
||||||
EndTime time.Time
|
EndTime time.Time
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type Runner struct {
|
||||||
|
Id uuid.UUID
|
||||||
|
Name string
|
||||||
|
Token string
|
||||||
|
}
|
||||||
|
|||||||
@@ -4,4 +4,4 @@ MAINTAINER restitux <restitux@ohea.xyz>
|
|||||||
RUN apt-get update && apt-get install -y \
|
RUN apt-get update && apt-get install -y \
|
||||||
ca-certificates \
|
ca-certificates \
|
||||||
&& rm -rf /var/lib/apt/lists/*
|
&& rm -rf /var/lib/apt/lists/*
|
||||||
ENTRYPOINT ["/build/server/docker/build-and-run.sh"]
|
ENTRYPOINT ["/build/server/docker/cursorius/build-and-run.sh"]
|
||||||
@@ -4,5 +4,5 @@
|
|||||||
set -e
|
set -e
|
||||||
|
|
||||||
cd /build/server
|
cd /build/server
|
||||||
go build .
|
go build -buildvcs=false .
|
||||||
./server
|
./server
|
||||||
@@ -2,15 +2,15 @@ version: "3.3"
|
|||||||
services:
|
services:
|
||||||
cursorius-server:
|
cursorius-server:
|
||||||
build:
|
build:
|
||||||
context: ..
|
context: ".."
|
||||||
dockerfile: docker/Dockerfile.dev
|
dockerfile: "docker/cursorius/Dockerfile.dev"
|
||||||
ports:
|
ports:
|
||||||
- "0.0.0.0:45420:45420"
|
- "0.0.0.0:45420:45420"
|
||||||
networks:
|
networks:
|
||||||
- cursorius
|
- cursorius
|
||||||
volumes:
|
volumes:
|
||||||
- "..:/build/server"
|
- "..:/build/server"
|
||||||
- "../server.toml:/root/.config/cursorius/server.toml"
|
- "./server.toml:/root/.config/cursorius/server.toml"
|
||||||
- "/var/run/docker.sock:/var/run/docker.sock"
|
- "/var/run/docker.sock:/var/run/docker.sock"
|
||||||
- "../_working/go:/go"
|
- "../_working/go:/go"
|
||||||
- "../_working/jobs:/cursorius/jobs"
|
- "../_working/jobs:/cursorius/jobs"
|
||||||
@@ -20,8 +20,31 @@ services:
|
|||||||
- POSTGRES_USER=cursorius
|
- POSTGRES_USER=cursorius
|
||||||
- POSTGRES_PASSWORD=cursorius
|
- POSTGRES_PASSWORD=cursorius
|
||||||
- POSTGRES_DB=cursorius
|
- POSTGRES_DB=cursorius
|
||||||
|
volumes:
|
||||||
|
- "../_working/postgres:/var/lib/postgresql/data"
|
||||||
networks:
|
networks:
|
||||||
- cursorius
|
- cursorius
|
||||||
|
graphiql:
|
||||||
|
build:
|
||||||
|
context: "graphiql"
|
||||||
|
dockerfile: "Dockerfile.graphiql"
|
||||||
|
ports:
|
||||||
|
- "0.0.0.0:45421:80"
|
||||||
|
networks:
|
||||||
|
- cursorius
|
||||||
|
gitea:
|
||||||
|
image: gitea/gitea:latest
|
||||||
|
profiles: ["gitea"]
|
||||||
|
environment:
|
||||||
|
- GITEA__webhook__ALLOWED_HOST_LIST=cursorius-server, external
|
||||||
|
ports:
|
||||||
|
- "127.0.0.1:2222:22"
|
||||||
|
- "127.0.0.1:3000:3000"
|
||||||
|
networks:
|
||||||
|
- cursorius
|
||||||
|
volumes:
|
||||||
|
- "../_working/gitea:/data"
|
||||||
|
|
||||||
networks:
|
networks:
|
||||||
cursorius:
|
cursorius:
|
||||||
|
external: true
|
||||||
|
|||||||
@@ -0,0 +1,3 @@
|
|||||||
|
FROM nginx:latest
|
||||||
|
COPY graphiql.html /usr/share/nginx/html/index.html
|
||||||
|
COPY graphiql.conf /etc/nginx/conf.d/default.conf
|
||||||
@@ -0,0 +1,52 @@
|
|||||||
|
upstream backend {
|
||||||
|
server cursorius-server:45420;
|
||||||
|
}
|
||||||
|
|
||||||
|
server {
|
||||||
|
listen 80;
|
||||||
|
listen [::]:80;
|
||||||
|
server_name localhost;
|
||||||
|
|
||||||
|
#access_log /var/log/nginx/host.access.log main;
|
||||||
|
|
||||||
|
location / {
|
||||||
|
root /usr/share/nginx/html;
|
||||||
|
index index.html index.htm;
|
||||||
|
}
|
||||||
|
|
||||||
|
location /graphql {
|
||||||
|
proxy_pass http://backend/graphql;
|
||||||
|
}
|
||||||
|
|
||||||
|
#error_page 404 /404.html;
|
||||||
|
|
||||||
|
# redirect server error pages to the static page /50x.html
|
||||||
|
#
|
||||||
|
error_page 500 502 503 504 /50x.html;
|
||||||
|
location = /50x.html {
|
||||||
|
root /usr/share/nginx/html;
|
||||||
|
}
|
||||||
|
|
||||||
|
# proxy the PHP scripts to Apache listening on 127.0.0.1:80
|
||||||
|
#
|
||||||
|
#location ~ \.php$ {
|
||||||
|
# proxy_pass http://127.0.0.1;
|
||||||
|
#}
|
||||||
|
|
||||||
|
# pass the PHP scripts to FastCGI server listening on 127.0.0.1:9000
|
||||||
|
#
|
||||||
|
#location ~ \.php$ {
|
||||||
|
# root html;
|
||||||
|
# fastcgi_pass 127.0.0.1:9000;
|
||||||
|
# fastcgi_index index.php;
|
||||||
|
# fastcgi_param SCRIPT_FILENAME /scripts$fastcgi_script_name;
|
||||||
|
# include fastcgi_params;
|
||||||
|
#}
|
||||||
|
|
||||||
|
# deny access to .htaccess files, if Apache's document root
|
||||||
|
# concurs with nginx's one
|
||||||
|
#
|
||||||
|
#location ~ /\.ht {
|
||||||
|
# deny all;
|
||||||
|
#}
|
||||||
|
}
|
||||||
@@ -0,0 +1,70 @@
|
|||||||
|
<!--
|
||||||
|
* Copyright (c) 2021 GraphQL Contributors
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under the license found in the
|
||||||
|
* LICENSE file in the root directory of this source tree.
|
||||||
|
-->
|
||||||
|
<!DOCTYPE html>
|
||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<title>GraphiQL</title>
|
||||||
|
<style>
|
||||||
|
body {
|
||||||
|
height: 100%;
|
||||||
|
margin: 0;
|
||||||
|
width: 100%;
|
||||||
|
overflow: hidden;
|
||||||
|
}
|
||||||
|
|
||||||
|
#graphiql {
|
||||||
|
height: 100vh;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
|
||||||
|
<!--
|
||||||
|
This GraphiQL example depends on Promise and fetch, which are available in
|
||||||
|
modern browsers, but can be "polyfilled" for older browsers.
|
||||||
|
GraphiQL itself depends on React DOM.
|
||||||
|
If you do not want to rely on a CDN, you can host these files locally or
|
||||||
|
include them directly in your favored resource bundler.
|
||||||
|
-->
|
||||||
|
<script
|
||||||
|
src="https://unpkg.com/react@17/umd/react.development.js"
|
||||||
|
integrity="sha512-Vf2xGDzpqUOEIKO+X2rgTLWPY+65++WPwCHkX2nFMu9IcstumPsf/uKKRd5prX3wOu8Q0GBylRpsDB26R6ExOg=="
|
||||||
|
crossorigin="anonymous"
|
||||||
|
></script>
|
||||||
|
<script
|
||||||
|
src="https://unpkg.com/react-dom@17/umd/react-dom.development.js"
|
||||||
|
integrity="sha512-Wr9OKCTtq1anK0hq5bY3X/AvDI5EflDSAh0mE9gma+4hl+kXdTJPKZ3TwLMBcrgUeoY0s3dq9JjhCQc7vddtFg=="
|
||||||
|
crossorigin="anonymous"
|
||||||
|
></script>
|
||||||
|
|
||||||
|
<!--
|
||||||
|
These two files can be found in the npm module, however you may wish to
|
||||||
|
copy them directly into your environment, or perhaps include them in your
|
||||||
|
favored resource bundler.
|
||||||
|
-->
|
||||||
|
<link rel="stylesheet" href="https://unpkg.com/graphiql/graphiql.min.css" />
|
||||||
|
</head>
|
||||||
|
|
||||||
|
<body>
|
||||||
|
<div id="graphiql">Loading...</div>
|
||||||
|
<script
|
||||||
|
src="https://unpkg.com/graphiql/graphiql.min.js"
|
||||||
|
type="application/javascript"
|
||||||
|
></script>
|
||||||
|
<script>
|
||||||
|
ReactDOM.render(
|
||||||
|
React.createElement(GraphiQL, {
|
||||||
|
fetcher: GraphiQL.createFetcher({
|
||||||
|
//url: 'https://swapi-graphql.netlify.app/.netlify/functions/index',
|
||||||
|
url: 'http://127.0.0.1:45421/graphql',
|
||||||
|
}),
|
||||||
|
defaultEditorToolsVisibility: true,
|
||||||
|
}),
|
||||||
|
document.getElementById('graphiql'),
|
||||||
|
);
|
||||||
|
</script>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
+1
-61
@@ -2,64 +2,4 @@
|
|||||||
|
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
mkdir -p _working/go
|
go run docker/run.go "$@"
|
||||||
mkdir -p _working/jobs
|
|
||||||
|
|
||||||
base_default_compose_files="docker/docker-compose.yml"
|
|
||||||
default_compose_files="$base_default_compose_files"
|
|
||||||
|
|
||||||
if [ -f "docker/docker-compose.override.yml" ]
|
|
||||||
then
|
|
||||||
default_compose_files+=" docker/docker-compose.override.yml"
|
|
||||||
else
|
|
||||||
default_compose_files="docker/docker-compose.yml"
|
|
||||||
fi
|
|
||||||
|
|
||||||
function stop_containers {
|
|
||||||
current_containers="$(cat _working/current_containers)"
|
|
||||||
if [ "$current_containers" == "default" ]
|
|
||||||
then
|
|
||||||
compose_files="$default_compose_files"
|
|
||||||
elif [ "$current_containers" == "webhook" ]
|
|
||||||
then
|
|
||||||
compose_files="$default_compose_files docker/webhook-override.yml"
|
|
||||||
fi
|
|
||||||
compose_file_flags=$(echo "$compose_files" | tr ' ' '\n' | xargs -I'{}' echo "-f {} " | tr -d '\n')
|
|
||||||
docker compose $compose_file_flags down
|
|
||||||
}
|
|
||||||
|
|
||||||
function show_logs {
|
|
||||||
current_containers="$(cat _working/current_containers)"
|
|
||||||
if [ "$current_containers" == "default" ]
|
|
||||||
then
|
|
||||||
compose_files="$default_compose_files"
|
|
||||||
elif [ "$current_containers" == "webhook" ]
|
|
||||||
then
|
|
||||||
compose_files="$default_compose_files docker/webhook-override.yml"
|
|
||||||
fi
|
|
||||||
compose_file_flags=$(echo "$compose_files" | tr ' ' '\n' | xargs -I'{}' echo "-f {} " | tr -d '\n')
|
|
||||||
docker compose $compose_file_flags logs -f
|
|
||||||
}
|
|
||||||
|
|
||||||
case $1 in
|
|
||||||
"default")
|
|
||||||
echo "default" > _working/current_containers
|
|
||||||
compose_files="$default_compose_files"
|
|
||||||
compose_file_flags=$(echo "$compose_files" | tr ' ' '\n' | xargs -I'{}' echo "-f {} " | tr -d '\n')
|
|
||||||
docker compose $compose_file_flags up --build -d
|
|
||||||
docker compose $compose_file_flags logs -f;;
|
|
||||||
"webhook")
|
|
||||||
echo "webhook" > _working/current_containers
|
|
||||||
stop_containers
|
|
||||||
compose_files="$default_compose_files docker/webhook-override.yml"
|
|
||||||
compose_file_flags=$(echo "$compose_files" | tr ' ' '\n' | xargs -I'{}' echo "-f {} " | tr -d '\n')
|
|
||||||
docker compose $compose_file_flags up --build -d
|
|
||||||
docker compose $compose_file_flags logs -f;;
|
|
||||||
"stop")
|
|
||||||
stop_containers;;
|
|
||||||
"logs")
|
|
||||||
show_logs;;
|
|
||||||
*) echo "ERROR: Unknown param \"$1\"" 2>&1
|
|
||||||
exit 255;;
|
|
||||||
esac
|
|
||||||
|
|
||||||
|
|||||||
@@ -0,0 +1,93 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
)
|
||||||
|
|
||||||
|
func panicError(errorString string, params ...any) {
|
||||||
|
fmt.Fprintf(os.Stderr, fmt.Sprintf("ERROR: %v\n", errorString), params...)
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func run(name string, arg ...string) {
|
||||||
|
cmd := exec.Command(name, arg...)
|
||||||
|
if err := cmd.Run(); err != nil {
|
||||||
|
panicError("could not run command %v: %v", name, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func runAttach(name string, arg ...string) {
|
||||||
|
cmd := exec.Command(name, arg...)
|
||||||
|
cmd.Stdin = os.Stdin
|
||||||
|
cmd.Stdout = os.Stdout
|
||||||
|
cmd.Stderr = os.Stderr
|
||||||
|
if err := cmd.Run(); err != nil {
|
||||||
|
panicError("could not run command %v: %v", name, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func runCompose(args []string) {
|
||||||
|
runAttach("docker", append([]string{"compose"}, args...)...)
|
||||||
|
}
|
||||||
|
|
||||||
|
func createDirs() {
|
||||||
|
run("mkdir", "-p", "_working/go")
|
||||||
|
run("mkdir", "-p", "_working/jobs")
|
||||||
|
}
|
||||||
|
|
||||||
|
func currentContainers() string {
|
||||||
|
bytes, err := os.ReadFile("_working/current_containers")
|
||||||
|
if err != nil {
|
||||||
|
panicError("could not read current containers: %v", err)
|
||||||
|
}
|
||||||
|
return string(bytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
func composeFlags() []string {
|
||||||
|
containers := currentContainers()
|
||||||
|
flags := []string{"-f", "docker/docker-compose.yml"}
|
||||||
|
switch containers {
|
||||||
|
case "gitea":
|
||||||
|
flags = append(flags, "--profile", "gitea")
|
||||||
|
}
|
||||||
|
return flags
|
||||||
|
}
|
||||||
|
|
||||||
|
func runContainers(containers string) {
|
||||||
|
err := os.WriteFile("_working/current_containers", []byte(containers), 0633)
|
||||||
|
if err != nil {
|
||||||
|
panicError("could not write current_containers file: %v", err)
|
||||||
|
}
|
||||||
|
runCompose(append(composeFlags(), "up", "--build", "-d"))
|
||||||
|
runCompose(append(composeFlags(), "logs", "-f"))
|
||||||
|
}
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
if len(os.Args) < 2 {
|
||||||
|
panicError("not enough arguments passed")
|
||||||
|
}
|
||||||
|
|
||||||
|
createDirs()
|
||||||
|
|
||||||
|
switch os.Args[1] {
|
||||||
|
case "default", "gitea":
|
||||||
|
runContainers(os.Args[1])
|
||||||
|
case "runprev":
|
||||||
|
runContainers(currentContainers())
|
||||||
|
case "stop":
|
||||||
|
runCompose(append(composeFlags(), "down"))
|
||||||
|
case "dbshell":
|
||||||
|
runCompose(append(composeFlags(), "exec", "cursorius-db", "psql", "--user=cursorius"))
|
||||||
|
case "logs":
|
||||||
|
runCompose(append(composeFlags(), "logs", "-f"))
|
||||||
|
case "ps":
|
||||||
|
runCompose(append(composeFlags(), "ps"))
|
||||||
|
case "help":
|
||||||
|
fmt.Println("commands: default, gitea, runprev, stop, dbshell, logs, ps, help")
|
||||||
|
default:
|
||||||
|
panicError("Unknown subcommand: %v", os.Args[1])
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,24 +0,0 @@
|
|||||||
version: "3.3"
|
|
||||||
services:
|
|
||||||
cursorius-server:
|
|
||||||
networks:
|
|
||||||
- gitea
|
|
||||||
gitea:
|
|
||||||
image: gitea/gitea:latest
|
|
||||||
environment:
|
|
||||||
- GITEA__webhook__ALLOWED_HOST_LIST=cursorius-server, external
|
|
||||||
ports:
|
|
||||||
- "127.0.0.1:2222:22"
|
|
||||||
- "127.0.0.1:3000:3000"
|
|
||||||
networks:
|
|
||||||
- gitea
|
|
||||||
volumes:
|
|
||||||
- gitea-data:/data
|
|
||||||
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
gitea-data:
|
|
||||||
|
|
||||||
networks:
|
|
||||||
gitea:
|
|
||||||
external: false
|
|
||||||
@@ -3,7 +3,7 @@ module git.ohea.xyz/cursorius/server
|
|||||||
go 1.19
|
go 1.19
|
||||||
|
|
||||||
require (
|
require (
|
||||||
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230109075652-ead0aeff2eb9
|
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230405234139-34d8875b72f4
|
||||||
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2
|
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2
|
||||||
git.ohea.xyz/golang/config v0.0.0-20220915224621-b9debd233173
|
git.ohea.xyz/golang/config v0.0.0-20220915224621-b9debd233173
|
||||||
github.com/bufbuild/connect-go v1.4.1
|
github.com/bufbuild/connect-go v1.4.1
|
||||||
@@ -14,9 +14,10 @@ require (
|
|||||||
github.com/graphql-go/graphql v0.8.0
|
github.com/graphql-go/graphql v0.8.0
|
||||||
github.com/graphql-go/handler v0.2.3
|
github.com/graphql-go/handler v0.2.3
|
||||||
github.com/jackc/pgx/v5 v5.2.0
|
github.com/jackc/pgx/v5 v5.2.0
|
||||||
|
github.com/jhoonb/archivex v0.0.0-20201016144719-6a343cdae81d
|
||||||
github.com/op/go-logging v0.0.0-20160315200505-970db520ece7
|
github.com/op/go-logging v0.0.0-20160315200505-970db520ece7
|
||||||
golang.org/x/net v0.2.0
|
golang.org/x/net v0.2.0
|
||||||
google.golang.org/protobuf v1.28.1
|
google.golang.org/protobuf v1.30.0
|
||||||
nhooyr.io/websocket v1.8.7
|
nhooyr.io/websocket v1.8.7
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -37,6 +38,7 @@ require (
|
|||||||
github.com/imdario/mergo v0.3.13 // indirect
|
github.com/imdario/mergo v0.3.13 // indirect
|
||||||
github.com/jackc/pgpassfile v1.0.0 // indirect
|
github.com/jackc/pgpassfile v1.0.0 // indirect
|
||||||
github.com/jackc/pgservicefile v0.0.0-20200714003250-2b9c44734f2b // indirect
|
github.com/jackc/pgservicefile v0.0.0-20200714003250-2b9c44734f2b // indirect
|
||||||
|
github.com/jackc/puddle/v2 v2.1.2 // indirect
|
||||||
github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99 // indirect
|
github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99 // indirect
|
||||||
github.com/json-iterator/go v1.1.12 // indirect
|
github.com/json-iterator/go v1.1.12 // indirect
|
||||||
github.com/kevinburke/ssh_config v1.2.0 // indirect
|
github.com/kevinburke/ssh_config v1.2.0 // indirect
|
||||||
@@ -53,8 +55,10 @@ require (
|
|||||||
github.com/sirupsen/logrus v1.9.0 // indirect
|
github.com/sirupsen/logrus v1.9.0 // indirect
|
||||||
github.com/stretchr/testify v1.8.1 // indirect
|
github.com/stretchr/testify v1.8.1 // indirect
|
||||||
github.com/xanzy/ssh-agent v0.3.2 // indirect
|
github.com/xanzy/ssh-agent v0.3.2 // indirect
|
||||||
|
go.uber.org/atomic v1.10.0 // indirect
|
||||||
golang.org/x/crypto v0.2.1-0.20221112162523-6fad3dfc1891 // indirect
|
golang.org/x/crypto v0.2.1-0.20221112162523-6fad3dfc1891 // indirect
|
||||||
golang.org/x/mod v0.6.0-dev.0.20220419223038-86c51ed26bb4 // indirect
|
golang.org/x/mod v0.6.0-dev.0.20220419223038-86c51ed26bb4 // indirect
|
||||||
|
golang.org/x/sync v0.0.0-20220923202941-7f9b1623fab7 // indirect
|
||||||
golang.org/x/sys v0.2.0 // indirect
|
golang.org/x/sys v0.2.0 // indirect
|
||||||
golang.org/x/text v0.4.0 // indirect
|
golang.org/x/text v0.4.0 // indirect
|
||||||
golang.org/x/tools v0.1.12 // indirect
|
golang.org/x/tools v0.1.12 // indirect
|
||||||
|
|||||||
@@ -78,8 +78,8 @@ contrib.go.opencensus.io/exporter/stackdriver v0.13.5/go.mod h1:aXENhDJ1Y4lIg4EU
|
|||||||
contrib.go.opencensus.io/integrations/ocsql v0.1.4/go.mod h1:8DsSdjz3F+APR+0z0WkU1aRorQCFfRxvqjUUPMbF3fE=
|
contrib.go.opencensus.io/integrations/ocsql v0.1.4/go.mod h1:8DsSdjz3F+APR+0z0WkU1aRorQCFfRxvqjUUPMbF3fE=
|
||||||
contrib.go.opencensus.io/resource v0.1.1/go.mod h1:F361eGI91LCmW1I/Saf+rX0+OFcigGlFvXwEGEnkRLA=
|
contrib.go.opencensus.io/resource v0.1.1/go.mod h1:F361eGI91LCmW1I/Saf+rX0+OFcigGlFvXwEGEnkRLA=
|
||||||
dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU=
|
dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU=
|
||||||
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230109075652-ead0aeff2eb9 h1:8p7Kw3B7dbi2zdgG+Me9ETRWrJzoNVjcase4YqXfGbs=
|
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230405234139-34d8875b72f4 h1:kKQQEg1nmWnqiNOqtUHteEuacyfy0NdxyDj6HPjbA2c=
|
||||||
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230109075652-ead0aeff2eb9/go.mod h1:D7GGcFIH421mo6KuRaXXXmlXPwWeEsemTZG/BOZA/4o=
|
git.ohea.xyz/cursorius/pipeline-api/go/api/v2 v2.0.0-20230405234139-34d8875b72f4/go.mod h1:D7GGcFIH421mo6KuRaXXXmlXPwWeEsemTZG/BOZA/4o=
|
||||||
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2 h1:G1XQEqhj1LZPQbH7avzvT7QL9Wfbb4CXMm0nLL39eDc=
|
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2 h1:G1XQEqhj1LZPQbH7avzvT7QL9Wfbb4CXMm0nLL39eDc=
|
||||||
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2/go.mod h1:F9y5Ck4Wchsaj5amSX2eDRUlQ/iYP1VNLFduvjNwmLc=
|
git.ohea.xyz/cursorius/runner-api/go/api/v2 v2.0.0-20230109074922-e20285fe6cf2/go.mod h1:F9y5Ck4Wchsaj5amSX2eDRUlQ/iYP1VNLFduvjNwmLc=
|
||||||
git.ohea.xyz/cursorius/webhooks/v6 v6.0.2-0.20221224221147-a2bdbf1756ed h1:gsK15m4Npow74+R6OfZKwwAg1sl7QWQCRXOeE2QLUco=
|
git.ohea.xyz/cursorius/webhooks/v6 v6.0.2-0.20221224221147-a2bdbf1756ed h1:gsK15m4Npow74+R6OfZKwwAg1sl7QWQCRXOeE2QLUco=
|
||||||
@@ -849,6 +849,8 @@ github.com/jackc/puddle v1.1.0/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dv
|
|||||||
github.com/jackc/puddle v1.1.1/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
github.com/jackc/puddle v1.1.1/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
||||||
github.com/jackc/puddle v1.1.3/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
github.com/jackc/puddle v1.1.3/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
||||||
github.com/jackc/puddle v1.2.1/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
github.com/jackc/puddle v1.2.1/go.mod h1:m4B5Dj62Y0fbyuIc15OsIqK0+JU8nkqQjsgx7dvjSWk=
|
||||||
|
github.com/jackc/puddle/v2 v2.1.2 h1:0f7vaaXINONKTsxYDn4otOAiJanX/BMeAtY//BXqzlg=
|
||||||
|
github.com/jackc/puddle/v2 v2.1.2/go.mod h1:2lpufsF5mRHO6SuZkm0fNYxM6SWHfvyFj62KwNzgels=
|
||||||
github.com/jarcoal/httpmock v0.0.0-20180424175123-9c70cfe4a1da/go.mod h1:ks+b9deReOc7jgqp+e7LuFiCBH6Rm5hL32cLcEAArb4=
|
github.com/jarcoal/httpmock v0.0.0-20180424175123-9c70cfe4a1da/go.mod h1:ks+b9deReOc7jgqp+e7LuFiCBH6Rm5hL32cLcEAArb4=
|
||||||
github.com/jarcoal/httpmock v1.0.5/go.mod h1:ATjnClrvW/3tijVmpL/va5Z3aAyGvqU3gCT8nX0Txik=
|
github.com/jarcoal/httpmock v1.0.5/go.mod h1:ATjnClrvW/3tijVmpL/va5Z3aAyGvqU3gCT8nX0Txik=
|
||||||
github.com/jaytaylor/html2text v0.0.0-20211105163654-bc68cce691ba/go.mod h1:CVKlgaMiht+LXvHG173ujK6JUhZXKb2u/BQtjPDIvyk=
|
github.com/jaytaylor/html2text v0.0.0-20211105163654-bc68cce691ba/go.mod h1:CVKlgaMiht+LXvHG173ujK6JUhZXKb2u/BQtjPDIvyk=
|
||||||
@@ -856,6 +858,8 @@ github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99 h1:BQSFePA1RWJOl
|
|||||||
github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99/go.mod h1:1lJo3i6rXxKeerYnT8Nvf0QmHCRC1n8sfWVwXF2Frvo=
|
github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99/go.mod h1:1lJo3i6rXxKeerYnT8Nvf0QmHCRC1n8sfWVwXF2Frvo=
|
||||||
github.com/jessevdk/go-flags v1.4.0/go.mod h1:4FA24M0QyGHXBuZZK/XkWh8h0e1EYbRYJSGM75WSRxI=
|
github.com/jessevdk/go-flags v1.4.0/go.mod h1:4FA24M0QyGHXBuZZK/XkWh8h0e1EYbRYJSGM75WSRxI=
|
||||||
github.com/jessevdk/go-flags v1.5.0/go.mod h1:Fw0T6WPc1dYxT4mKEZRfG5kJhaTDP9pj1c2EWnYs/m4=
|
github.com/jessevdk/go-flags v1.5.0/go.mod h1:Fw0T6WPc1dYxT4mKEZRfG5kJhaTDP9pj1c2EWnYs/m4=
|
||||||
|
github.com/jhoonb/archivex v0.0.0-20201016144719-6a343cdae81d h1:q7n+5taxmM+9T2Q7Ydo7YN90FkoDuR5bbzByZwkQqPo=
|
||||||
|
github.com/jhoonb/archivex v0.0.0-20201016144719-6a343cdae81d/go.mod h1:GN1Mg/uXQ6qwXA0HypnUO3xlcQJS9/y68EsHNeuuRa4=
|
||||||
github.com/jhump/protoreflect v1.6.1/go.mod h1:RZQ/lnuN+zqeRVpQigTwO6o0AJUkxbnSnpuG7toUTG4=
|
github.com/jhump/protoreflect v1.6.1/go.mod h1:RZQ/lnuN+zqeRVpQigTwO6o0AJUkxbnSnpuG7toUTG4=
|
||||||
github.com/jhump/protoreflect v1.8.2/go.mod h1:7GcYQDdMU/O/BBrl/cX6PNHpXh6cenjd8pneu5yW7Tg=
|
github.com/jhump/protoreflect v1.8.2/go.mod h1:7GcYQDdMU/O/BBrl/cX6PNHpXh6cenjd8pneu5yW7Tg=
|
||||||
github.com/jmespath/go-jmespath v0.0.0-20160202185014-0b12d6b521d8/go.mod h1:Nht3zPeWKUH0NzdCt2Blrr5ys8VGpn0CEB0cQHVjt7k=
|
github.com/jmespath/go-jmespath v0.0.0-20160202185014-0b12d6b521d8/go.mod h1:Nht3zPeWKUH0NzdCt2Blrr5ys8VGpn0CEB0cQHVjt7k=
|
||||||
@@ -1443,6 +1447,7 @@ go.uber.org/atomic v1.4.0/go.mod h1:gD2HeocX3+yG+ygLZcrzQJaqmWj9AIm7n08wl/qW/PE=
|
|||||||
go.uber.org/atomic v1.5.0/go.mod h1:sABNBOSYdrvTF6hTgEIbc7YasKWGhgEQZyfxyTvoXHQ=
|
go.uber.org/atomic v1.5.0/go.mod h1:sABNBOSYdrvTF6hTgEIbc7YasKWGhgEQZyfxyTvoXHQ=
|
||||||
go.uber.org/atomic v1.6.0/go.mod h1:sABNBOSYdrvTF6hTgEIbc7YasKWGhgEQZyfxyTvoXHQ=
|
go.uber.org/atomic v1.6.0/go.mod h1:sABNBOSYdrvTF6hTgEIbc7YasKWGhgEQZyfxyTvoXHQ=
|
||||||
go.uber.org/atomic v1.7.0/go.mod h1:fEN4uk6kAWBTFdckzkM89CLk9XfWZrxpCo0nPH17wJc=
|
go.uber.org/atomic v1.7.0/go.mod h1:fEN4uk6kAWBTFdckzkM89CLk9XfWZrxpCo0nPH17wJc=
|
||||||
|
go.uber.org/atomic v1.10.0 h1:9qC72Qh0+3MqyJbAn8YU5xVq1frD8bn3JtD2oXtafVQ=
|
||||||
go.uber.org/atomic v1.10.0/go.mod h1:LUxbIzbOniOlMKjJjyPfpl4v+PKK2cNJn91OQbhoJI0=
|
go.uber.org/atomic v1.10.0/go.mod h1:LUxbIzbOniOlMKjJjyPfpl4v+PKK2cNJn91OQbhoJI0=
|
||||||
go.uber.org/goleak v1.1.11/go.mod h1:cwTWslyiVhfpKIDGSZEM2HlOvcqm+tG4zioyIeLoqMQ=
|
go.uber.org/goleak v1.1.11/go.mod h1:cwTWslyiVhfpKIDGSZEM2HlOvcqm+tG4zioyIeLoqMQ=
|
||||||
go.uber.org/multierr v1.1.0/go.mod h1:wR5kodmAFQ0UK8QlbwjlSNy0Z68gJhDJUG5sjR94q/0=
|
go.uber.org/multierr v1.1.0/go.mod h1:wR5kodmAFQ0UK8QlbwjlSNy0Z68gJhDJUG5sjR94q/0=
|
||||||
@@ -1670,6 +1675,8 @@ golang.org/x/sync v0.0.0-20210220032951-036812b2e83c/go.mod h1:RxMgew5VJxzue5/jJ
|
|||||||
golang.org/x/sync v0.0.0-20220513210516-0976fa681c29/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20220513210516-0976fa681c29/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.0.0-20220601150217-0de741cfad7f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20220601150217-0de741cfad7f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.0.0-20220722155255-886fb9371eb4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20220722155255-886fb9371eb4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
|
golang.org/x/sync v0.0.0-20220923202941-7f9b1623fab7 h1:ZrnxWX62AgTKOSagEqxvb3ffipvEDX2pl7E1TdqLqIc=
|
||||||
|
golang.org/x/sync v0.0.0-20220923202941-7f9b1623fab7/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sys v0.0.0-20180823144017-11551d06cbcc/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
golang.org/x/sys v0.0.0-20180823144017-11551d06cbcc/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||||
golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||||
golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||||
@@ -2122,8 +2129,9 @@ google.golang.org/protobuf v1.26.0-rc.1/go.mod h1:jlhhOSvTdKEhbULTjvd4ARK9grFBp0
|
|||||||
google.golang.org/protobuf v1.26.0/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc=
|
google.golang.org/protobuf v1.26.0/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc=
|
||||||
google.golang.org/protobuf v1.27.1/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc=
|
google.golang.org/protobuf v1.27.1/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc=
|
||||||
google.golang.org/protobuf v1.28.0/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I=
|
google.golang.org/protobuf v1.28.0/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I=
|
||||||
google.golang.org/protobuf v1.28.1 h1:d0NfwRgPtno5B1Wa6L2DAG+KivqkdutMf1UhdNx175w=
|
|
||||||
google.golang.org/protobuf v1.28.1/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I=
|
google.golang.org/protobuf v1.28.1/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I=
|
||||||
|
google.golang.org/protobuf v1.30.0 h1:kPPoIgf3TsEvrm0PFe15JQ+570QVxYzEvvHqChK+cng=
|
||||||
|
google.golang.org/protobuf v1.30.0/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I=
|
||||||
gopkg.in/alecthomas/kingpin.v2 v2.2.6/go.mod h1:FMv+mEhP44yOT+4EoQTLFTRgOQ1FBLkstjWtayDeSgw=
|
gopkg.in/alecthomas/kingpin.v2 v2.2.6/go.mod h1:FMv+mEhP44yOT+4EoQTLFTRgOQ1FBLkstjWtayDeSgw=
|
||||||
gopkg.in/alexcesaro/quotedprintable.v3 v3.0.0-20150716171945-2caba252f4dc/go.mod h1:m7x9LTH6d71AHyAX77c9yqWCCa3UKHcVEj9y7hAtKDk=
|
gopkg.in/alexcesaro/quotedprintable.v3 v3.0.0-20150716171945-2caba252f4dc/go.mod h1:m7x9LTH6d71AHyAX77c9yqWCCa3UKHcVEj9y7hAtKDk=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||||
|
|||||||
+14
-12
@@ -10,6 +10,8 @@ import (
|
|||||||
"git.ohea.xyz/cursorius/server/pipeline_api"
|
"git.ohea.xyz/cursorius/server/pipeline_api"
|
||||||
"git.ohea.xyz/cursorius/server/runnermanager"
|
"git.ohea.xyz/cursorius/server/runnermanager"
|
||||||
"git.ohea.xyz/cursorius/server/webhook"
|
"git.ohea.xyz/cursorius/server/webhook"
|
||||||
|
|
||||||
|
"github.com/google/uuid"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
"golang.org/x/net/http2"
|
"golang.org/x/net/http2"
|
||||||
"golang.org/x/net/http2/h2c"
|
"golang.org/x/net/http2/h2c"
|
||||||
@@ -20,17 +22,17 @@ var log = logging.MustGetLogger("cursorius-server")
|
|||||||
|
|
||||||
func setupHTTPServer(
|
func setupHTTPServer(
|
||||||
mux *http.ServeMux,
|
mux *http.ServeMux,
|
||||||
conf config.Config,
|
conf config.PipelineConf,
|
||||||
db database.Database,
|
db database.Database,
|
||||||
registerCh chan runnermanager.RunnerRegistration,
|
runnerManagerChans runnermanager.RunnerManagerChans,
|
||||||
getRunnerCh chan runnermanager.GetRunnerRequest,
|
pollChan chan uuid.UUID,
|
||||||
) error {
|
) error {
|
||||||
|
|
||||||
webhook.CreateWebhookHandler(conf, mux)
|
webhook.CreateWebhookHandler(db, conf, mux)
|
||||||
|
|
||||||
pipeline_api.CreateHandler(getRunnerCh, mux)
|
pipeline_api.CreateHandler(runnerManagerChans.Allocation, runnerManagerChans.Release, mux)
|
||||||
|
|
||||||
err := admin_api.CreateHandler(db, mux)
|
err := admin_api.CreateHandler(db, pollChan, mux)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Could not create admin api handler: %w", err)
|
return fmt.Errorf("Could not create admin api handler: %w", err)
|
||||||
}
|
}
|
||||||
@@ -41,7 +43,7 @@ func setupHTTPServer(
|
|||||||
log.Errorf("Could not upgrade runner connection to websocket: %v", err)
|
log.Errorf("Could not upgrade runner connection to websocket: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
go runnermanager.RegisterRunner(conn, registerCh)
|
go runnermanager.RegisterRunner(conn, runnerManagerChans.Registration)
|
||||||
})
|
})
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -50,18 +52,18 @@ func Listen(
|
|||||||
mux *http.ServeMux,
|
mux *http.ServeMux,
|
||||||
address string,
|
address string,
|
||||||
port int,
|
port int,
|
||||||
conf config.Config,
|
conf config.PipelineConf,
|
||||||
db database.Database,
|
db database.Database,
|
||||||
registerCh chan runnermanager.RunnerRegistration,
|
runnerManagerChans runnermanager.RunnerManagerChans,
|
||||||
getRunnerCh chan runnermanager.GetRunnerRequest,
|
pollChan chan uuid.UUID,
|
||||||
) error {
|
) error {
|
||||||
|
|
||||||
err := setupHTTPServer(
|
err := setupHTTPServer(
|
||||||
mux,
|
mux,
|
||||||
conf,
|
conf,
|
||||||
db,
|
db,
|
||||||
registerCh,
|
runnerManagerChans,
|
||||||
getRunnerCh,
|
pollChan,
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Could not setup http endpoints: %w", err)
|
return fmt.Errorf("Could not setup http endpoints: %w", err)
|
||||||
|
|||||||
@@ -40,13 +40,13 @@ func main() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
getRunnerCh, registerCh, err := runnermanager.StartRunnerManager(configData.Config.Runners)
|
runnerManagerChans, err := runnermanager.StartRunnerManager(configData.Config.Runners, db)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("Could not start runner: %v", err)
|
log.Errorf("Could not start runner: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
poll.StartPolling(configData.Config)
|
pollChan := poll.StartPolling(configData.Config.PipelineConf, db)
|
||||||
|
|
||||||
mux := http.NewServeMux()
|
mux := http.NewServeMux()
|
||||||
|
|
||||||
@@ -54,9 +54,9 @@ func main() {
|
|||||||
mux,
|
mux,
|
||||||
configData.Config.Address,
|
configData.Config.Address,
|
||||||
configData.Config.Port,
|
configData.Config.Port,
|
||||||
configData.Config,
|
configData.Config.PipelineConf,
|
||||||
db,
|
db,
|
||||||
registerCh,
|
runnerManagerChans,
|
||||||
getRunnerCh,
|
pollChan,
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,12 +4,13 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
apiv2 "git.ohea.xyz/cursorius/pipeline-api/go/api/v2"
|
apiv2 "git.ohea.xyz/cursorius/pipeline-api/go/api/v2"
|
||||||
"git.ohea.xyz/cursorius/pipeline-api/go/api/v2/apiv2connect"
|
"git.ohea.xyz/cursorius/pipeline-api/go/api/v2/apiv2connect"
|
||||||
"git.ohea.xyz/cursorius/server/runnermanager"
|
"git.ohea.xyz/cursorius/server/runnermanager"
|
||||||
|
"git.ohea.xyz/cursorius/server/util"
|
||||||
"github.com/bufbuild/connect-go"
|
"github.com/bufbuild/connect-go"
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
@@ -18,7 +19,8 @@ import (
|
|||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
type ApiServer struct {
|
type ApiServer struct {
|
||||||
getRunnerCh chan runnermanager.GetRunnerRequest
|
allocationCh chan runnermanager.RunnerAllocationRequest
|
||||||
|
releaseCh chan runnermanager.RunnerReleaseRequest
|
||||||
allocatedRunners map[uuid.UUID]*RunnerWrapper
|
allocatedRunners map[uuid.UUID]*RunnerWrapper
|
||||||
allocatedRunnersMutex sync.RWMutex
|
allocatedRunnersMutex sync.RWMutex
|
||||||
}
|
}
|
||||||
@@ -34,15 +36,17 @@ func (r *RunnerWrapper) RunCommand(cmd string, args []string) (int64, string, st
|
|||||||
|
|
||||||
return_code, stdout, stderr, err := r.runner.RunCommand(cmd, args)
|
return_code, stdout, stderr, err := r.runner.RunCommand(cmd, args)
|
||||||
|
|
||||||
// TODO: run command by sending websocket packet
|
|
||||||
// TODO: get stdout and stderr response
|
|
||||||
return return_code, stdout, stderr, err
|
return return_code, stdout, stderr, err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RunnerWrapper) Release() {
|
func (r *RunnerWrapper) Release(releaseCh chan runnermanager.RunnerReleaseRequest) {
|
||||||
r.mutex.Lock()
|
r.mutex.Lock()
|
||||||
defer r.mutex.Unlock()
|
defer r.mutex.Unlock()
|
||||||
r.runner.Release()
|
|
||||||
|
releaseCh <- runnermanager.RunnerReleaseRequest{
|
||||||
|
Runner: r.runner,
|
||||||
|
}
|
||||||
|
r.runner = nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *ApiServer) GetRunnerFromMap(u uuid.UUID) (*RunnerWrapper, bool) {
|
func (s *ApiServer) GetRunnerFromMap(u uuid.UUID) (*RunnerWrapper, bool) {
|
||||||
@@ -52,37 +56,71 @@ func (s *ApiServer) GetRunnerFromMap(u uuid.UUID) (*RunnerWrapper, bool) {
|
|||||||
return runner, ok
|
return runner, ok
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (s *ApiServer) AddRunnerToMap(u uuid.UUID, runner *runnermanager.Runner) {
|
||||||
|
s.allocatedRunnersMutex.Lock()
|
||||||
|
defer s.allocatedRunnersMutex.Unlock()
|
||||||
|
s.allocatedRunners[u] = &RunnerWrapper{runner: runner}
|
||||||
|
}
|
||||||
|
|
||||||
func (s *ApiServer) GetRunner(
|
func (s *ApiServer) GetRunner(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
req *connect.Request[apiv2.GetRunnerRequest],
|
req *connect.Request[apiv2.GetRunnerRequest],
|
||||||
) (*connect.Response[apiv2.GetRunnerResponse], error) {
|
) (*connect.Response[apiv2.GetRunnerResponse], error) {
|
||||||
|
|
||||||
respChan := make(chan runnermanager.GetRunnerResponse)
|
var response runnermanager.RunnerAllocationResponse
|
||||||
s.getRunnerCh <- runnermanager.GetRunnerRequest{
|
var timeoutCtx *context.Context
|
||||||
|
var retryInterval int64 = 0
|
||||||
|
|
||||||
|
respChan := make(chan runnermanager.RunnerAllocationResponse)
|
||||||
|
|
||||||
|
tagsStr := util.FormatTags(req.Msg.Tags)
|
||||||
|
|
||||||
|
if req.Msg.Options != nil {
|
||||||
|
if req.Msg.Options.Timeout != 0 {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Duration(req.Msg.Options.Timeout)*time.Second)
|
||||||
|
timeoutCtx = &ctx
|
||||||
|
defer cancel()
|
||||||
|
}
|
||||||
|
|
||||||
|
retryInterval = req.Msg.Options.RetryInterval
|
||||||
|
}
|
||||||
|
|
||||||
|
for {
|
||||||
|
s.allocationCh <- runnermanager.RunnerAllocationRequest{
|
||||||
Tags: req.Msg.Tags,
|
Tags: req.Msg.Tags,
|
||||||
RespChan: respChan,
|
RespChan: respChan,
|
||||||
}
|
}
|
||||||
|
|
||||||
var runnerTagsStr strings.Builder
|
response = <-respChan
|
||||||
fmt.Fprintf(&runnerTagsStr, "[%v", req.Msg.Tags[0])
|
if response.Err == nil {
|
||||||
for _, tag := range req.Msg.Tags[1:] {
|
break
|
||||||
fmt.Fprintf(&runnerTagsStr, ", %v", tag)
|
|
||||||
}
|
}
|
||||||
fmt.Fprintf(&runnerTagsStr, "]")
|
|
||||||
|
|
||||||
response := <-respChan
|
log.Infof("Could not get runner with tags \"%v\": %v", tagsStr, response.Err)
|
||||||
if response.Err != nil {
|
|
||||||
log.Errorf("Could not get runner with tags \"%v\": %v", runnerTagsStr.String(), response.Err)
|
// If no timeout is specified, skip after one attempt
|
||||||
|
if timeoutCtx == nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
// If timeout is expired, stop trying to allocate runner
|
||||||
|
if (*timeoutCtx).Err() != nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Infof("Sleeping for %v seconds before retry...", retryInterval)
|
||||||
|
time.Sleep(time.Duration(retryInterval) * time.Second)
|
||||||
|
}
|
||||||
|
|
||||||
|
if response.Runner == nil {
|
||||||
return nil, connect.NewError(connect.CodeNotFound, fmt.Errorf("Could not get runner"))
|
return nil, connect.NewError(connect.CodeNotFound, fmt.Errorf("Could not get runner"))
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Infof("Got runner with tags: %v", runnerTagsStr.String())
|
log.Infof("Got runner with tags: %v", tagsStr)
|
||||||
|
|
||||||
runnerUuid := uuid.New()
|
runnerUuid := uuid.New()
|
||||||
|
|
||||||
s.allocatedRunnersMutex.Lock()
|
s.AddRunnerToMap(runnerUuid, response.Runner)
|
||||||
s.allocatedRunners[runnerUuid] = &RunnerWrapper{runner: response.Runner}
|
|
||||||
s.allocatedRunnersMutex.Unlock()
|
|
||||||
|
|
||||||
res := connect.NewResponse(&apiv2.GetRunnerResponse{
|
res := connect.NewResponse(&apiv2.GetRunnerResponse{
|
||||||
Id: runnerUuid.String(),
|
Id: runnerUuid.String(),
|
||||||
@@ -105,7 +143,7 @@ func (s *ApiServer) ReleaseRunner(
|
|||||||
s.allocatedRunnersMutex.Lock()
|
s.allocatedRunnersMutex.Lock()
|
||||||
runner := s.allocatedRunners[uuid]
|
runner := s.allocatedRunners[uuid]
|
||||||
delete(s.allocatedRunners, uuid)
|
delete(s.allocatedRunners, uuid)
|
||||||
runner.Release()
|
runner.Release(s.releaseCh)
|
||||||
s.allocatedRunnersMutex.Unlock()
|
s.allocatedRunnersMutex.Unlock()
|
||||||
|
|
||||||
res := connect.NewResponse(&apiv2.ReleaseRunnerResponse{})
|
res := connect.NewResponse(&apiv2.ReleaseRunnerResponse{})
|
||||||
@@ -144,9 +182,10 @@ func (s *ApiServer) RunCommand(
|
|||||||
return res, nil
|
return res, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func CreateHandler(getRunnerCh chan runnermanager.GetRunnerRequest, mux *http.ServeMux) {
|
func CreateHandler(allocationCh chan runnermanager.RunnerAllocationRequest, releaseCh chan runnermanager.RunnerReleaseRequest, mux *http.ServeMux) {
|
||||||
api_server := &ApiServer{
|
api_server := &ApiServer{
|
||||||
getRunnerCh: getRunnerCh,
|
allocationCh: allocationCh,
|
||||||
|
releaseCh: releaseCh,
|
||||||
allocatedRunners: make(map[uuid.UUID]*RunnerWrapper),
|
allocatedRunners: make(map[uuid.UUID]*RunnerWrapper),
|
||||||
}
|
}
|
||||||
path, handler := apiv2connect.NewGetRunnerServiceHandler(api_server)
|
path, handler := apiv2connect.NewGetRunnerServiceHandler(api_server)
|
||||||
|
|||||||
@@ -4,81 +4,155 @@ import (
|
|||||||
"bytes"
|
"bytes"
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/jhoonb/archivex"
|
||||||
|
|
||||||
"git.ohea.xyz/cursorius/server/config"
|
|
||||||
"github.com/docker/docker/api/types"
|
"github.com/docker/docker/api/types"
|
||||||
"github.com/docker/docker/api/types/container"
|
"github.com/docker/docker/api/types/container"
|
||||||
"github.com/docker/docker/api/types/mount"
|
"github.com/docker/docker/api/types/mount"
|
||||||
"github.com/docker/docker/client"
|
"github.com/docker/docker/client"
|
||||||
"github.com/docker/docker/pkg/stdcopy"
|
"github.com/docker/docker/pkg/stdcopy"
|
||||||
|
"github.com/go-git/go-git/v5"
|
||||||
|
"github.com/go-git/go-git/v5/plumbing/transport"
|
||||||
|
"github.com/go-git/go-git/v5/plumbing/transport/http"
|
||||||
|
"github.com/go-git/go-git/v5/plumbing/transport/ssh"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
|
|
||||||
|
"git.ohea.xyz/cursorius/server/config"
|
||||||
|
"git.ohea.xyz/cursorius/server/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
type PipelineExecution struct {
|
type PipelineExecution struct {
|
||||||
Name string
|
Pipeline database.Pipeline
|
||||||
Job config.Job
|
|
||||||
Ref string
|
Ref string
|
||||||
|
Run database.Run
|
||||||
}
|
}
|
||||||
|
|
||||||
func ExecutePipeline(pe PipelineExecution, pipelineConf config.PipelineConf) error {
|
func ExecutePipeline(pe PipelineExecution, db database.Database, pipelineConf config.PipelineConf) {
|
||||||
jobFolder := filepath.Join(pipelineConf.WorkingDir, pe.Name)
|
jobFolder := filepath.Join(pipelineConf.WorkingDir, pe.Pipeline.Id.String(), pe.Run.Id.String())
|
||||||
|
cloneFolder := filepath.Join(jobFolder, "repo")
|
||||||
|
|
||||||
log.Debugf("Job %v configured with URL \"%v\"", pe.Name, pe.Job.URL)
|
log.Debugf("Job %v configured with URL \"%v\"", pe.Pipeline.Name, pe.Pipeline.Url)
|
||||||
|
|
||||||
log.Debugf("Job %v configured with folder \"%v\"", pe.Name, jobFolder)
|
log.Debugf("Job %v configured with folder \"%v\"", pe.Pipeline.Name, jobFolder)
|
||||||
|
|
||||||
err := os.RemoveAll(jobFolder)
|
err := os.RemoveAll(jobFolder)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not delete existing folder %v", jobFolder)
|
log.Errorf("could not delete existing folder %v", jobFolder)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
err = os.MkdirAll(jobFolder, 0755)
|
err = os.MkdirAll(cloneFolder, 0755)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not create working directory for job %v: %w", pe.Name, err)
|
log.Errorf("could not create working directory for job %v: %w", pe.Pipeline.Name, err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Infof("Cloning source from URL %v", pe.Job.URL)
|
log.Infof("Cloning source from URL %v", pe.Pipeline.Url)
|
||||||
// TODO: should I use go-git here instead of shelling out to raw git?
|
|
||||||
cloneCmd := exec.Command("git", "clone", pe.Job.URL, jobFolder)
|
var auth transport.AuthMethod
|
||||||
output, err := cloneCmd.CombinedOutput()
|
|
||||||
|
if pe.Pipeline.CloneCredential != nil {
|
||||||
|
credential, err := db.GetCloneCredentialById(*pe.Pipeline.CloneCredential)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Debugf("%s", output)
|
log.Errorf("could not get credenital from db: %v", err)
|
||||||
return fmt.Errorf("could not clone source: %w", err)
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
switch credential.Type {
|
||||||
|
case "USER_PASS":
|
||||||
|
log.Debugf("job %v configured to use credential %v", pe.Pipeline.Name, credential.Name)
|
||||||
|
auth = transport.AuthMethod(&http.BasicAuth{
|
||||||
|
Username: credential.Username,
|
||||||
|
Password: credential.Secret,
|
||||||
|
})
|
||||||
|
case "SSH_KEY":
|
||||||
|
publicKeys, err := ssh.NewPublicKeys(credential.Username, []byte(credential.Secret), "")
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("could not parse credential %v", credential.Name)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
auth = transport.AuthMethod(publicKeys)
|
||||||
|
default:
|
||||||
|
log.Errorf("unsupported credential type %v", credential.Type)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
auth = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err = git.PlainClone(cloneFolder, false, &git.CloneOptions{
|
||||||
|
URL: pe.Pipeline.Url,
|
||||||
|
Auth: auth,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("could not clone repo: %v", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
cli, err := client.NewClientWithOpts(client.FromEnv)
|
cli, err := client.NewClientWithOpts(client.FromEnv)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Could not create docker client: %w", err)
|
log.Errorf("Could not create docker client: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
log.Info("Source cloned successfully")
|
log.Info("Source cloned successfully")
|
||||||
|
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
imageName := "git.ohea.xyz/cursorius/pipeline-api/cursorius-pipeline:v2"
|
log.Info("Building container")
|
||||||
|
|
||||||
log.Infof("Pulling image %v", imageName)
|
tarFile := filepath.Join(jobFolder, "archive.tar")
|
||||||
pullOutput, err := cli.ImagePull(ctx, imageName, types.ImagePullOptions{})
|
tar := new(archivex.TarFile)
|
||||||
|
err = tar.Create(tarFile)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not pull image %v: %w", imageName, err)
|
log.Errorf("could not create tarfile: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
buf, err := io.ReadAll(pullOutput)
|
err = tar.AddAll(cloneFolder, false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not read from io.ReadCloser:, %w", err)
|
log.Errorf("could not add repo to tarfile: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
log.Infof("%s", buf)
|
|
||||||
|
|
||||||
err = pullOutput.Close()
|
err = tar.Close()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not close io.ReadCloser: %w", err)
|
log.Errorf("could not close tarfile: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
log.Info("Image pulled sucessfully")
|
|
||||||
|
dockerBuildContext, err := os.Open(tarFile)
|
||||||
|
defer dockerBuildContext.Close()
|
||||||
|
|
||||||
|
imageName := fmt.Sprintf("%v-%v:latest", pe.Pipeline.Id.String(), pe.Run.Id.String())
|
||||||
|
|
||||||
|
buildResponse, err := cli.ImageBuild(context.Background(), dockerBuildContext, types.ImageBuildOptions{
|
||||||
|
Tags: []string{imageName},
|
||||||
|
Dockerfile: ".cursorius/Dockerfile",
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("could not build container: %w", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
err = db.UpdateRunBuildOutput(pe.Run.Id, cleanupBuildOutput(buildResponse.Body))
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("could not update build output for run: %w", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
err = buildResponse.Body.Close()
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not close build response body: %w", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Info("Image built sucessfully")
|
||||||
|
|
||||||
hostConfig := container.HostConfig{}
|
hostConfig := container.HostConfig{}
|
||||||
|
|
||||||
@@ -87,10 +161,12 @@ func ExecutePipeline(pe PipelineExecution, pipelineConf config.PipelineConf) err
|
|||||||
}
|
}
|
||||||
|
|
||||||
if pipelineConf.MountConf.Type == config.Bind {
|
if pipelineConf.MountConf.Type == config.Bind {
|
||||||
|
sourceDir := filepath.Join(pipelineConf.MountConf.Source, pe.Pipeline.Id.String(), pe.Run.Id.String())
|
||||||
|
|
||||||
hostConfig.Mounts = append(hostConfig.Mounts,
|
hostConfig.Mounts = append(hostConfig.Mounts,
|
||||||
mount.Mount{
|
mount.Mount{
|
||||||
Type: mount.TypeBind,
|
Type: mount.TypeBind,
|
||||||
Source: fmt.Sprintf("%v/%v", pipelineConf.MountConf.Source, pe.Name),
|
Source: sourceDir,
|
||||||
Target: "/cursorius/src",
|
Target: "/cursorius/src",
|
||||||
ReadOnly: false,
|
ReadOnly: false,
|
||||||
Consistency: mount.ConsistencyDefault,
|
Consistency: mount.ConsistencyDefault,
|
||||||
@@ -108,42 +184,71 @@ func ExecutePipeline(pe PipelineExecution, pipelineConf config.PipelineConf) err
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
env := make([]string, 0)
|
||||||
|
|
||||||
|
// set cursorius environment variables
|
||||||
|
env = append(env, []string{
|
||||||
|
fmt.Sprintf("CURSORIUS_RUN_ID=%v", pe.Run.Id),
|
||||||
|
"CURSORIUS_SRC_DIR=/cursorius/src",
|
||||||
|
fmt.Sprintf("CURSORIUS_SERVER_URL=%v", pipelineConf.AccessURL),
|
||||||
|
}...)
|
||||||
|
|
||||||
|
// load secrets into environment
|
||||||
|
secrets, err := db.GetSecretsForPipeline(pe.Pipeline.Id)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not get secrets for pipeline", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, secret := range secrets {
|
||||||
|
// the env name is validated to be just uppercase letters, numbers, and underscores on ingestion
|
||||||
|
env = append(env, fmt.Sprintf("%v=%v", strings.ToUpper(secret.Name), secret.Secret))
|
||||||
|
}
|
||||||
|
|
||||||
resp, err := cli.ContainerCreate(ctx,
|
resp, err := cli.ContainerCreate(ctx,
|
||||||
&container.Config{
|
&container.Config{
|
||||||
Image: imageName,
|
Image: imageName,
|
||||||
Cmd: []string{"/launcher.sh"},
|
|
||||||
Tty: false,
|
Tty: false,
|
||||||
Env: []string{
|
Env: env,
|
||||||
"CURSORIUS_SRC_DIR=/cursorius/src",
|
|
||||||
fmt.Sprintf("CUROSRIUS_SERVER_URL=%v", pipelineConf.AccessURL),
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
// TODO: fix running the runner in docker (add VolumesFrom to HostConfig)
|
// TODO: fix running the runner in docker (add VolumesFrom to HostConfig)
|
||||||
&hostConfig,
|
&hostConfig,
|
||||||
nil, nil, "",
|
nil, nil, "",
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not create container: %w", err)
|
log.Errorf("could not create container: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Info("Launching container")
|
log.Info("Launching container")
|
||||||
|
|
||||||
if err := cli.ContainerStart(ctx, resp.ID, types.ContainerStartOptions{}); err != nil {
|
if err := cli.ContainerStart(ctx, resp.ID, types.ContainerStartOptions{}); err != nil {
|
||||||
return fmt.Errorf("could not start container: %w", err)
|
log.Errorf("could not start container: %v", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
statusCh, errCh := cli.ContainerWait(ctx, resp.ID, container.WaitConditionNotRunning)
|
statusCh, errCh := cli.ContainerWait(ctx, resp.ID, container.WaitConditionNotRunning)
|
||||||
select {
|
select {
|
||||||
case err := <-errCh:
|
case err := <-errCh:
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("container returned error: %w", err)
|
log.Errorf("container returned error: %v", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
case retCode := <-statusCh:
|
case okBody := <-statusCh:
|
||||||
log.Debugf("Container finished running with return code: %v", retCode)
|
if okBody.Error != nil {
|
||||||
|
log.Errorf("Could not wait on container: %v", err)
|
||||||
|
return
|
||||||
|
} else {
|
||||||
|
log.Debugf("Container finished running with return code: %v", okBody.StatusCode)
|
||||||
|
pe.Run.Result = &okBody.StatusCode
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pe.Run.InProgress = false
|
||||||
|
|
||||||
out, err := cli.ContainerLogs(ctx, resp.ID, types.ContainerLogsOptions{ShowStdout: true, ShowStderr: true})
|
out, err := cli.ContainerLogs(ctx, resp.ID, types.ContainerLogsOptions{ShowStdout: true, ShowStderr: true})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("could not get container logs: %w", err)
|
log.Errorf("could not get container logs: %w", err)
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
var stdOut bytes.Buffer
|
var stdOut bytes.Buffer
|
||||||
@@ -151,8 +256,10 @@ func ExecutePipeline(pe PipelineExecution, pipelineConf config.PipelineConf) err
|
|||||||
|
|
||||||
stdcopy.StdCopy(&stdOut, &stdErr, out)
|
stdcopy.StdCopy(&stdOut, &stdErr, out)
|
||||||
|
|
||||||
log.Debugf("%s", stdOut.Bytes())
|
pe.Run.Stdout = stdOut.Bytes()
|
||||||
log.Debugf("%s", stdErr.Bytes())
|
pe.Run.Stderr = stdErr.Bytes()
|
||||||
|
|
||||||
return nil
|
db.UpdateRunResult(pe.Run)
|
||||||
|
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,25 @@
|
|||||||
|
package pipeline_executor
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"encoding/json"
|
||||||
|
"io"
|
||||||
|
)
|
||||||
|
|
||||||
|
func cleanupBuildOutput(input io.ReadCloser) string {
|
||||||
|
output := ""
|
||||||
|
|
||||||
|
scanner := bufio.NewScanner(input)
|
||||||
|
for scanner.Scan() {
|
||||||
|
var log map[string]any
|
||||||
|
json.Unmarshal(scanner.Bytes(), &log)
|
||||||
|
|
||||||
|
if logVar, ok := log["stream"]; ok {
|
||||||
|
if log, ok := logVar.(string); ok {
|
||||||
|
output += log
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return output
|
||||||
|
}
|
||||||
+112
-32
@@ -1,15 +1,18 @@
|
|||||||
package poll
|
package poll
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/op/go-logging"
|
|
||||||
|
|
||||||
"git.ohea.xyz/cursorius/server/config"
|
"git.ohea.xyz/cursorius/server/config"
|
||||||
|
"git.ohea.xyz/cursorius/server/database"
|
||||||
"git.ohea.xyz/cursorius/server/pipeline_executor"
|
"git.ohea.xyz/cursorius/server/pipeline_executor"
|
||||||
|
|
||||||
"github.com/go-git/go-git/v5"
|
"github.com/go-git/go-git/v5"
|
||||||
"github.com/go-git/go-git/v5/plumbing"
|
"github.com/go-git/go-git/v5/plumbing"
|
||||||
"github.com/go-git/go-git/v5/storage/memory"
|
"github.com/go-git/go-git/v5/storage/memory"
|
||||||
|
"github.com/google/uuid"
|
||||||
|
"github.com/op/go-logging"
|
||||||
)
|
)
|
||||||
|
|
||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
@@ -24,17 +27,43 @@ type tag struct {
|
|||||||
commitHash string
|
commitHash string
|
||||||
}
|
}
|
||||||
|
|
||||||
func pollJob(repoName string, jobConfig config.Job, pipelineConf config.PipelineConf) {
|
func pollJob(ctx context.Context, pipeline database.Pipeline, pipelineConf config.PipelineConf, db database.Database) {
|
||||||
prevCommits := make(map[string]string)
|
firstScan := true
|
||||||
for {
|
for {
|
||||||
time.Sleep(time.Duration(jobConfig.PollInterval) * time.Second)
|
// Don't sleep on first scan to ease testing
|
||||||
log.Infof("Polling repo %v", repoName)
|
// TODO: this should be replaced with a script that mocks a webhook
|
||||||
|
if !firstScan {
|
||||||
|
|
||||||
|
ctx, cancel := context.WithTimeout(ctx, time.Duration(pipeline.PollInterval)*time.Second)
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
switch ctx.Err() {
|
||||||
|
case context.Canceled:
|
||||||
|
log.Infof("Polling for pipeline %v canceled, stopping", pipeline.Name)
|
||||||
|
cancel()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
cancel()
|
||||||
|
|
||||||
|
log.Infof("Polling repo %v", pipeline.Name)
|
||||||
|
} else {
|
||||||
|
firstScan = false
|
||||||
|
}
|
||||||
|
|
||||||
|
prevRefs, err := db.GetPipelineRefs(pipeline.Id)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not get pipeline refs from db: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
repo, err := git.Clone(memory.NewStorage(), nil, &git.CloneOptions{
|
repo, err := git.Clone(memory.NewStorage(), nil, &git.CloneOptions{
|
||||||
URL: jobConfig.URL,
|
URL: pipeline.Url,
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("Could not clone repo %v from url %v: %v", repoName, jobConfig.URL, err)
|
log.Errorf("Could not clone repo %v from url %v: %v", pipeline.Name, pipeline.Url, err)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
refsToRunFor := []string{}
|
refsToRunFor := []string{}
|
||||||
@@ -42,24 +71,24 @@ func pollJob(repoName string, jobConfig config.Job, pipelineConf config.Pipeline
|
|||||||
// get branches
|
// get branches
|
||||||
branches, err := repo.Branches()
|
branches, err := repo.Branches()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("Could not enumerate branches in repo %v: %v", repoName, err)
|
log.Errorf("Could not enumerate branches in repo %v: %v", pipeline.Name, err)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
branches.ForEach(func(branch *plumbing.Reference) error {
|
branches.ForEach(func(branch *plumbing.Reference) error {
|
||||||
log.Debugf("Processing branch %v from repo %v", branch.Name().String(), repoName)
|
log.Debugf("Processing branch %v from repo %v (id: %v)", branch.Name().String(), pipeline.Name, pipeline.Id)
|
||||||
prevRef, ok := prevCommits[branch.Name().String()]
|
prevRef, ok := prevRefs[branch.Name().String()]
|
||||||
if ok {
|
if ok {
|
||||||
if branch.Hash().String() != prevRef {
|
if branch.Hash().String() != prevRef {
|
||||||
log.Debugf("Queuing job for branch %v in repo %v with hash %v", branch.Name().String(), repoName, branch.Hash().String())
|
log.Debugf("Queuing job for branch %v in repo %v (id: %v) with hash %v", branch.Name().String(), pipeline.Name, pipeline.Id, branch.Hash().String())
|
||||||
prevCommits[branch.Name().String()] = branch.Hash().String()
|
prevRefs[branch.Name().String()] = branch.Hash().String()
|
||||||
refsToRunFor = append(refsToRunFor, branch.Name().String())
|
refsToRunFor = append(refsToRunFor, branch.Name().String())
|
||||||
} else {
|
} else {
|
||||||
log.Debugf("Branch %v in repo %v has hash %v, which matches the previously seen hash of %v", branch.Name().String(), repoName, branch.Hash().String(), prevRef)
|
log.Debugf("Branch %v in repo %v (id: %v) has hash %v, which matches the previously seen hash of %v", branch.Name().String(), pipeline.Name, pipeline.Id, branch.Hash().String(), prevRef)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
log.Debugf("Queuing job for newly discovered branch %v in repo %v with hash %v", branch.Name().String(), repoName, branch.Hash().String())
|
log.Debugf("Queuing job for newly discovered branch %v in repo %v (id: %v) with hash %v", branch.Name().String(), pipeline.Name, pipeline.Id, branch.Hash().String())
|
||||||
prevCommits[branch.Name().String()] = branch.Hash().String()
|
prevRefs[branch.Name().String()] = branch.Hash().String()
|
||||||
refsToRunFor = append(refsToRunFor, branch.Name().String())
|
refsToRunFor = append(refsToRunFor, branch.Name().String())
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@@ -67,48 +96,99 @@ func pollJob(repoName string, jobConfig config.Job, pipelineConf config.Pipeline
|
|||||||
|
|
||||||
tags, err := repo.Tags()
|
tags, err := repo.Tags()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("Could not enumerate tags in repo %v: %v", repoName, err)
|
log.Errorf("Could not enumerate tags in repo %v: %v", pipeline.Name, err)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
tags.ForEach(func(tag *plumbing.Reference) error {
|
tags.ForEach(func(tag *plumbing.Reference) error {
|
||||||
log.Debugf("Processing tag %v from repo %v", tag.Name().String(), repoName)
|
log.Debugf("Processing tag %v from repo %v (id: %v)", tag.Name().String(), pipeline.Name, pipeline.Id)
|
||||||
prevRef, ok := prevCommits[tag.Name().String()]
|
prevRef, ok := prevRefs[tag.Name().String()]
|
||||||
if ok {
|
if ok {
|
||||||
if tag.Hash().String() != prevRef {
|
if tag.Hash().String() != prevRef {
|
||||||
log.Debugf("Queuing job for tag %v in repo %v with hash %v", tag.Name().String(), repoName, tag.Hash().String())
|
log.Debugf("Queuing job for tag %v in repo %v (id: %v) with hash %v", tag.Name().String(), pipeline.Name, pipeline.Id, tag.Hash().String())
|
||||||
prevCommits[tag.Name().String()] = tag.Hash().String()
|
prevRefs[tag.Name().String()] = tag.Hash().String()
|
||||||
refsToRunFor = append(refsToRunFor, tag.Name().String())
|
refsToRunFor = append(refsToRunFor, tag.Name().String())
|
||||||
} else {
|
} else {
|
||||||
log.Debugf("Tag %v in repo %v has hash %v, which matches the previously seen hash of %v", tag.Name().String(), repoName, tag.Hash().String(), prevRef)
|
log.Debugf("Tag %v in repo %v (id: %v) has hash %v, which matches the previously seen hash of %v", tag.Name().String(), pipeline.Name, pipeline.Id, tag.Hash().String(), prevRef)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
log.Debugf("Queuing job for newly discovered tag %v in repo %v with hash %v", tag.Name().String(), repoName, tag.Hash().String())
|
log.Debugf("Queuing job for newly discovered tag %v in repo %v (id: %v) with hash %v", tag.Name().String(), pipeline.Name, pipeline.Id, tag.Hash().String())
|
||||||
prevCommits[tag.Name().String()] = tag.Hash().String()
|
prevRefs[tag.Name().String()] = tag.Hash().String()
|
||||||
refsToRunFor = append(refsToRunFor, tag.Name().String())
|
refsToRunFor = append(refsToRunFor, tag.Name().String())
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
|
|
||||||
|
err = db.UpdatePipelineRefs(pipeline.Id, prevRefs)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not update pipeline refs: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
for _, ref := range refsToRunFor {
|
for _, ref := range refsToRunFor {
|
||||||
log.Debugf("Dispatching job for ref %v in repo %v", ref, repoName)
|
log.Debugf("Dispatching job for ref %v in repo %v (id: %v)", ref, pipeline.Name, pipeline.Id)
|
||||||
|
|
||||||
|
run, err := db.CreateRun(pipeline.Id)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not create run for pipeline with id \"%v\": ", pipeline.Id, err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
pe := pipeline_executor.PipelineExecution{
|
pe := pipeline_executor.PipelineExecution{
|
||||||
Name: repoName,
|
Pipeline: pipeline,
|
||||||
Job: jobConfig,
|
|
||||||
Ref: ref,
|
Ref: ref,
|
||||||
|
Run: run,
|
||||||
}
|
}
|
||||||
|
|
||||||
pipeline_executor.ExecutePipeline(pe, pipelineConf)
|
go pipeline_executor.ExecutePipeline(pe, db, pipelineConf)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func StartPolling(conf config.Config) {
|
func launchPollJobs(conf config.PipelineConf, db database.Database, pollChan chan uuid.UUID) {
|
||||||
for jobName, job := range conf.Jobs {
|
pipelines, err := db.GetPipelines()
|
||||||
if job.PollInterval == 0 {
|
if err != nil {
|
||||||
|
log.Errorf("Could not get pipelines from database: %w", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
pipelineCancelations := make(map[uuid.UUID]context.CancelFunc)
|
||||||
|
|
||||||
|
for _, pipeline := range pipelines {
|
||||||
|
if pipeline.PollInterval == 0 {
|
||||||
continue
|
continue
|
||||||
} else {
|
} else {
|
||||||
go pollJob(jobName, job, conf.PipelineConf)
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
pipelineCancelations[pipeline.Id] = cancel
|
||||||
|
|
||||||
|
log.Infof("Starting polling for pipeline %v with id %v", pipeline.Name, pipeline.Id)
|
||||||
|
go pollJob(ctx, pipeline, conf, db)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
for {
|
||||||
|
jobUUID := <-pollChan
|
||||||
|
pipeline, err := db.GetPipelineById(jobUUID)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not get pipeline with id \"%v\" from database: %v", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cancel existing polling job if it exists
|
||||||
|
if cancelFunc, ok := pipelineCancelations[pipeline.Id]; ok {
|
||||||
|
cancelFunc()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start new polling job
|
||||||
|
log.Infof("Starting polling for pipeline %v with id %v", pipeline.Name, pipeline.Id)
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
pipelineCancelations[pipeline.Id] = cancel
|
||||||
|
go pollJob(ctx, pipeline, conf, db)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func StartPolling(conf config.PipelineConf, db database.Database) chan uuid.UUID {
|
||||||
|
pollChan := make(chan uuid.UUID)
|
||||||
|
go launchPollJobs(conf, db, pollChan)
|
||||||
|
return pollChan
|
||||||
}
|
}
|
||||||
|
|||||||
+24
-7
@@ -4,6 +4,7 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/google/uuid"
|
||||||
"google.golang.org/protobuf/proto"
|
"google.golang.org/protobuf/proto"
|
||||||
"google.golang.org/protobuf/reflect/protoreflect"
|
"google.golang.org/protobuf/reflect/protoreflect"
|
||||||
"nhooyr.io/websocket"
|
"nhooyr.io/websocket"
|
||||||
@@ -17,23 +18,37 @@ type RunnerData struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Runner struct {
|
type Runner struct {
|
||||||
id string
|
id uuid.UUID
|
||||||
tags []string
|
tags []string
|
||||||
conn *websocket.Conn
|
conn *websocket.Conn
|
||||||
receiveChan chan []byte
|
receiveChan chan []byte
|
||||||
running bool
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *Runner) Id() string {
|
func (r *Runner) HasTags(requestedTags []string) bool {
|
||||||
|
tagIter:
|
||||||
|
for _, requestedTag := range requestedTags {
|
||||||
|
for _, posessedTag := range r.tags {
|
||||||
|
// if we find the tag, move on to search for the next one
|
||||||
|
if posessedTag == requestedTag {
|
||||||
|
continue tagIter
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// if we don't find the tag
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *Runner) Id() uuid.UUID {
|
||||||
return r.id
|
return r.id
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *Runner) Release() {
|
|
||||||
r.running = false
|
|
||||||
}
|
|
||||||
|
|
||||||
func (r *Runner) RunCommand(cmd string, args []string) (returnCode int64, stdout string, stderr string, err error) {
|
func (r *Runner) RunCommand(cmd string, args []string) (returnCode int64, stdout string, stderr string, err error) {
|
||||||
|
|
||||||
|
if r.conn == nil {
|
||||||
|
return 0, "", "", fmt.Errorf("runner with id %v has nil conn, THIS IS A BUG", r.id)
|
||||||
|
}
|
||||||
|
|
||||||
// Write RunCommand message to client
|
// Write RunCommand message to client
|
||||||
serverToRunnerMsg := &runner_api.ServerToRunnerMsg{
|
serverToRunnerMsg := &runner_api.ServerToRunnerMsg{
|
||||||
Msg: &runner_api.ServerToRunnerMsg_RunCommandMsg{
|
Msg: &runner_api.ServerToRunnerMsg_RunCommandMsg{
|
||||||
@@ -86,6 +101,8 @@ func (r *Runner) sendProtoStruct(p protoreflect.ProtoMessage) error {
|
|||||||
|
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
|
log.Debugf("r.conn: %p", r.conn)
|
||||||
|
|
||||||
if err := r.conn.Write(ctx, websocket.MessageBinary, protoOut); err != nil {
|
if err := r.conn.Write(ctx, websocket.MessageBinary, protoOut); err != nil {
|
||||||
return fmt.Errorf("Could not send proto to websocket: %w", err)
|
return fmt.Errorf("Could not send proto to websocket: %w", err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,71 +3,33 @@ package runnermanager
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/google/uuid"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
"google.golang.org/protobuf/proto"
|
"google.golang.org/protobuf/proto"
|
||||||
"nhooyr.io/websocket"
|
"nhooyr.io/websocket"
|
||||||
|
|
||||||
"git.ohea.xyz/cursorius/server/config"
|
"git.ohea.xyz/cursorius/server/config"
|
||||||
|
"git.ohea.xyz/cursorius/server/database"
|
||||||
|
"git.ohea.xyz/cursorius/server/util"
|
||||||
|
|
||||||
runner_api "git.ohea.xyz/cursorius/runner-api/go/api/v2"
|
runner_api "git.ohea.xyz/cursorius/runner-api/go/api/v2"
|
||||||
)
|
)
|
||||||
|
|
||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
type RunnerRegistration struct {
|
func (r *runnerManager) processRunnerAllocation(req RunnerAllocationRequest) {
|
||||||
Secret string
|
tagsStr := util.FormatTags(req.Tags)
|
||||||
Id string
|
log.Infof("Got request for runner with tags \"%v\"", tagsStr)
|
||||||
Tags []string
|
|
||||||
conn *websocket.Conn
|
|
||||||
}
|
|
||||||
|
|
||||||
type runnerManager struct {
|
log.Debugf("Finding runner with tags %v", tagsStr)
|
||||||
getRunnerCh chan GetRunnerRequest
|
|
||||||
registerCh chan RunnerRegistration
|
|
||||||
connectedRunners []Runner
|
|
||||||
numConnectedRunners uint64
|
|
||||||
configuredRunners map[string]config.Runner
|
|
||||||
}
|
|
||||||
|
|
||||||
type GetRunnerRequest struct {
|
|
||||||
Tags []string
|
|
||||||
RespChan chan GetRunnerResponse
|
|
||||||
}
|
|
||||||
|
|
||||||
type GetRunnerResponse struct {
|
|
||||||
Runner *Runner
|
|
||||||
Err error
|
|
||||||
}
|
|
||||||
|
|
||||||
type runnerJob struct {
|
|
||||||
Id string
|
|
||||||
URL string
|
|
||||||
}
|
|
||||||
|
|
||||||
func (r *runnerManager) processRequest(req GetRunnerRequest) {
|
|
||||||
var runnerTagsStr strings.Builder
|
|
||||||
fmt.Fprintf(&runnerTagsStr, "[%v", req.Tags[0])
|
|
||||||
for _, tag := range req.Tags[1:] {
|
|
||||||
fmt.Fprintf(&runnerTagsStr, ", %v", tag)
|
|
||||||
}
|
|
||||||
fmt.Fprintf(&runnerTagsStr, "]")
|
|
||||||
log.Infof("Got request for runner with tags \"%v\"", runnerTagsStr.String())
|
|
||||||
|
|
||||||
log.Debugf("Finding runner with tags %v", runnerTagsStr.String())
|
|
||||||
|
|
||||||
foundRunner := false
|
foundRunner := false
|
||||||
|
|
||||||
runnersToRemove := []int{}
|
runnersToRemove := []int{}
|
||||||
runnerIter:
|
runnerIter:
|
||||||
for i, runner := range r.connectedRunners {
|
for i, runner := range r.connectedRunners {
|
||||||
// don't allocate runner that is already occupied
|
|
||||||
if runner.running {
|
|
||||||
log.Debugf("Skipping runner %v, as runner is activly running another job", runner.id)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
// don't allocate runner with closed receiveChan (is defunct)
|
// don't allocate runner with closed receiveChan (is defunct)
|
||||||
// there should never be messages to read on an inactive runner,
|
// there should never be messages to read on an inactive runner,
|
||||||
// so we aren't losing any data here
|
// so we aren't losing any data here
|
||||||
@@ -84,29 +46,25 @@ runnerIter:
|
|||||||
default:
|
default:
|
||||||
log.Debugf("Checking runner %v for requested tags", runner.id)
|
log.Debugf("Checking runner %v for requested tags", runner.id)
|
||||||
|
|
||||||
tagIter:
|
if !runner.HasTags(req.Tags) {
|
||||||
for _, requestedTag := range req.Tags {
|
|
||||||
for _, posessedTag := range runner.tags {
|
|
||||||
if requestedTag == posessedTag {
|
|
||||||
continue tagIter
|
|
||||||
}
|
|
||||||
}
|
|
||||||
continue runnerIter
|
continue runnerIter
|
||||||
}
|
}
|
||||||
|
|
||||||
r.connectedRunners[i].running = true
|
runnersToRemove = append(runnersToRemove, i)
|
||||||
foundRunner = true
|
foundRunner = true
|
||||||
req.RespChan <- GetRunnerResponse{
|
log.Debugf("Runner %v has requested tags, allocating", runner.id)
|
||||||
|
req.RespChan <- RunnerAllocationResponse{
|
||||||
Runner: &r.connectedRunners[i],
|
Runner: &r.connectedRunners[i],
|
||||||
Err: nil,
|
Err: nil,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
// remove allocated runner plus defunct runners
|
||||||
// since we iterate, all the indexes will be in accending order
|
// since we iterate, all the indexes will be in accending order
|
||||||
for i, runnerInd := range runnersToRemove {
|
for i, runnerInd := range runnersToRemove {
|
||||||
r.connectedRunners[runnerInd-i] = r.connectedRunners[len(r.connectedRunners)-1]
|
r.connectedRunners[runnerInd-i] = r.connectedRunners[len(r.connectedRunners)-1]
|
||||||
r.connectedRunners = r.connectedRunners[0 : len(r.connectedRunners)-2]
|
r.connectedRunners = r.connectedRunners[0 : len(r.connectedRunners)-1]
|
||||||
}
|
}
|
||||||
|
|
||||||
if foundRunner {
|
if foundRunner {
|
||||||
@@ -117,24 +75,42 @@ runnerIter:
|
|||||||
if len(r.connectedRunners) == 0 {
|
if len(r.connectedRunners) == 0 {
|
||||||
errorMsg = "no connected runners"
|
errorMsg = "no connected runners"
|
||||||
}
|
}
|
||||||
req.RespChan <- GetRunnerResponse{
|
req.RespChan <- RunnerAllocationResponse{
|
||||||
Runner: &Runner{},
|
Runner: nil,
|
||||||
Err: fmt.Errorf("Could not allocate runner: %v", errorMsg),
|
Err: fmt.Errorf("Could not allocate runner: %v", errorMsg),
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *runnerManager) processRegistration(reg RunnerRegistration) {
|
func (r *runnerManager) processRunnerRegistration(req RunnerRegistrationRequest) {
|
||||||
log.Debugf("New runner appeared with id: %v and secret: %v", reg.Id, reg.Secret)
|
log.Debugf("New runner appeared with id: %v and secret: %v", req.Id, req.Secret)
|
||||||
if configuredRunner, doesExist := r.configuredRunners[reg.Id]; doesExist {
|
|
||||||
if configuredRunner.Secret == reg.Secret {
|
// Get runner with give id from database
|
||||||
log.Infof("Registering runner \"%v\" with tags %v", reg.Id, reg.Tags)
|
runnerId, err := uuid.Parse(req.Id)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Disconnecting runner with id: %v, could not parse as UUID: %v", req.Id, err)
|
||||||
|
req.conn.Close(websocket.StatusNormalClosure, "registration invalid")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
dbRunner, err := r.db.GetRunnerById(runnerId)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Disconnecting runner with id: %v, could not find runner in DB: %v", runnerId, err)
|
||||||
|
req.conn.Close(websocket.StatusNormalClosure, "registration invalid")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if req.Secret != dbRunner.Token {
|
||||||
|
log.Errorf("Disconnecting runner with id: %v, invalid secret", runnerId)
|
||||||
|
req.conn.Close(websocket.StatusNormalClosure, "registration invalid")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Infof("Registering runner \"%v\" with tags %v", req.Id, req.Tags)
|
||||||
runner := Runner{
|
runner := Runner{
|
||||||
id: reg.Id,
|
id: runnerId,
|
||||||
tags: reg.Tags,
|
tags: req.Tags,
|
||||||
conn: reg.conn,
|
conn: req.conn,
|
||||||
receiveChan: make(chan []byte),
|
receiveChan: make(chan []byte),
|
||||||
running: false,
|
|
||||||
}
|
}
|
||||||
r.connectedRunners = append(r.connectedRunners, runner)
|
r.connectedRunners = append(r.connectedRunners, runner)
|
||||||
// start goroutine to call Read function on websocket connection
|
// start goroutine to call Read function on websocket connection
|
||||||
@@ -143,7 +119,7 @@ func (r *runnerManager) processRegistration(reg RunnerRegistration) {
|
|||||||
defer log.Noticef("Deregistered runner with id: %v", runner.id)
|
defer log.Noticef("Deregistered runner with id: %v", runner.id)
|
||||||
defer close(runner.receiveChan)
|
defer close(runner.receiveChan)
|
||||||
for {
|
for {
|
||||||
msgType, data, err := reg.conn.Read(context.Background())
|
msgType, data, err := req.conn.Read(context.Background())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// TODO: this is still racy, since a runner could be allocated between the
|
// TODO: this is still racy, since a runner could be allocated between the
|
||||||
// connection returning an err and the channel closing
|
// connection returning an err and the channel closing
|
||||||
@@ -161,47 +137,47 @@ func (r *runnerManager) processRegistration(reg RunnerRegistration) {
|
|||||||
|
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
} else {
|
func (r *runnerManager) processRunnerRelease(req RunnerReleaseRequest) {
|
||||||
log.Errorf("Disconnecting runner with id: %v and invalid secret: %v", reg.Id, reg.Secret)
|
r.connectedRunners = append(r.connectedRunners, *req.Runner)
|
||||||
reg.conn.Close(websocket.StatusNormalClosure, "registration invalid")
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
log.Errorf("Disconnecting runner with invalid id: %v", reg.Id)
|
|
||||||
reg.conn.Close(websocket.StatusNormalClosure, "registration invalid")
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func runRunnerManager(r runnerManager) {
|
func runRunnerManager(r runnerManager) {
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case request := <-r.getRunnerCh:
|
case request := <-r.chans.Allocation:
|
||||||
r.processRequest(request)
|
r.processRunnerAllocation(request)
|
||||||
|
case release := <-r.chans.Release:
|
||||||
case registration := <-r.registerCh:
|
r.processRunnerRelease(release)
|
||||||
r.processRegistration(registration)
|
case registration := <-r.chans.Registration:
|
||||||
|
r.processRunnerRegistration(registration)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func StartRunnerManager(configuredRunners map[string]config.Runner) (chan GetRunnerRequest, chan RunnerRegistration, error) {
|
func StartRunnerManager(configuredRunners map[string]config.Runner, db database.Database) (RunnerManagerChans, error) {
|
||||||
scheduler := runnerManager{
|
scheduler := runnerManager{
|
||||||
getRunnerCh: make(chan GetRunnerRequest),
|
chans: RunnerManagerChans{
|
||||||
registerCh: make(chan RunnerRegistration),
|
Allocation: make(chan RunnerAllocationRequest),
|
||||||
|
Release: make(chan RunnerReleaseRequest),
|
||||||
|
Registration: make(chan RunnerRegistrationRequest),
|
||||||
|
},
|
||||||
connectedRunners: make([]Runner, 0),
|
connectedRunners: make([]Runner, 0),
|
||||||
configuredRunners: configuredRunners,
|
configuredRunners: configuredRunners,
|
||||||
|
db: db,
|
||||||
}
|
}
|
||||||
|
|
||||||
go runRunnerManager(scheduler)
|
go runRunnerManager(scheduler)
|
||||||
|
|
||||||
return scheduler.getRunnerCh, scheduler.registerCh, nil
|
return scheduler.chans, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func RegisterRunner(conn *websocket.Conn, registerCh chan RunnerRegistration) {
|
func RegisterRunner(conn *websocket.Conn, registerCh chan RunnerRegistrationRequest) {
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), time.Second*10)
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second*10)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
var registration RunnerRegistration
|
var registration RunnerRegistrationRequest
|
||||||
registration.conn = conn
|
registration.conn = conn
|
||||||
|
|
||||||
typ, r, err := conn.Read(ctx)
|
typ, r, err := conn.Read(ctx)
|
||||||
|
|||||||
@@ -0,0 +1,49 @@
|
|||||||
|
package runnermanager
|
||||||
|
|
||||||
|
import (
|
||||||
|
"nhooyr.io/websocket"
|
||||||
|
|
||||||
|
"git.ohea.xyz/cursorius/server/config"
|
||||||
|
"git.ohea.xyz/cursorius/server/database"
|
||||||
|
)
|
||||||
|
|
||||||
|
type RunnerManagerChans struct {
|
||||||
|
Allocation chan RunnerAllocationRequest
|
||||||
|
Release chan RunnerReleaseRequest
|
||||||
|
Registration chan RunnerRegistrationRequest
|
||||||
|
}
|
||||||
|
|
||||||
|
type runnerManager struct {
|
||||||
|
chans RunnerManagerChans
|
||||||
|
connectedRunners []Runner
|
||||||
|
numConnectedRunners uint64
|
||||||
|
configuredRunners map[string]config.Runner
|
||||||
|
db database.Database
|
||||||
|
}
|
||||||
|
|
||||||
|
type RunnerAllocationRequest struct {
|
||||||
|
Tags []string
|
||||||
|
RespChan chan RunnerAllocationResponse
|
||||||
|
CancelChan chan string
|
||||||
|
}
|
||||||
|
|
||||||
|
type RunnerAllocationResponse struct {
|
||||||
|
Runner *Runner
|
||||||
|
Err error
|
||||||
|
}
|
||||||
|
|
||||||
|
type RunnerReleaseRequest struct {
|
||||||
|
Runner *Runner
|
||||||
|
}
|
||||||
|
|
||||||
|
type RunnerRegistrationRequest struct {
|
||||||
|
Secret string
|
||||||
|
Id string
|
||||||
|
Tags []string
|
||||||
|
conn *websocket.Conn
|
||||||
|
}
|
||||||
|
|
||||||
|
type runnerJob struct {
|
||||||
|
Id string
|
||||||
|
URL string
|
||||||
|
}
|
||||||
@@ -0,0 +1,18 @@
|
|||||||
|
package util
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
func FormatTags(tags []string) string {
|
||||||
|
var tagsStr strings.Builder
|
||||||
|
if len(tags) > 0 {
|
||||||
|
fmt.Fprintf(&tagsStr, "[%v", tags[0])
|
||||||
|
for _, tag := range tags[1:] {
|
||||||
|
fmt.Fprintf(&tagsStr, ", %v", tag)
|
||||||
|
}
|
||||||
|
fmt.Fprintf(&tagsStr, "]")
|
||||||
|
}
|
||||||
|
return tagsStr.String()
|
||||||
|
}
|
||||||
+47
-29
@@ -5,36 +5,62 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"git.ohea.xyz/cursorius/server/config"
|
"git.ohea.xyz/cursorius/server/config"
|
||||||
|
"git.ohea.xyz/cursorius/server/database"
|
||||||
"git.ohea.xyz/cursorius/server/pipeline_executor"
|
"git.ohea.xyz/cursorius/server/pipeline_executor"
|
||||||
"github.com/go-playground/webhooks/v6/gitea"
|
"github.com/go-playground/webhooks/v6/gitea"
|
||||||
|
"github.com/google/uuid"
|
||||||
"github.com/op/go-logging"
|
"github.com/op/go-logging"
|
||||||
)
|
)
|
||||||
|
|
||||||
var log = logging.MustGetLogger("cursorius-server")
|
var log = logging.MustGetLogger("cursorius-server")
|
||||||
|
|
||||||
func CreateWebhookHandler(conf config.Config, mux *http.ServeMux) {
|
func webhookHandler(w http.ResponseWriter, r *http.Request, db database.Database, conf config.PipelineConf) {
|
||||||
mux.HandleFunc("/webhook/", func(w http.ResponseWriter, r *http.Request) {
|
|
||||||
switch r.Method {
|
switch r.Method {
|
||||||
case "POST":
|
case "POST":
|
||||||
splitUrl := strings.Split(r.URL.Path, "/")
|
splitUrl := strings.Split(r.URL.Path, "/")
|
||||||
if len(splitUrl) != 3 {
|
if len(splitUrl) != 4 {
|
||||||
log.Errorf("Webhook recieved with invalid url \"%v\", ignoring...", r.URL)
|
log.Errorf("Webhook recieved with invalid url \"%v\", ignoring...", r.URL)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// get URL path after /webhook/
|
// get URL path after /webhook/
|
||||||
// TODO: verify that this handles all valid URL formats
|
// TODO: verify that this handles all valid URL formats
|
||||||
webhookJobName := splitUrl[2]
|
pipelineUUIDStr := splitUrl[2]
|
||||||
|
webhookUUIDStr := splitUrl[3]
|
||||||
|
|
||||||
for jobName, jobConfig := range conf.Jobs {
|
pipelineUUID, err := uuid.Parse(pipelineUUIDStr)
|
||||||
if webhookJobName == jobName {
|
if err != nil {
|
||||||
if jobConfig.Webhook == nil {
|
log.Errorf("Could not parse pipeline UUID: %v", err)
|
||||||
log.Errorf("Matching job does not have webhook configuration, ignoring....")
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
switch jobConfig.Webhook.Sender {
|
webhookUUID, err := uuid.Parse(webhookUUIDStr)
|
||||||
case config.Gitea:
|
if err != nil {
|
||||||
hook, err := gitea.New(gitea.Options.Secret(jobConfig.Webhook.Secret))
|
log.Errorf("Could not parse webhook UUID: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
pipeline, err := db.GetPipelineById(pipelineUUID)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not get webhooks for pipeline with UUID \"%v\": %v", pipelineUUID, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
webhooks, err := db.GetWebhooksForPipeline(pipelineUUID)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("Could not get webhooks for pipeline with UUID \"%v\": %v", webhookUUID, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(webhooks) < 1 {
|
||||||
|
log.Errorf("No webhooks configured for pipeline with UUID \"%v\"", webhookUUID)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, webhook := range webhooks {
|
||||||
|
if webhook.Id == webhookUUID {
|
||||||
|
switch webhook.ServerType {
|
||||||
|
case database.Gitea:
|
||||||
|
hook, err := gitea.New(gitea.Options.Secret(webhook.Secret))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("Could not create Gitea webhook handler: %v", err)
|
log.Errorf("Could not create Gitea webhook handler: %v", err)
|
||||||
return
|
return
|
||||||
@@ -42,42 +68,34 @@ func CreateWebhookHandler(conf config.Config, mux *http.ServeMux) {
|
|||||||
payload, err := hook.Parse(r, gitea.PushEvent)
|
payload, err := hook.Parse(r, gitea.PushEvent)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if err == gitea.ErrEventNotFound {
|
if err == gitea.ErrEventNotFound {
|
||||||
log.Info("Got webhook for unexpected event type, ignoring...")
|
log.Warning("Got webhook \"%v\" for unexpected event type, ignoring...", webhookUUID)
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
log.Errorf("Could not parse webhook: %v", err)
|
|
||||||
return
|
|
||||||
}
|
}
|
||||||
log.Infof("Got webhook with payload %v", payload)
|
|
||||||
|
|
||||||
switch payload.(type) {
|
switch payload.(type) {
|
||||||
case gitea.PushPayload:
|
case gitea.PushPayload:
|
||||||
pushPayload := payload.(gitea.PushPayload)
|
pushPayload := payload.(gitea.PushPayload)
|
||||||
|
|
||||||
pe := pipeline_executor.PipelineExecution{
|
pe := pipeline_executor.PipelineExecution{
|
||||||
Name: webhookJobName,
|
Pipeline: pipeline,
|
||||||
Job: jobConfig,
|
|
||||||
Ref: pushPayload.Ref,
|
Ref: pushPayload.Ref,
|
||||||
}
|
}
|
||||||
|
go pipeline_executor.ExecutePipeline(pe, db, conf)
|
||||||
pipeline_executor.ExecutePipeline(pe, conf.PipelineConf)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return
|
|
||||||
default:
|
|
||||||
log.Errorf("Job configured with unknown webhook sender \"%v\", igonring...", jobConfig.Webhook.Sender)
|
|
||||||
return
|
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Errorf("Not job configured with name \"%v\", required by webhook with url \"%v\", ignoring...",
|
log.Errorf("No webhook found with ID \"%v\"", webhookUUID)
|
||||||
webhookJobName, r.URL)
|
|
||||||
|
|
||||||
default:
|
default:
|
||||||
log.Errorf("Got request with method \"%v\", ignoring...", r.Method)
|
log.Errorf("Got request with method \"%v\", ignoring...", r.Method)
|
||||||
}
|
}
|
||||||
})
|
}
|
||||||
|
|
||||||
|
func CreateWebhookHandler(db database.Database, conf config.PipelineConf, mux *http.ServeMux) {
|
||||||
|
mux.HandleFunc("/webhook/", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
webhookHandler(w, r, db, conf)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user