Compare commits
8 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
39509e9ad0 | ||
|
|
9924aea786 | ||
|
|
65c232c4a5 | ||
|
|
5da4954b65 | ||
|
|
ec091ad269 | ||
|
|
1656206765 | ||
|
|
6cdf1e5788 | ||
|
|
ab381649da |
@@ -16,6 +16,7 @@ func NewInterpeter(
|
|||||||
gitCtx *model.GithubContext,
|
gitCtx *model.GithubContext,
|
||||||
results map[string]*JobResult,
|
results map[string]*JobResult,
|
||||||
vars map[string]string,
|
vars map[string]string,
|
||||||
|
inputs map[string]interface{},
|
||||||
) exprparser.Interpreter {
|
) exprparser.Interpreter {
|
||||||
strategy := make(map[string]interface{})
|
strategy := make(map[string]interface{})
|
||||||
if job.Strategy != nil {
|
if job.Strategy != nil {
|
||||||
@@ -62,7 +63,7 @@ func NewInterpeter(
|
|||||||
Strategy: strategy,
|
Strategy: strategy,
|
||||||
Matrix: matrix,
|
Matrix: matrix,
|
||||||
Needs: using,
|
Needs: using,
|
||||||
Inputs: nil, // not supported yet
|
Inputs: inputs,
|
||||||
Vars: vars,
|
Vars: vars,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ import (
|
|||||||
|
|
||||||
"gopkg.in/yaml.v3"
|
"gopkg.in/yaml.v3"
|
||||||
|
|
||||||
|
"github.com/nektos/act/pkg/exprparser"
|
||||||
"github.com/nektos/act/pkg/model"
|
"github.com/nektos/act/pkg/model"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -40,6 +41,10 @@ func Parse(content []byte, options ...ParseOption) ([]*SingleWorkflow, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("invalid jobs: %w", err)
|
return nil, fmt.Errorf("invalid jobs: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
evaluator := NewExpressionEvaluator(exprparser.NewInterpeter(&exprparser.EvaluationEnvironment{Github: pc.gitContext, Vars: pc.vars}, exprparser.Config{}))
|
||||||
|
workflow.RunName = evaluator.Interpolate(workflow.RunName)
|
||||||
|
|
||||||
for i, id := range ids {
|
for i, id := range ids {
|
||||||
job := jobs[i]
|
job := jobs[i]
|
||||||
matricxes, err := getMatrixes(origin.GetJob(id))
|
matricxes, err := getMatrixes(origin.GetJob(id))
|
||||||
@@ -52,7 +57,7 @@ func Parse(content []byte, options ...ParseOption) ([]*SingleWorkflow, error) {
|
|||||||
job.Name = id
|
job.Name = id
|
||||||
}
|
}
|
||||||
job.Strategy.RawMatrix = encodeMatrix(matrix)
|
job.Strategy.RawMatrix = encodeMatrix(matrix)
|
||||||
evaluator := NewExpressionEvaluator(NewInterpeter(id, origin.GetJob(id), matrix, pc.gitContext, results, pc.vars))
|
evaluator := NewExpressionEvaluator(NewInterpeter(id, origin.GetJob(id), matrix, pc.gitContext, results, pc.vars, nil))
|
||||||
job.Name = nameWithMatrix(job.Name, matrix, evaluator)
|
job.Name = nameWithMatrix(job.Name, matrix, evaluator)
|
||||||
runsOn := origin.GetJob(id).RunsOn()
|
runsOn := origin.GetJob(id).RunsOn()
|
||||||
for i, v := range runsOn {
|
for i, v := range runsOn {
|
||||||
@@ -60,10 +65,12 @@ func Parse(content []byte, options ...ParseOption) ([]*SingleWorkflow, error) {
|
|||||||
}
|
}
|
||||||
job.RawRunsOn = encodeRunsOn(runsOn)
|
job.RawRunsOn = encodeRunsOn(runsOn)
|
||||||
swf := &SingleWorkflow{
|
swf := &SingleWorkflow{
|
||||||
Name: workflow.Name,
|
Name: workflow.Name,
|
||||||
RawOn: workflow.RawOn,
|
RawOn: workflow.RawOn,
|
||||||
Env: workflow.Env,
|
Env: workflow.Env,
|
||||||
Defaults: workflow.Defaults,
|
Defaults: workflow.Defaults,
|
||||||
|
RawPermissions: workflow.RawPermissions,
|
||||||
|
RunName: workflow.RunName,
|
||||||
}
|
}
|
||||||
if err := swf.SetJob(id, job); err != nil {
|
if err := swf.SetJob(id, job); err != nil {
|
||||||
return nil, fmt.Errorf("SetJob: %w", err)
|
return nil, fmt.Errorf("SetJob: %w", err)
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
package jobparser
|
package jobparser
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"github.com/nektos/act/pkg/model"
|
"github.com/nektos/act/pkg/model"
|
||||||
@@ -9,11 +10,13 @@ import (
|
|||||||
|
|
||||||
// SingleWorkflow is a workflow with single job and single matrix
|
// SingleWorkflow is a workflow with single job and single matrix
|
||||||
type SingleWorkflow struct {
|
type SingleWorkflow struct {
|
||||||
Name string `yaml:"name,omitempty"`
|
Name string `yaml:"name,omitempty"`
|
||||||
RawOn yaml.Node `yaml:"on,omitempty"`
|
RawOn yaml.Node `yaml:"on,omitempty"`
|
||||||
Env map[string]string `yaml:"env,omitempty"`
|
Env map[string]string `yaml:"env,omitempty"`
|
||||||
RawJobs yaml.Node `yaml:"jobs,omitempty"`
|
RawJobs yaml.Node `yaml:"jobs,omitempty"`
|
||||||
Defaults Defaults `yaml:"defaults,omitempty"`
|
Defaults Defaults `yaml:"defaults,omitempty"`
|
||||||
|
RawPermissions yaml.Node `yaml:"permissions,omitempty"`
|
||||||
|
RunName string `yaml:"run-name,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *SingleWorkflow) Job() (string, *Job) {
|
func (w *SingleWorkflow) Job() (string, *Job) {
|
||||||
@@ -82,6 +85,8 @@ type Job struct {
|
|||||||
Uses string `yaml:"uses,omitempty"`
|
Uses string `yaml:"uses,omitempty"`
|
||||||
With map[string]interface{} `yaml:"with,omitempty"`
|
With map[string]interface{} `yaml:"with,omitempty"`
|
||||||
RawSecrets yaml.Node `yaml:"secrets,omitempty"`
|
RawSecrets yaml.Node `yaml:"secrets,omitempty"`
|
||||||
|
RawConcurrency *model.RawConcurrency `yaml:"concurrency,omitempty"`
|
||||||
|
RawPermissions yaml.Node `yaml:"permissions,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func (j *Job) Clone() *Job {
|
func (j *Job) Clone() *Job {
|
||||||
@@ -104,6 +109,8 @@ func (j *Job) Clone() *Job {
|
|||||||
Uses: j.Uses,
|
Uses: j.Uses,
|
||||||
With: j.With,
|
With: j.With,
|
||||||
RawSecrets: j.RawSecrets,
|
RawSecrets: j.RawSecrets,
|
||||||
|
RawConcurrency: j.RawConcurrency,
|
||||||
|
RawPermissions: j.RawPermissions,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -172,10 +179,20 @@ type RunDefaults struct {
|
|||||||
WorkingDirectory string `yaml:"working-directory,omitempty"`
|
WorkingDirectory string `yaml:"working-directory,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type WorkflowDispatchInput struct {
|
||||||
|
Name string `yaml:"name"`
|
||||||
|
Description string `yaml:"description"`
|
||||||
|
Required bool `yaml:"required"`
|
||||||
|
Default string `yaml:"default"`
|
||||||
|
Type string `yaml:"type"`
|
||||||
|
Options []string `yaml:"options"`
|
||||||
|
}
|
||||||
|
|
||||||
type Event struct {
|
type Event struct {
|
||||||
Name string
|
Name string
|
||||||
acts map[string][]string
|
acts map[string][]string
|
||||||
schedules []map[string]string
|
schedules []map[string]string
|
||||||
|
inputs []WorkflowDispatchInput
|
||||||
}
|
}
|
||||||
|
|
||||||
func (evt *Event) IsSchedule() bool {
|
func (evt *Event) IsSchedule() bool {
|
||||||
@@ -190,6 +207,126 @@ func (evt *Event) Schedules() []map[string]string {
|
|||||||
return evt.schedules
|
return evt.schedules
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (evt *Event) Inputs() []WorkflowDispatchInput {
|
||||||
|
return evt.inputs
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseWorkflowDispatchInputs(inputs map[string]interface{}) ([]WorkflowDispatchInput, error) {
|
||||||
|
var results []WorkflowDispatchInput
|
||||||
|
for name, input := range inputs {
|
||||||
|
inputMap, ok := input.(map[string]interface{})
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("invalid input: %v", input)
|
||||||
|
}
|
||||||
|
input := WorkflowDispatchInput{
|
||||||
|
Name: name,
|
||||||
|
}
|
||||||
|
if desc, ok := inputMap["description"].(string); ok {
|
||||||
|
input.Description = desc
|
||||||
|
}
|
||||||
|
if required, ok := inputMap["required"].(bool); ok {
|
||||||
|
input.Required = required
|
||||||
|
}
|
||||||
|
if defaultVal, ok := inputMap["default"].(string); ok {
|
||||||
|
input.Default = defaultVal
|
||||||
|
}
|
||||||
|
if inputType, ok := inputMap["type"].(string); ok {
|
||||||
|
input.Type = inputType
|
||||||
|
}
|
||||||
|
if options, ok := inputMap["options"].([]string); ok {
|
||||||
|
input.Options = options
|
||||||
|
} else if options, ok := inputMap["options"].([]interface{}); ok {
|
||||||
|
for _, option := range options {
|
||||||
|
if opt, ok := option.(string); ok {
|
||||||
|
input.Options = append(input.Options, opt)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
results = append(results, input)
|
||||||
|
}
|
||||||
|
return results, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func ReadWorkflowRawConcurrency(content []byte) (*model.RawConcurrency, error) {
|
||||||
|
w := new(model.Workflow)
|
||||||
|
err := yaml.NewDecoder(bytes.NewReader(content)).Decode(w)
|
||||||
|
return w.RawConcurrency, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func EvaluateConcurrency(rc *model.RawConcurrency, jobID string, job *Job, gitCtx map[string]any, results map[string]*JobResult, vars map[string]string, inputs map[string]any) (string, bool, error) {
|
||||||
|
actJob := &model.Job{}
|
||||||
|
if job != nil {
|
||||||
|
actJob.Strategy = &model.Strategy{
|
||||||
|
FailFastString: job.Strategy.FailFastString,
|
||||||
|
MaxParallelString: job.Strategy.MaxParallelString,
|
||||||
|
RawMatrix: job.Strategy.RawMatrix,
|
||||||
|
}
|
||||||
|
actJob.Strategy.FailFast = actJob.Strategy.GetFailFast()
|
||||||
|
actJob.Strategy.MaxParallel = actJob.Strategy.GetMaxParallel()
|
||||||
|
}
|
||||||
|
|
||||||
|
matrix := make(map[string]any)
|
||||||
|
matrixes, err := actJob.GetMatrixes()
|
||||||
|
if err != nil {
|
||||||
|
return "", false, err
|
||||||
|
}
|
||||||
|
if len(matrixes) > 0 {
|
||||||
|
matrix = matrixes[0]
|
||||||
|
}
|
||||||
|
|
||||||
|
evaluator := NewExpressionEvaluator(NewInterpeter(jobID, actJob, matrix, toGitContext(gitCtx), results, vars, inputs))
|
||||||
|
var node yaml.Node
|
||||||
|
if err := node.Encode(rc); err != nil {
|
||||||
|
return "", false, fmt.Errorf("failed to encode concurrency: %w", err)
|
||||||
|
}
|
||||||
|
if err := evaluator.EvaluateYamlNode(&node); err != nil {
|
||||||
|
return "", false, fmt.Errorf("failed to evaluate concurrency: %w", err)
|
||||||
|
}
|
||||||
|
var evaluated model.RawConcurrency
|
||||||
|
if err := node.Decode(&evaluated); err != nil {
|
||||||
|
return "", false, fmt.Errorf("failed to unmarshal evaluated concurrency: %w", err)
|
||||||
|
}
|
||||||
|
if evaluated.RawExpression != "" {
|
||||||
|
return evaluated.RawExpression, false, nil
|
||||||
|
}
|
||||||
|
return evaluated.Group, evaluated.CancelInProgress == "true", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func toGitContext(input map[string]any) *model.GithubContext {
|
||||||
|
gitContext := &model.GithubContext{
|
||||||
|
EventPath: asString(input["event_path"]),
|
||||||
|
Workflow: asString(input["workflow"]),
|
||||||
|
RunID: asString(input["run_id"]),
|
||||||
|
RunNumber: asString(input["run_number"]),
|
||||||
|
Actor: asString(input["actor"]),
|
||||||
|
Repository: asString(input["repository"]),
|
||||||
|
EventName: asString(input["event_name"]),
|
||||||
|
Sha: asString(input["sha"]),
|
||||||
|
Ref: asString(input["ref"]),
|
||||||
|
RefName: asString(input["ref_name"]),
|
||||||
|
RefType: asString(input["ref_type"]),
|
||||||
|
HeadRef: asString(input["head_ref"]),
|
||||||
|
BaseRef: asString(input["base_ref"]),
|
||||||
|
Token: asString(input["token"]),
|
||||||
|
Workspace: asString(input["workspace"]),
|
||||||
|
Action: asString(input["action"]),
|
||||||
|
ActionPath: asString(input["action_path"]),
|
||||||
|
ActionRef: asString(input["action_ref"]),
|
||||||
|
ActionRepository: asString(input["action_repository"]),
|
||||||
|
Job: asString(input["job"]),
|
||||||
|
RepositoryOwner: asString(input["repository_owner"]),
|
||||||
|
RetentionDays: asString(input["retention_days"]),
|
||||||
|
}
|
||||||
|
|
||||||
|
event, ok := input["event"].(map[string]any)
|
||||||
|
if ok {
|
||||||
|
gitContext.Event = event
|
||||||
|
}
|
||||||
|
|
||||||
|
return gitContext
|
||||||
|
}
|
||||||
|
|
||||||
func ParseRawOn(rawOn *yaml.Node) ([]*Event, error) {
|
func ParseRawOn(rawOn *yaml.Node) ([]*Event, error) {
|
||||||
switch rawOn.Kind {
|
switch rawOn.Kind {
|
||||||
case yaml.ScalarNode:
|
case yaml.ScalarNode:
|
||||||
@@ -218,79 +355,126 @@ func ParseRawOn(rawOn *yaml.Node) ([]*Event, error) {
|
|||||||
}
|
}
|
||||||
return res, nil
|
return res, nil
|
||||||
case yaml.MappingNode:
|
case yaml.MappingNode:
|
||||||
events, triggers, err := parseMappingNode[interface{}](rawOn)
|
events, triggers, err := parseMappingNode[yaml.Node](rawOn)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
res := make([]*Event, 0, len(events))
|
res := make([]*Event, 0, len(events))
|
||||||
for i, k := range events {
|
for i, k := range events {
|
||||||
v := triggers[i]
|
v := triggers[i]
|
||||||
if v == nil {
|
switch v.Kind {
|
||||||
|
case yaml.ScalarNode:
|
||||||
res = append(res, &Event{
|
res = append(res, &Event{
|
||||||
Name: k,
|
Name: k,
|
||||||
acts: map[string][]string{},
|
|
||||||
})
|
})
|
||||||
continue
|
case yaml.SequenceNode:
|
||||||
}
|
var t []interface{}
|
||||||
switch t := v.(type) {
|
err := v.Decode(&t)
|
||||||
case string:
|
if err != nil {
|
||||||
res = append(res, &Event{
|
return nil, err
|
||||||
Name: k,
|
|
||||||
acts: map[string][]string{},
|
|
||||||
})
|
|
||||||
case []string:
|
|
||||||
res = append(res, &Event{
|
|
||||||
Name: k,
|
|
||||||
acts: map[string][]string{},
|
|
||||||
})
|
|
||||||
case map[string]interface{}:
|
|
||||||
acts := make(map[string][]string, len(t))
|
|
||||||
for act, branches := range t {
|
|
||||||
switch b := branches.(type) {
|
|
||||||
case string:
|
|
||||||
acts[act] = []string{b}
|
|
||||||
case []string:
|
|
||||||
acts[act] = b
|
|
||||||
case []interface{}:
|
|
||||||
acts[act] = make([]string, len(b))
|
|
||||||
for i, v := range b {
|
|
||||||
var ok bool
|
|
||||||
if acts[act][i], ok = v.(string); !ok {
|
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", branches)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", branches)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
res = append(res, &Event{
|
|
||||||
Name: k,
|
|
||||||
acts: acts,
|
|
||||||
})
|
|
||||||
case []interface{}:
|
|
||||||
if k != "schedule" {
|
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", v)
|
|
||||||
}
|
}
|
||||||
schedules := make([]map[string]string, len(t))
|
schedules := make([]map[string]string, len(t))
|
||||||
for i, tt := range t {
|
if k == "schedule" {
|
||||||
vv, ok := tt.(map[string]interface{})
|
for i, tt := range t {
|
||||||
if !ok {
|
vv, ok := tt.(map[string]interface{})
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", v)
|
if !ok {
|
||||||
}
|
return nil, fmt.Errorf("unknown on type(schedule): %#v", v)
|
||||||
schedules[i] = make(map[string]string, len(vv))
|
}
|
||||||
for k, vvv := range vv {
|
schedules[i] = make(map[string]string, len(vv))
|
||||||
var ok bool
|
for k, vvv := range vv {
|
||||||
if schedules[i][k], ok = vvv.(string); !ok {
|
var ok bool
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", v)
|
if schedules[i][k], ok = vvv.(string); !ok {
|
||||||
|
return nil, fmt.Errorf("unknown on type(schedule): %#v", v)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if len(schedules) == 0 {
|
||||||
|
schedules = nil
|
||||||
|
}
|
||||||
res = append(res, &Event{
|
res = append(res, &Event{
|
||||||
Name: k,
|
Name: k,
|
||||||
schedules: schedules,
|
schedules: schedules,
|
||||||
})
|
})
|
||||||
|
case yaml.MappingNode:
|
||||||
|
acts := make(map[string][]string, len(v.Content)/2)
|
||||||
|
var inputs []WorkflowDispatchInput
|
||||||
|
expectedKey := true
|
||||||
|
var act string
|
||||||
|
for _, content := range v.Content {
|
||||||
|
if expectedKey {
|
||||||
|
if content.Kind != yaml.ScalarNode {
|
||||||
|
return nil, fmt.Errorf("key type not string: %#v", content)
|
||||||
|
}
|
||||||
|
act = ""
|
||||||
|
err := content.Decode(&act)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
switch content.Kind {
|
||||||
|
case yaml.SequenceNode:
|
||||||
|
var t []string
|
||||||
|
err := content.Decode(&t)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
acts[act] = t
|
||||||
|
case yaml.ScalarNode:
|
||||||
|
var t string
|
||||||
|
err := content.Decode(&t)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
acts[act] = []string{t}
|
||||||
|
case yaml.MappingNode:
|
||||||
|
if k != "workflow_dispatch" || act != "inputs" {
|
||||||
|
return nil, fmt.Errorf("map should only for workflow_dispatch but %s: %#v", act, content)
|
||||||
|
}
|
||||||
|
|
||||||
|
var key string
|
||||||
|
for i, vv := range content.Content {
|
||||||
|
if i%2 == 0 {
|
||||||
|
if vv.Kind != yaml.ScalarNode {
|
||||||
|
return nil, fmt.Errorf("key type not string: %#v", vv)
|
||||||
|
}
|
||||||
|
key = ""
|
||||||
|
if err := vv.Decode(&key); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if vv.Kind != yaml.MappingNode {
|
||||||
|
return nil, fmt.Errorf("key type not map(%s): %#v", key, vv)
|
||||||
|
}
|
||||||
|
|
||||||
|
input := WorkflowDispatchInput{}
|
||||||
|
if err := vv.Decode(&input); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
input.Name = key
|
||||||
|
inputs = append(inputs, input)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("unknown on type: %#v", content)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
expectedKey = !expectedKey
|
||||||
|
}
|
||||||
|
if len(inputs) == 0 {
|
||||||
|
inputs = nil
|
||||||
|
}
|
||||||
|
if len(acts) == 0 {
|
||||||
|
acts = nil
|
||||||
|
}
|
||||||
|
res = append(res, &Event{
|
||||||
|
Name: k,
|
||||||
|
acts: acts,
|
||||||
|
inputs: inputs,
|
||||||
|
})
|
||||||
default:
|
default:
|
||||||
return nil, fmt.Errorf("unknown on type: %#v", v)
|
return nil, fmt.Errorf("unknown on type: %v", v.Kind)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return res, nil
|
return res, nil
|
||||||
@@ -331,3 +515,12 @@ func parseMappingNode[T any](node *yaml.Node) ([]string, []T, error) {
|
|||||||
|
|
||||||
return scalars, datas, nil
|
return scalars, datas, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func asString(v interface{}) string {
|
||||||
|
if v == nil {
|
||||||
|
return ""
|
||||||
|
} else if s, ok := v.(string); ok {
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|||||||
@@ -58,6 +58,19 @@ func TestParseRawOn(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
input: "on:\n push:\n branches: main",
|
||||||
|
result: []*Event{
|
||||||
|
{
|
||||||
|
Name: "push",
|
||||||
|
acts: map[string][]string{
|
||||||
|
"branches": {
|
||||||
|
"main",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
{
|
{
|
||||||
input: "on:\n branch_protection_rule:\n types: [created, deleted]",
|
input: "on:\n branch_protection_rule:\n types: [created, deleted]",
|
||||||
result: []*Event{
|
result: []*Event{
|
||||||
@@ -186,6 +199,60 @@ func TestParseRawOn(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
input: `on:
|
||||||
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
logLevel:
|
||||||
|
description: 'Log level'
|
||||||
|
required: true
|
||||||
|
default: 'warning'
|
||||||
|
type: choice
|
||||||
|
options:
|
||||||
|
- info
|
||||||
|
- warning
|
||||||
|
- debug
|
||||||
|
tags:
|
||||||
|
description: 'Test scenario tags'
|
||||||
|
required: false
|
||||||
|
type: boolean
|
||||||
|
environment:
|
||||||
|
description: 'Environment to run tests against'
|
||||||
|
type: environment
|
||||||
|
required: true
|
||||||
|
push:
|
||||||
|
`,
|
||||||
|
result: []*Event{
|
||||||
|
{
|
||||||
|
Name: "workflow_dispatch",
|
||||||
|
inputs: []WorkflowDispatchInput{
|
||||||
|
{
|
||||||
|
Name: "logLevel",
|
||||||
|
Description: "Log level",
|
||||||
|
Required: true,
|
||||||
|
Default: "warning",
|
||||||
|
Type: "choice",
|
||||||
|
Options: []string{"info", "warning", "debug"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "tags",
|
||||||
|
Description: "Test scenario tags",
|
||||||
|
Required: false,
|
||||||
|
Type: "boolean",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "environment",
|
||||||
|
Description: "Environment to run tests against",
|
||||||
|
Type: "environment",
|
||||||
|
Required: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "push",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
}
|
}
|
||||||
for _, kase := range kases {
|
for _, kase := range kases {
|
||||||
t.Run(kase.input, func(t *testing.T) {
|
t.Run(kase.input, func(t *testing.T) {
|
||||||
@@ -230,8 +297,7 @@ func TestParseMappingNode(t *testing.T) {
|
|||||||
{
|
{
|
||||||
input: "on:\n push:\n branches:\n - master",
|
input: "on:\n push:\n branches:\n - master",
|
||||||
scalars: []string{"push"},
|
scalars: []string{"push"},
|
||||||
datas: []interface {
|
datas: []interface{}{
|
||||||
}{
|
|
||||||
map[string]interface{}{
|
map[string]interface{}{
|
||||||
"branches": []interface{}{"master"},
|
"branches": []interface{}{"master"},
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -17,12 +17,14 @@ import (
|
|||||||
|
|
||||||
// Workflow is the structure of the files in .github/workflows
|
// Workflow is the structure of the files in .github/workflows
|
||||||
type Workflow struct {
|
type Workflow struct {
|
||||||
File string
|
File string
|
||||||
Name string `yaml:"name"`
|
Name string `yaml:"name"`
|
||||||
RawOn yaml.Node `yaml:"on"`
|
RawOn yaml.Node `yaml:"on"`
|
||||||
Env map[string]string `yaml:"env"`
|
Env map[string]string `yaml:"env"`
|
||||||
Jobs map[string]*Job `yaml:"jobs"`
|
Jobs map[string]*Job `yaml:"jobs"`
|
||||||
Defaults Defaults `yaml:"defaults"`
|
Defaults Defaults `yaml:"defaults"`
|
||||||
|
RawConcurrency *RawConcurrency `yaml:"concurrency"`
|
||||||
|
RawPermissions yaml.Node `yaml:"permissions"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// On events for the workflow
|
// On events for the workflow
|
||||||
@@ -199,6 +201,7 @@ type Job struct {
|
|||||||
Uses string `yaml:"uses"`
|
Uses string `yaml:"uses"`
|
||||||
With map[string]interface{} `yaml:"with"`
|
With map[string]interface{} `yaml:"with"`
|
||||||
RawSecrets yaml.Node `yaml:"secrets"`
|
RawSecrets yaml.Node `yaml:"secrets"`
|
||||||
|
RawPermissions yaml.Node `yaml:"permissions"`
|
||||||
Result string
|
Result string
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -769,3 +772,28 @@ func decodeNode(node yaml.Node, out interface{}) bool {
|
|||||||
}
|
}
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// For Gitea
|
||||||
|
// RawConcurrency represents a workflow concurrency or a job concurrency with uninterpolated options
|
||||||
|
type RawConcurrency struct {
|
||||||
|
Group string `yaml:"group,omitempty"`
|
||||||
|
CancelInProgress string `yaml:"cancel-in-progress,omitempty"`
|
||||||
|
RawExpression string `yaml:"-,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type objectConcurrency RawConcurrency
|
||||||
|
|
||||||
|
func (r *RawConcurrency) UnmarshalYAML(n *yaml.Node) error {
|
||||||
|
if err := n.Decode(&r.RawExpression); err == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return n.Decode((*objectConcurrency)(r))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RawConcurrency) MarshalYAML() (interface{}, error) {
|
||||||
|
if r.RawExpression != "" {
|
||||||
|
return r.RawExpression, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return (*objectConcurrency)(r), nil
|
||||||
|
}
|
||||||
|
|||||||
@@ -73,10 +73,6 @@ func newJobExecutor(info jobInfo, sf stepFactory, rc *RunContext) common.Executo
|
|||||||
|
|
||||||
preExec := step.pre()
|
preExec := step.pre()
|
||||||
preSteps = append(preSteps, useStepLogger(rc, stepModel, stepStagePre, func(ctx context.Context) error {
|
preSteps = append(preSteps, useStepLogger(rc, stepModel, stepStagePre, func(ctx context.Context) error {
|
||||||
if rc.caller != nil { // For Gitea
|
|
||||||
rc.caller.reusedWorkflowJobResults[rc.JobName] = "pending"
|
|
||||||
}
|
|
||||||
|
|
||||||
logger := common.Logger(ctx)
|
logger := common.Logger(ctx)
|
||||||
preErr := preExec(ctx)
|
preErr := preExec(ctx)
|
||||||
if preErr != nil {
|
if preErr != nil {
|
||||||
@@ -189,34 +185,7 @@ func setJobResult(ctx context.Context, info jobInfo, rc *RunContext, success boo
|
|||||||
info.result(jobResult)
|
info.result(jobResult)
|
||||||
if rc.caller != nil {
|
if rc.caller != nil {
|
||||||
// set reusable workflow job result
|
// set reusable workflow job result
|
||||||
|
rc.caller.setReusedWorkflowJobResult(rc.JobName, jobResult) // For Gitea
|
||||||
rc.caller.updateResultLock.Lock()
|
|
||||||
rc.caller.reusedWorkflowJobResults[rc.JobName] = jobResult
|
|
||||||
|
|
||||||
allJobDone := true
|
|
||||||
hasFailure := false
|
|
||||||
for _, result := range rc.caller.reusedWorkflowJobResults {
|
|
||||||
if result == "pending" {
|
|
||||||
allJobDone = false
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if result == "failure" {
|
|
||||||
hasFailure = true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if allJobDone {
|
|
||||||
reusedWorkflowJobResult := "success"
|
|
||||||
reusedWorkflowJobResultMessage := "succeeded"
|
|
||||||
if hasFailure {
|
|
||||||
reusedWorkflowJobResult = "failure"
|
|
||||||
reusedWorkflowJobResultMessage = "failed"
|
|
||||||
}
|
|
||||||
rc.caller.runContext.result(reusedWorkflowJobResult)
|
|
||||||
logger.WithField("jobResult", reusedWorkflowJobResult).Infof("\U0001F3C1 Job %s", reusedWorkflowJobResultMessage)
|
|
||||||
}
|
|
||||||
|
|
||||||
rc.caller.updateResultLock.Unlock()
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -175,7 +175,11 @@ func newReusableWorkflowExecutor(rc *RunContext, directory string, workflow stri
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
return runner.NewPlanExecutor(plan)(ctx)
|
// return runner.NewPlanExecutor(plan)(ctx)
|
||||||
|
return common.NewPipelineExecutor( // For Gitea
|
||||||
|
runner.NewPlanExecutor(plan),
|
||||||
|
setReusedWorkflowCallerResult(rc, runner),
|
||||||
|
)(ctx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -271,3 +275,47 @@ func newRemoteReusableWorkflow(uses string) *remoteReusableWorkflow {
|
|||||||
URL: "https://github.com",
|
URL: "https://github.com",
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// For Gitea
|
||||||
|
func setReusedWorkflowCallerResult(rc *RunContext, runner Runner) common.Executor {
|
||||||
|
return func(ctx context.Context) error {
|
||||||
|
logger := common.Logger(ctx)
|
||||||
|
|
||||||
|
runnerImpl, ok := runner.(*runnerImpl)
|
||||||
|
if !ok {
|
||||||
|
logger.Warn("Failed to get caller from runner")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
caller := runnerImpl.caller
|
||||||
|
|
||||||
|
allJobDone := true
|
||||||
|
hasFailure := false
|
||||||
|
for _, result := range caller.reusedWorkflowJobResults {
|
||||||
|
if result == "pending" {
|
||||||
|
allJobDone = false
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if result == "failure" {
|
||||||
|
hasFailure = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if allJobDone {
|
||||||
|
reusedWorkflowJobResult := "success"
|
||||||
|
reusedWorkflowJobResultMessage := "succeeded"
|
||||||
|
if hasFailure {
|
||||||
|
reusedWorkflowJobResult = "failure"
|
||||||
|
reusedWorkflowJobResultMessage = "failed"
|
||||||
|
}
|
||||||
|
|
||||||
|
if rc.caller != nil {
|
||||||
|
rc.caller.setReusedWorkflowJobResult(rc.JobName, reusedWorkflowJobResult)
|
||||||
|
} else {
|
||||||
|
rc.result(reusedWorkflowJobResult)
|
||||||
|
logger.WithField("jobResult", reusedWorkflowJobResult).Infof("\U0001F3C1 Job %s", reusedWorkflowJobResultMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -92,7 +92,12 @@ func (rc *RunContext) GetEnv() map[string]string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (rc *RunContext) jobContainerName() string {
|
func (rc *RunContext) jobContainerName() string {
|
||||||
return createSimpleContainerName(rc.Config.ContainerNamePrefix, "WORKFLOW-"+rc.Run.Workflow.Name, "JOB-"+rc.Name)
|
nameParts := []string{rc.Config.ContainerNamePrefix, "WORKFLOW-" + rc.Run.Workflow.Name, "JOB-" + rc.Name}
|
||||||
|
if rc.caller != nil {
|
||||||
|
nameParts = append(nameParts, "CALLED-BY-"+rc.caller.runContext.JobName)
|
||||||
|
}
|
||||||
|
// return createSimpleContainerName(rc.Config.ContainerNamePrefix, "WORKFLOW-"+rc.Run.Workflow.Name, "JOB-"+rc.Name)
|
||||||
|
return createSimpleContainerName(nameParts...) // For Gitea
|
||||||
}
|
}
|
||||||
|
|
||||||
// Deprecated: use `networkNameForGitea`
|
// Deprecated: use `networkNameForGitea`
|
||||||
@@ -653,6 +658,7 @@ func (rc *RunContext) Executor() (common.Executor, error) {
|
|||||||
return func(ctx context.Context) error {
|
return func(ctx context.Context) error {
|
||||||
res, err := rc.isEnabled(ctx)
|
res, err := rc.isEnabled(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
rc.caller.setReusedWorkflowJobResult(rc.JobName, "failure") // For Gitea
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if res {
|
if res {
|
||||||
@@ -748,6 +754,10 @@ func (rc *RunContext) isEnabled(ctx context.Context) (bool, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if !runJob {
|
if !runJob {
|
||||||
|
if rc.caller != nil { // For Gitea
|
||||||
|
rc.caller.setReusedWorkflowJobResult(rc.JobName, "skipped")
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
l.WithField("jobResult", "skipped").Debugf("Skipping job '%s' due to '%s'", job.Name, job.If.Value)
|
l.WithField("jobResult", "skipped").Debugf("Skipping job '%s' due to '%s'", job.Name, job.If.Value)
|
||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -210,6 +210,9 @@ func (runner *runnerImpl) NewPlanExecutor(plan *model.Plan) common.Executor {
|
|||||||
if len(rc.String()) > maxJobNameLen {
|
if len(rc.String()) > maxJobNameLen {
|
||||||
maxJobNameLen = len(rc.String())
|
maxJobNameLen = len(rc.String())
|
||||||
}
|
}
|
||||||
|
if rc.caller != nil { // For Gitea
|
||||||
|
rc.caller.setReusedWorkflowJobResult(rc.JobName, "pending")
|
||||||
|
}
|
||||||
stageExecutor = append(stageExecutor, func(ctx context.Context) error {
|
stageExecutor = append(stageExecutor, func(ctx context.Context) error {
|
||||||
jobName := fmt.Sprintf("%-*s", maxJobNameLen, rc.String())
|
jobName := fmt.Sprintf("%-*s", maxJobNameLen, rc.String())
|
||||||
executor, err := rc.Executor()
|
executor, err := rc.Executor()
|
||||||
@@ -281,3 +284,10 @@ func (runner *runnerImpl) newRunContext(ctx context.Context, run *model.Run, mat
|
|||||||
|
|
||||||
return rc
|
return rc
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// For Gitea
|
||||||
|
func (c *caller) setReusedWorkflowJobResult(jobName string, result string) {
|
||||||
|
c.updateResultLock.Lock()
|
||||||
|
defer c.updateResultLock.Unlock()
|
||||||
|
c.reusedWorkflowJobResults[jobName] = result
|
||||||
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user