summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorValery Piashchynski <[email protected]>2020-10-27 15:58:46 +0300
committerValery Piashchynski <[email protected]>2020-10-27 15:58:46 +0300
commitc9af916ae4d78334d348ed1ef7238206f3ecb7a1 (patch)
tree43c04f7c538ecbd0b29fc280d41df5d00f203436
parentfd60aff1cddf32b0c45d934fcf14b070df80adcf (diff)
parent105bde0e0c1a7c133d1daa10603ca5ce9a9ade4d (diff)
Merge remote-tracking branch 'origin/release_2.0' into feature/pool_supervisor
# Conflicts: # static_pool.go # sync_worker.go
-rwxr-xr-xerrors.go6
-rwxr-xr-xerrors_test.go2
-rwxr-xr-xpool.go3
-rwxr-xr-xstatic_pool.go89
-rwxr-xr-xstatic_pool_test.go41
-rwxr-xr-xsync_worker.go10
-rwxr-xr-xsync_worker_test.go2
-rwxr-xr-xworker_watcher.go2
8 files changed, 109 insertions, 46 deletions
diff --git a/errors.go b/errors.go
index 52356549..7c91a92b 100755
--- a/errors.go
+++ b/errors.go
@@ -1,11 +1,11 @@
package roadrunner
-// JobError is job level error (no WorkerProcess halt), wraps at top
+// ExecError is job level error (no WorkerProcess halt), wraps at top
// of error context
-type JobError []byte
+type ExecError []byte
// Error converts error context to string
-func (te JobError) Error() string {
+func (te ExecError) Error() string {
return string(te)
}
diff --git a/errors_test.go b/errors_test.go
index 75a86840..86ab908d 100755
--- a/errors_test.go
+++ b/errors_test.go
@@ -8,7 +8,7 @@ import (
)
func Test_JobError_Error(t *testing.T) {
- e := JobError([]byte("error"))
+ e := ExecError([]byte("error"))
assert.Equal(t, "error", e.Error())
}
diff --git a/pool.go b/pool.go
index 721b67c1..a95b8cfb 100755
--- a/pool.go
+++ b/pool.go
@@ -68,6 +68,9 @@ type Pool interface {
// Configures the pool behaviour.
type Config struct {
+ // Debug flag creates new fresh worker before every request.
+ Debug bool
+
// NumWorkers defines how many sub-processes can be run at once. This value
// might be doubled by Swapper while hot-swap. Defaults to number of CPU cores.
NumWorkers int64
diff --git a/static_pool.go b/static_pool.go
index 3af933c3..7a5f6103 100755
--- a/static_pool.go
+++ b/static_pool.go
@@ -39,6 +39,11 @@ type StaticPool struct {
func NewPool(ctx context.Context, cmd func() *exec.Cmd, factory Factory, cfg Config) (Pool, error) {
cfg.InitDefaults()
+ if cfg.Debug {
+ cfg.NumWorkers = 0
+ cfg.MaxJobs = 1
+ }
+
p := &StaticPool{
cfg: cfg,
cmd: cmd,
@@ -82,27 +87,30 @@ func NewPool(ctx context.Context, cmd func() *exec.Cmd, factory Factory, cfg Con
}
// AddListener connects event listener to the pool.
-func (p *StaticPool) AddListener(listener util.EventListener) {
- p.events.AddListener(listener)
+func (sp *StaticPool) AddListener(listener util.EventListener) {
+ sp.events.AddListener(listener)
}
// Config returns associated pool configuration. Immutable.
-func (p *StaticPool) GetConfig() Config {
- return p.cfg
+func (sp *StaticPool) GetConfig() Config {
+ return sp.cfg
}
// Workers returns worker list associated with the pool.
-func (p *StaticPool) Workers() (workers []WorkerBase) {
- return p.ww.WorkersList()
+func (sp *StaticPool) Workers() (workers []WorkerBase) {
+ return sp.ww.WorkersList()
}
-func (p *StaticPool) RemoveWorker(ctx context.Context, wb WorkerBase) error {
- return p.ww.RemoveWorker(ctx, wb)
+func (sp *StaticPool) RemoveWorker(ctx context.Context, wb WorkerBase) error {
+ return sp.ww.RemoveWorker(ctx, wb)
}
-func (p *StaticPool) Exec(rqs Payload) (Payload, error) {
+func (sp *StaticPool) Exec(p Payload) (Payload, error) {
const op = errors.Op("Exec")
- w, err := p.ww.GetFreeWorker(context.Background())
+ if sp.cfg.Debug {
+ return sp.execDebug(p)
+ }
+ w, err := sp.ww.GetFreeWorker(context.Background())
if err != nil && errors.Is(errors.ErrWatcherStopped, err) {
return EmptyPayload, errors.E(op, err)
} else if err != nil {
@@ -111,30 +119,30 @@ func (p *StaticPool) Exec(rqs Payload) (Payload, error) {
sw := w.(SyncWorker)
- rsp, err := sw.Exec(rqs)
+ rsp, err := sw.Exec(p)
if err != nil {
// soft job errors are allowed
- if _, jobError := err.(JobError); jobError {
- if p.cfg.MaxJobs != 0 && w.State().NumExecs() >= p.cfg.MaxJobs {
- err := p.ww.AllocateNew(bCtx)
+ if _, jobError := err.(ExecError); jobError {
+ if sp.cfg.MaxJobs != 0 && w.State().NumExecs() >= sp.cfg.MaxJobs {
+ err := sp.ww.AllocateNew(bCtx)
if err != nil {
- p.events.Push(PoolEvent{Event: EventPoolError, Payload: err})
+ sp.events.Push(PoolEvent{Event: EventPoolError, Payload: err})
}
w.State().Set(StateInvalid)
err = w.Stop(bCtx)
if err != nil {
- p.events.Push(WorkerEvent{Event: EventWorkerError, Worker: w, Payload: err})
+ sp.events.Push(WorkerEvent{Event: EventWorkerError, Worker: w, Payload: err})
}
} else {
- p.ww.PushWorker(w)
+ sp.ww.PushWorker(w)
}
return EmptyPayload, err
}
sw.State().Set(StateInvalid)
- p.events.Push(PoolEvent{Event: EventWorkerDestruct, Payload: w})
+ sp.events.Push(PoolEvent{Event: EventWorkerDestruct, Payload: w})
errS := w.Stop(bCtx)
if errS != nil {
@@ -149,25 +157,38 @@ func (p *StaticPool) Exec(rqs Payload) (Payload, error) {
w.State().Set(StateInvalid)
err = w.Stop(bCtx)
if err != nil {
- p.events.Push(WorkerEvent{Event: EventWorkerError, Worker: w, Payload: err})
+ sp.events.Push(WorkerEvent{Event: EventWorkerError, Worker: w, Payload: err})
}
- return p.Exec(rqs)
+ return sp.Exec(p)
}
- if p.cfg.MaxJobs != 0 && w.State().NumExecs() >= p.cfg.MaxJobs {
- err = p.ww.AllocateNew(bCtx)
+ if sp.cfg.MaxJobs != 0 && w.State().NumExecs() >= sp.cfg.MaxJobs {
+ err = sp.ww.AllocateNew(bCtx)
if err != nil {
return EmptyPayload, err
}
} else {
- p.muw.Lock()
- p.ww.PushWorker(w)
- p.muw.Unlock()
+ sp.ww.PushWorker(w)
}
return rsp, nil
}
+func (sp *StaticPool) execDebug(p Payload) (Payload, error) {
+ sw, err := sp.ww.allocator()
+ if err != nil {
+ return EmptyPayload, err
+ }
+
+ r, err := sw.(SyncWorker).Exec(p)
+
+ if stopErr := sw.Stop(context.Background()); stopErr != nil {
+ sp.events.Push(WorkerEvent{Event: EventWorkerError, Worker: sw, Payload: err})
+ }
+
+ return r, err
+}
+
func (p *StaticPool) ExecWithContext(ctx context.Context, rqs Payload) (Payload, error) {
const op = errors.Op("Exec")
w, err := p.ww.GetFreeWorker(context.Background())
@@ -237,18 +258,18 @@ func (p *StaticPool) ExecWithContext(ctx context.Context, rqs Payload) (Payload,
}
// Destroy all underlying stack (but let them to complete the task).
-func (p *StaticPool) Destroy(ctx context.Context) {
- p.ww.Destroy(ctx)
+func (sp *StaticPool) Destroy(ctx context.Context) {
+ sp.ww.Destroy(ctx)
}
// allocate required number of stack
-func (p *StaticPool) allocateWorkers(ctx context.Context, numWorkers int64) ([]WorkerBase, error) {
+func (sp *StaticPool) allocateWorkers(ctx context.Context, numWorkers int64) ([]WorkerBase, error) {
var workers []WorkerBase
// constant number of stack simplify logic
for i := int64(0); i < numWorkers; i++ {
- ctx, cancel := context.WithTimeout(ctx, p.cfg.AllocateTimeout)
- w, err := p.factory.SpawnWorkerWithContext(ctx, p.cmd())
+ ctx, cancel := context.WithTimeout(ctx, sp.cfg.AllocateTimeout)
+ w, err := sp.factory.SpawnWorkerWithContext(ctx, sp.cmd())
if err != nil {
cancel()
return nil, err
@@ -259,11 +280,11 @@ func (p *StaticPool) allocateWorkers(ctx context.Context, numWorkers int64) ([]W
return workers, nil
}
-func (p *StaticPool) checkMaxJobs(ctx context.Context, w WorkerBase) error {
- if p.cfg.MaxJobs != 0 && w.State().NumExecs() >= p.cfg.MaxJobs {
- err := p.ww.AllocateNew(ctx)
+func (sp *StaticPool) checkMaxJobs(ctx context.Context, w WorkerBase) error {
+ if sp.cfg.MaxJobs != 0 && w.State().NumExecs() >= sp.cfg.MaxJobs {
+ err := sp.ww.AllocateNew(ctx)
if err != nil {
- p.events.Push(PoolEvent{Event: EventPoolError, Payload: err})
+ sp.events.Push(PoolEvent{Event: EventPoolError, Payload: err})
return err
}
}
diff --git a/static_pool_test.go b/static_pool_test.go
index 8633f9c5..f1e3e4e4 100755
--- a/static_pool_test.go
+++ b/static_pool_test.go
@@ -152,7 +152,7 @@ func Test_StaticPool_JobError(t *testing.T) {
assert.Nil(t, res.Body)
assert.Nil(t, res.Context)
- assert.IsType(t, JobError{}, err)
+ assert.IsType(t, ExecError{}, err)
assert.Equal(t, "hello", err.Error())
}
@@ -301,6 +301,45 @@ func Test_StaticPool_Replace_Worker(t *testing.T) {
}
}
+func Test_StaticPool_Debug_Worker(t *testing.T) {
+ ctx := context.Background()
+ p, err := NewPool(
+ ctx,
+ func() *exec.Cmd { return exec.Command("php", "tests/client.php", "pid", "pipes") },
+ NewPipeFactory(),
+ Config{
+ Debug: true,
+ AllocateTimeout: time.Second,
+ DestroyTimeout: time.Second,
+ },
+ )
+ assert.NoError(t, err)
+ defer p.Destroy(ctx)
+
+ assert.NotNil(t, p)
+
+ assert.Len(t, p.Workers(), 0)
+
+ var lastPID string
+ res, _ := p.Exec(Payload{Body: []byte("hello")})
+ assert.NotEqual(t, lastPID, string(res.Body))
+
+ assert.Len(t, p.Workers(), 0)
+
+ for i := 0; i < 10; i++ {
+ assert.Len(t, p.Workers(), 0)
+ res, err := p.Exec(Payload{Body: []byte("hello")})
+
+ assert.NoError(t, err)
+ assert.NotNil(t, res)
+ assert.NotNil(t, res.Body)
+ assert.Nil(t, res.Context)
+
+ assert.NotEqual(t, lastPID, string(res.Body))
+ lastPID = string(res.Body)
+ }
+}
+
// identical to replace but controlled on worker side
func Test_StaticPool_Stop_Worker(t *testing.T) {
ctx := context.Background()
diff --git a/sync_worker.go b/sync_worker.go
index 31d68168..6dd8d8e8 100755
--- a/sync_worker.go
+++ b/sync_worker.go
@@ -49,7 +49,7 @@ func (tw *syncWorker) Exec(p Payload) (Payload, error) {
rsp, err := tw.execPayload(p)
if err != nil {
- if _, ok := err.(JobError); !ok {
+ if _, ok := err.(ExecError); !ok {
tw.w.State().Set(StateErrored)
tw.w.State().RegisterExec()
}
@@ -129,14 +129,14 @@ func (tw *syncWorker) ExecWithContext(ctx context.Context, p Payload) (Payload,
}
}
-func (tw *syncWorker) execPayload(rqs Payload) (Payload, error) {
+func (tw *syncWorker) execPayload(p Payload) (Payload, error) {
const op = errors.Op("exec_payload")
// two things; todo: merge
- if err := sendControl(tw.w.Relay(), rqs.Context); err != nil {
+ if err := sendControl(tw.w.Relay(), p.Context); err != nil {
return EmptyPayload, errors.E(op, err, "header error")
}
- if err := tw.w.Relay().Send(rqs.Body, 0); err != nil {
+ if err := tw.w.Relay().Send(p.Body, 0); err != nil {
return EmptyPayload, errors.E(op, err, "sender error")
}
@@ -153,7 +153,7 @@ func (tw *syncWorker) execPayload(rqs Payload) (Payload, error) {
}
if pr.HasFlag(goridge.PayloadError) {
- return EmptyPayload, JobError(rsp.Context)
+ return EmptyPayload, ExecError(rsp.Context)
}
// add streaming support :)
diff --git a/sync_worker_test.go b/sync_worker_test.go
index 7f969283..1bc2deb1 100755
--- a/sync_worker_test.go
+++ b/sync_worker_test.go
@@ -206,7 +206,7 @@ func Test_Error(t *testing.T) {
assert.Nil(t, res.Body)
assert.Nil(t, res.Context)
- assert.IsType(t, JobError{}, err)
+ assert.IsType(t, ExecError{}, err)
assert.Equal(t, "hello", err.Error())
}
diff --git a/worker_watcher.go b/worker_watcher.go
index 0eb8152b..f49eeacf 100755
--- a/worker_watcher.go
+++ b/worker_watcher.go
@@ -19,7 +19,7 @@ type Stack struct {
func NewWorkersStack() *Stack {
return &Stack{
- workers: make([]WorkerBase, 0, 12),
+ workers: make([]WorkerBase, 0),
}
}