360 lines
8.8 KiB
Go
360 lines
8.8 KiB
Go
package containerd
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"io"
|
|
"runtime"
|
|
"strings"
|
|
"syscall"
|
|
|
|
"github.com/containerd/containerd/api/services/containers/v1"
|
|
eventsapi "github.com/containerd/containerd/api/services/events/v1"
|
|
"github.com/containerd/containerd/api/services/tasks/v1"
|
|
tasktypes "github.com/containerd/containerd/api/types/task"
|
|
"github.com/containerd/containerd/content"
|
|
"github.com/containerd/containerd/events"
|
|
"github.com/containerd/containerd/rootfs"
|
|
"github.com/opencontainers/image-spec/specs-go/v1"
|
|
specs "github.com/opencontainers/runtime-spec/specs-go"
|
|
)
|
|
|
|
const UnknownExitStatus = 255
|
|
|
|
type TaskStatus string
|
|
|
|
const (
|
|
Running TaskStatus = "running"
|
|
Created TaskStatus = "created"
|
|
Stopped TaskStatus = "stopped"
|
|
Paused TaskStatus = "paused"
|
|
Pausing TaskStatus = "pausing"
|
|
)
|
|
|
|
type IOCloserOpts func(*tasks.CloseIORequest)
|
|
|
|
func WithStdinCloser(r *tasks.CloseIORequest) {
|
|
r.Stdin = true
|
|
}
|
|
|
|
type CheckpointOpts func(*tasks.CheckpointTaskRequest) error
|
|
|
|
type Task interface {
|
|
Pid() uint32
|
|
Delete(context.Context) (uint32, error)
|
|
Kill(context.Context, syscall.Signal) error
|
|
Pause(context.Context) error
|
|
Resume(context.Context) error
|
|
Start(context.Context) error
|
|
Status(context.Context) (TaskStatus, error)
|
|
Wait(context.Context) (uint32, error)
|
|
Exec(context.Context, *specs.Process, IOCreation) (Process, error)
|
|
Processes(context.Context) ([]uint32, error)
|
|
CloseIO(context.Context, ...IOCloserOpts) error
|
|
Resize(ctx context.Context, w, h uint32) error
|
|
IO() *IO
|
|
Checkpoint(context.Context, ...CheckpointOpts) (v1.Descriptor, error)
|
|
}
|
|
|
|
type Process interface {
|
|
Pid() uint32
|
|
Start(context.Context) error
|
|
Delete(context.Context) (uint32, error)
|
|
Kill(context.Context, syscall.Signal) error
|
|
Wait(context.Context) (uint32, error)
|
|
CloseIO(context.Context, ...IOCloserOpts) error
|
|
Resize(ctx context.Context, w, h uint32) error
|
|
IO() *IO
|
|
}
|
|
|
|
var _ = (Task)(&task{})
|
|
|
|
type task struct {
|
|
client *Client
|
|
|
|
io *IO
|
|
containerID string
|
|
pid uint32
|
|
|
|
deferred *tasks.CreateTaskRequest
|
|
pidSync chan struct{}
|
|
}
|
|
|
|
// Pid returns the pid or process id for the task
|
|
func (t *task) Pid() uint32 {
|
|
return t.pid
|
|
}
|
|
|
|
func (t *task) Start(ctx context.Context) error {
|
|
if t.deferred != nil {
|
|
response, err := t.client.TaskService().Create(ctx, t.deferred)
|
|
t.deferred = nil
|
|
if err != nil {
|
|
return err
|
|
}
|
|
t.pid = response.Pid
|
|
close(t.pidSync)
|
|
return nil
|
|
}
|
|
_, err := t.client.TaskService().Start(ctx, &tasks.StartTaskRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
return err
|
|
}
|
|
|
|
func (t *task) Kill(ctx context.Context, s syscall.Signal) error {
|
|
_, err := t.client.TaskService().Kill(ctx, &tasks.KillRequest{
|
|
Signal: uint32(s),
|
|
ContainerID: t.containerID,
|
|
PidOrAll: &tasks.KillRequest_All{
|
|
All: true,
|
|
},
|
|
})
|
|
return err
|
|
}
|
|
|
|
func (t *task) Pause(ctx context.Context) error {
|
|
_, err := t.client.TaskService().Pause(ctx, &tasks.PauseTaskRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
return err
|
|
}
|
|
|
|
func (t *task) Resume(ctx context.Context) error {
|
|
_, err := t.client.TaskService().Resume(ctx, &tasks.ResumeTaskRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
return err
|
|
}
|
|
|
|
func (t *task) Status(ctx context.Context) (TaskStatus, error) {
|
|
r, err := t.client.TaskService().Get(ctx, &tasks.GetTaskRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
return TaskStatus(strings.ToLower(r.Task.Status.String())), nil
|
|
}
|
|
|
|
// Wait is a blocking call that will wait for the task to exit and return the exit status
|
|
func (t *task) Wait(ctx context.Context) (uint32, error) {
|
|
// TODO (ehazlett): add filtering for specific event
|
|
eventstream, err := t.client.EventService().Stream(ctx, &eventsapi.StreamEventsRequest{})
|
|
if err != nil {
|
|
return UnknownExitStatus, err
|
|
}
|
|
<-t.pidSync
|
|
|
|
var e eventsapi.RuntimeEvent
|
|
|
|
for {
|
|
evt, err := eventstream.Recv()
|
|
if err != nil {
|
|
return UnknownExitStatus, err
|
|
}
|
|
|
|
if !events.Is(evt.Event, &eventsapi.RuntimeEvent{}) {
|
|
continue
|
|
}
|
|
|
|
if err := events.UnmarshalEvent(evt.Event, &e); err != nil {
|
|
return UnknownExitStatus, err
|
|
}
|
|
|
|
if e.Type != tasktypes.Event_EXIT {
|
|
continue
|
|
}
|
|
|
|
if e.ID == t.containerID && e.Pid == t.pid {
|
|
return e.ExitStatus, nil
|
|
}
|
|
}
|
|
}
|
|
|
|
// Delete deletes the task and its runtime state
|
|
// it returns the exit status of the task and any errors that were encountered
|
|
// during cleanup
|
|
func (t *task) Delete(ctx context.Context) (uint32, error) {
|
|
var cerr error
|
|
if t.io != nil {
|
|
cerr = t.io.Close()
|
|
}
|
|
r, err := t.client.TaskService().Delete(ctx, &tasks.DeleteTaskRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
if err != nil {
|
|
return UnknownExitStatus, err
|
|
}
|
|
return r.ExitStatus, cerr
|
|
}
|
|
|
|
func (t *task) Exec(ctx context.Context, spec *specs.Process, ioCreate IOCreation) (Process, error) {
|
|
i, err := ioCreate()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &process{
|
|
task: t,
|
|
io: i,
|
|
spec: spec,
|
|
pidSync: make(chan struct{}),
|
|
}, nil
|
|
}
|
|
|
|
func (t *task) Processes(ctx context.Context) ([]uint32, error) {
|
|
response, err := t.client.TaskService().ListProcesses(ctx, &tasks.ListProcessesRequest{
|
|
ContainerID: t.containerID,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
var out []uint32
|
|
for _, p := range response.Processes {
|
|
out = append(out, p.Pid)
|
|
}
|
|
return out, nil
|
|
}
|
|
|
|
func (t *task) CloseIO(ctx context.Context, opts ...IOCloserOpts) error {
|
|
r := &tasks.CloseIORequest{
|
|
ContainerID: t.containerID,
|
|
Pid: t.pid,
|
|
}
|
|
for _, o := range opts {
|
|
o(r)
|
|
}
|
|
_, err := t.client.TaskService().CloseIO(ctx, r)
|
|
return err
|
|
}
|
|
|
|
func (t *task) IO() *IO {
|
|
return t.io
|
|
}
|
|
|
|
func (t *task) Resize(ctx context.Context, w, h uint32) error {
|
|
_, err := t.client.TaskService().ResizePty(ctx, &tasks.ResizePtyRequest{
|
|
ContainerID: t.containerID,
|
|
Width: w,
|
|
Height: h,
|
|
Pid: t.pid,
|
|
})
|
|
return err
|
|
}
|
|
|
|
func WithExit(r *tasks.CheckpointTaskRequest) error {
|
|
r.Options["exit"] = "true"
|
|
return nil
|
|
}
|
|
|
|
func (t *task) Checkpoint(ctx context.Context, opts ...CheckpointOpts) (d v1.Descriptor, err error) {
|
|
request := &tasks.CheckpointTaskRequest{
|
|
ContainerID: t.containerID,
|
|
Options: make(map[string]string),
|
|
}
|
|
for _, o := range opts {
|
|
if err := o(request); err != nil {
|
|
return d, err
|
|
}
|
|
}
|
|
// make sure we pause it and resume after all other filesystem operations are completed
|
|
if err := t.Pause(ctx); err != nil {
|
|
return d, err
|
|
}
|
|
defer t.Resume(ctx)
|
|
cr, err := t.client.ContainerService().Get(ctx, &containers.GetContainerRequest{
|
|
ID: t.containerID,
|
|
})
|
|
if err != nil {
|
|
return d, err
|
|
}
|
|
var index v1.Index
|
|
if err := t.checkpointTask(ctx, &index, request); err != nil {
|
|
return d, err
|
|
}
|
|
if err := t.checkpointImage(ctx, &index, cr.Container.Image); err != nil {
|
|
return d, err
|
|
}
|
|
if err := t.checkpointRWSnapshot(ctx, &index, cr.Container.RootFS); err != nil {
|
|
return d, err
|
|
}
|
|
index.Annotations = make(map[string]string)
|
|
index.Annotations["image.name"] = cr.Container.Image
|
|
return t.writeIndex(ctx, &index)
|
|
}
|
|
|
|
func (t *task) checkpointTask(ctx context.Context, index *v1.Index, request *tasks.CheckpointTaskRequest) error {
|
|
response, err := t.client.TaskService().Checkpoint(ctx, request)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
// add the checkpoint descriptors to the index
|
|
for _, d := range response.Descriptors {
|
|
index.Manifests = append(index.Manifests, v1.Descriptor{
|
|
MediaType: d.MediaType,
|
|
Size: d.Size_,
|
|
Digest: d.Digest,
|
|
Platform: &v1.Platform{
|
|
OS: runtime.GOOS,
|
|
Architecture: runtime.GOARCH,
|
|
},
|
|
})
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (t *task) checkpointRWSnapshot(ctx context.Context, index *v1.Index, id string) error {
|
|
rw, err := rootfs.Diff(ctx, id, fmt.Sprintf("checkpoint-rw-%s", id), t.client.SnapshotService(), t.client.DiffService())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rw.Platform = &v1.Platform{
|
|
OS: runtime.GOOS,
|
|
Architecture: runtime.GOARCH,
|
|
}
|
|
index.Manifests = append(index.Manifests, rw)
|
|
return nil
|
|
}
|
|
|
|
func (t *task) checkpointImage(ctx context.Context, index *v1.Index, image string) error {
|
|
if image == "" {
|
|
return fmt.Errorf("cannot checkpoint image with empty name")
|
|
}
|
|
ir, err := t.client.ImageService().Get(ctx, image)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
index.Manifests = append(index.Manifests, ir.Target)
|
|
return nil
|
|
}
|
|
|
|
func (t *task) writeIndex(ctx context.Context, index *v1.Index) (v1.Descriptor, error) {
|
|
buf := bytes.NewBuffer(nil)
|
|
if err := json.NewEncoder(buf).Encode(index); err != nil {
|
|
return v1.Descriptor{}, err
|
|
}
|
|
return writeContent(ctx, t.client.ContentStore(), v1.MediaTypeImageIndex, t.containerID, buf)
|
|
}
|
|
|
|
func writeContent(ctx context.Context, store content.Store, mediaType, ref string, r io.Reader) (d v1.Descriptor, err error) {
|
|
writer, err := store.Writer(ctx, ref, 0, "")
|
|
if err != nil {
|
|
return d, err
|
|
}
|
|
defer writer.Close()
|
|
size, err := io.Copy(writer, r)
|
|
if err != nil {
|
|
return d, err
|
|
}
|
|
if err := writer.Commit(0, ""); err != nil {
|
|
return d, err
|
|
}
|
|
return v1.Descriptor{
|
|
MediaType: mediaType,
|
|
Digest: writer.Digest(),
|
|
Size: size,
|
|
}, nil
|
|
}
|