771 lines
20 KiB
Go
771 lines
20 KiB
Go
/*
|
|
Copyright The containerd Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package v2
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"strings"
|
|
"time"
|
|
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/connectivity"
|
|
"google.golang.org/grpc/credentials/insecure"
|
|
|
|
eventstypes "github.com/containerd/containerd/api/events"
|
|
task "github.com/containerd/containerd/api/runtime/task/v3"
|
|
"github.com/containerd/containerd/api/types"
|
|
"github.com/containerd/containerd/v2/core/events/exchange"
|
|
"github.com/containerd/containerd/v2/core/runtime"
|
|
"github.com/containerd/containerd/v2/pkg/atomicfile"
|
|
"github.com/containerd/containerd/v2/pkg/dialer"
|
|
"github.com/containerd/containerd/v2/pkg/identifiers"
|
|
"github.com/containerd/containerd/v2/pkg/protobuf"
|
|
ptypes "github.com/containerd/containerd/v2/pkg/protobuf/types"
|
|
client "github.com/containerd/containerd/v2/pkg/shim"
|
|
"github.com/containerd/containerd/v2/pkg/timeout"
|
|
"github.com/containerd/errdefs"
|
|
"github.com/containerd/log"
|
|
"github.com/containerd/ttrpc"
|
|
"github.com/containerd/typeurl/v2"
|
|
)
|
|
|
|
const (
|
|
loadTimeout = "io.containerd.timeout.shim.load"
|
|
cleanupTimeout = "io.containerd.timeout.shim.cleanup"
|
|
shutdownTimeout = "io.containerd.timeout.shim.shutdown"
|
|
)
|
|
|
|
func init() {
|
|
timeout.Set(loadTimeout, 5*time.Second)
|
|
timeout.Set(cleanupTimeout, 5*time.Second)
|
|
timeout.Set(shutdownTimeout, 3*time.Second)
|
|
}
|
|
|
|
func loadShim(ctx context.Context, bundle *Bundle, onClose func()) (_ ShimInstance, retErr error) {
|
|
shimCtx, cancelShimLog := context.WithCancel(ctx)
|
|
defer func() {
|
|
if retErr != nil {
|
|
cancelShimLog()
|
|
}
|
|
}()
|
|
f, err := openShimLog(shimCtx, bundle, client.AnonReconnectDialer)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("open shim log pipe when reload: %w", err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil {
|
|
f.Close()
|
|
}
|
|
}()
|
|
// open the log pipe and block until the writer is ready
|
|
// this helps with synchronization of the shim
|
|
// copy the shim's logs to containerd's output
|
|
go func() {
|
|
defer f.Close()
|
|
_, err := io.Copy(os.Stderr, f)
|
|
// To prevent flood of error messages, the expected error
|
|
// should be reset, like os.ErrClosed or os.ErrNotExist, which
|
|
// depends on platform.
|
|
err = checkCopyShimLogError(ctx, err)
|
|
if err != nil {
|
|
log.G(ctx).WithError(err).Error("copy shim log after reload")
|
|
}
|
|
}()
|
|
onCloseWithShimLog := func() {
|
|
onClose()
|
|
cancelShimLog()
|
|
f.Close()
|
|
}
|
|
|
|
params, err := restoreBootstrapParams(bundle.Path)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to read boostrap.json when restoring bundle %q: %w", bundle.ID, err)
|
|
}
|
|
|
|
conn, err := makeConnection(ctx, bundle.ID, params, onCloseWithShimLog)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("unable to make connection: %w", err)
|
|
}
|
|
|
|
defer func() {
|
|
if retErr != nil {
|
|
conn.Close()
|
|
}
|
|
}()
|
|
|
|
// The address is in the form like ttrpc+unix://<uds-path> or grpc+vsock://<cid>:<port>
|
|
address := fmt.Sprintf("%s+%s", params.Protocol, params.Address)
|
|
|
|
shim := &shim{
|
|
bundle: bundle,
|
|
client: conn,
|
|
address: address,
|
|
version: params.Version,
|
|
}
|
|
|
|
return shim, nil
|
|
}
|
|
|
|
func cleanupAfterDeadShim(ctx context.Context, id string, rt *runtime.NSMap[ShimInstance], events *exchange.Exchange, binaryCall *binary) {
|
|
ctx, cancel := timeout.WithContext(ctx, cleanupTimeout)
|
|
defer cancel()
|
|
|
|
log.G(ctx).WithField("id", id).Warn("cleaning up after shim disconnected")
|
|
response, err := binaryCall.Delete(ctx)
|
|
if err != nil {
|
|
log.G(ctx).WithError(err).WithField("id", id).Warn("failed to clean up after shim disconnected")
|
|
}
|
|
|
|
if _, err := rt.Get(ctx, id); err != nil {
|
|
// Task was never started or was already successfully deleted
|
|
// No need to publish events
|
|
return
|
|
}
|
|
|
|
var (
|
|
pid uint32
|
|
exitStatus uint32
|
|
exitedAt time.Time
|
|
)
|
|
if response != nil {
|
|
pid = response.Pid
|
|
exitStatus = response.Status
|
|
exitedAt = response.Timestamp
|
|
} else {
|
|
exitStatus = 255
|
|
exitedAt = time.Now()
|
|
}
|
|
events.Publish(ctx, runtime.TaskExitEventTopic, &eventstypes.TaskExit{
|
|
ContainerID: id,
|
|
ID: id,
|
|
Pid: pid,
|
|
ExitStatus: exitStatus,
|
|
ExitedAt: protobuf.ToTimestamp(exitedAt),
|
|
})
|
|
|
|
events.Publish(ctx, runtime.TaskDeleteEventTopic, &eventstypes.TaskDelete{
|
|
ContainerID: id,
|
|
Pid: pid,
|
|
ExitStatus: exitStatus,
|
|
ExitedAt: protobuf.ToTimestamp(exitedAt),
|
|
})
|
|
}
|
|
|
|
// CurrentShimVersion is the latest shim version supported by containerd (e.g. TaskService v3).
|
|
const CurrentShimVersion = 3
|
|
|
|
// ShimInstance represents running shim process managed by ShimManager.
|
|
type ShimInstance interface {
|
|
io.Closer
|
|
|
|
// ID of the shim.
|
|
ID() string
|
|
// Namespace of this shim.
|
|
Namespace() string
|
|
// Bundle is a file system path to shim's bundle.
|
|
Bundle() string
|
|
// Client returns the underlying TTRPC or GRPC client object for this shim.
|
|
// The underlying object can be either *ttrpc.Client or grpc.ClientConnInterface.
|
|
Client() any
|
|
// Delete will close the client and remove bundle from disk.
|
|
Delete(ctx context.Context) error
|
|
// Endpoint returns shim's endpoint information,
|
|
// including address and version.
|
|
Endpoint() (string, int)
|
|
}
|
|
|
|
func parseStartResponse(response []byte) (client.BootstrapParams, error) {
|
|
var params client.BootstrapParams
|
|
|
|
if err := json.Unmarshal(response, ¶ms); err != nil || params.Version < 2 {
|
|
// Use TTRPC for legacy shims
|
|
params.Address = string(response)
|
|
params.Protocol = "ttrpc"
|
|
params.Version = 2
|
|
}
|
|
|
|
if params.Version > CurrentShimVersion {
|
|
return client.BootstrapParams{}, fmt.Errorf("unsupported shim version (%d): %w", params.Version, errdefs.ErrNotImplemented)
|
|
}
|
|
|
|
return params, nil
|
|
}
|
|
|
|
// writeBootstrapParams writes shim's bootstrap configuration (e.g. how to connect, version, etc).
|
|
func writeBootstrapParams(path string, params client.BootstrapParams) error {
|
|
path, err := filepath.Abs(path)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
data, err := json.Marshal(¶ms)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
f, err := atomicfile.New(path, 0o666)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
_, err = f.Write(data)
|
|
if err != nil {
|
|
f.Cancel()
|
|
return err
|
|
}
|
|
|
|
return f.Close()
|
|
}
|
|
|
|
func readBootstrapParams(path string) (client.BootstrapParams, error) {
|
|
path, err := filepath.Abs(path)
|
|
if err != nil {
|
|
return client.BootstrapParams{}, err
|
|
}
|
|
|
|
data, err := os.ReadFile(path)
|
|
if err != nil {
|
|
return client.BootstrapParams{}, err
|
|
}
|
|
|
|
return parseStartResponse(data)
|
|
}
|
|
|
|
// makeConnection creates a new TTRPC or GRPC connection object from address.
|
|
// address can be either a socket path for TTRPC or JSON serialized BootstrapParams.
|
|
func makeConnection(ctx context.Context, id string, params client.BootstrapParams, onClose func()) (_ io.Closer, retErr error) {
|
|
log.G(ctx).WithFields(log.Fields{
|
|
"address": params.Address,
|
|
"protocol": params.Protocol,
|
|
"version": params.Version,
|
|
}).Infof("connecting to shim %s", id)
|
|
|
|
switch strings.ToLower(params.Protocol) {
|
|
case "ttrpc":
|
|
conn, err := client.Connect(params.Address, client.AnonReconnectDialer)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to create TTRPC connection: %w", err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil {
|
|
conn.Close()
|
|
}
|
|
}()
|
|
|
|
return ttrpc.NewClient(conn, ttrpc.WithOnClose(onClose)), nil
|
|
case "grpc":
|
|
ctx, cancel := context.WithTimeout(ctx, time.Second*100)
|
|
defer cancel()
|
|
|
|
gopts := []grpc.DialOption{
|
|
grpc.WithTransportCredentials(insecure.NewCredentials()),
|
|
grpc.WithBlock(),
|
|
}
|
|
return grpcDialContext(ctx, params.Address, onClose, gopts...)
|
|
default:
|
|
return nil, fmt.Errorf("unexpected protocol: %q", params.Protocol)
|
|
}
|
|
}
|
|
|
|
// grpcDialContext and the underlying grpcConn type exist solely
|
|
// so we can have something similar to ttrpc.WithOnClose to have
|
|
// a callback run when the connection is severed or explicitly closed.
|
|
func grpcDialContext(
|
|
ctx context.Context,
|
|
address string,
|
|
onClose func(),
|
|
gopts ...grpc.DialOption,
|
|
) (*grpcConn, error) {
|
|
// If grpc.WithBlock is specified in gopts this causes the connection to block waiting for
|
|
// a connection regardless of if the socket exists or has a listener when Dial begins. This
|
|
// specific behavior of WithBlock is mostly undesirable for shims, as if the socket isn't
|
|
// there when we go to load/connect there's likely an issue. However, getting rid of WithBlock is
|
|
// also undesirable as we don't want the background connection behavior, we want to ensure
|
|
// a connection before moving on. To bring this in line with the ttrpc connection behavior
|
|
// lets do an initial dial to ensure the shims socket is actually available. stat wouldn't suffice
|
|
// here as if the shim exited unexpectedly its socket may still be on the filesystem, but it'd return
|
|
// ECONNREFUSED which grpc.DialContext will happily trudge along through for the full timeout.
|
|
//
|
|
// This is especially helpful on restart of containerd as if the shim died while containerd
|
|
// was down, we end up waiting the full timeout.
|
|
conn, err := net.DialTimeout("unix", address, time.Second*10)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
conn.Close()
|
|
|
|
target := dialer.DialAddress(address)
|
|
client, err := grpc.DialContext(ctx, target, gopts...)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to create GRPC connection: %w", err)
|
|
}
|
|
|
|
done := make(chan struct{})
|
|
go func() {
|
|
gctx := context.Background()
|
|
sourceState := connectivity.Ready
|
|
for {
|
|
if client.WaitForStateChange(gctx, sourceState) {
|
|
state := client.GetState()
|
|
if state == connectivity.Idle || state == connectivity.Shutdown {
|
|
break
|
|
}
|
|
// Could be transient failure. Lets see if we can get back to a working
|
|
// state.
|
|
log.G(gctx).WithFields(log.Fields{
|
|
"state": state,
|
|
"addr": target,
|
|
}).Warn("shim grpc connection unexpected state")
|
|
sourceState = state
|
|
}
|
|
}
|
|
onClose()
|
|
close(done)
|
|
}()
|
|
|
|
return &grpcConn{
|
|
ClientConn: client,
|
|
onCloseDone: done,
|
|
}, nil
|
|
}
|
|
|
|
type grpcConn struct {
|
|
*grpc.ClientConn
|
|
onCloseDone chan struct{}
|
|
}
|
|
|
|
func (gc *grpcConn) UserOnCloseWait(ctx context.Context) error {
|
|
select {
|
|
case <-gc.onCloseDone:
|
|
return nil
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
|
|
type shim struct {
|
|
bundle *Bundle
|
|
client any
|
|
address string
|
|
version int
|
|
}
|
|
|
|
var _ ShimInstance = (*shim)(nil)
|
|
|
|
// ID of the shim/task
|
|
func (s *shim) ID() string {
|
|
return s.bundle.ID
|
|
}
|
|
|
|
func (s *shim) Endpoint() (string, int) {
|
|
return s.address, s.version
|
|
}
|
|
|
|
func (s *shim) Namespace() string {
|
|
return s.bundle.Namespace
|
|
}
|
|
|
|
func (s *shim) Bundle() string {
|
|
return s.bundle.Path
|
|
}
|
|
|
|
func (s *shim) Client() any {
|
|
return s.client
|
|
}
|
|
|
|
// Close closes the underlying client connection.
|
|
func (s *shim) Close() error {
|
|
if ttrpcClient, ok := s.client.(*ttrpc.Client); ok {
|
|
return ttrpcClient.Close()
|
|
}
|
|
|
|
if grpcClient, ok := s.client.(*grpcConn); ok {
|
|
return grpcClient.Close()
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *shim) Delete(ctx context.Context) error {
|
|
var result []error
|
|
|
|
if ttrpcClient, ok := s.client.(*ttrpc.Client); ok {
|
|
if err := ttrpcClient.Close(); err != nil {
|
|
result = append(result, fmt.Errorf("failed to close ttrpc client: %w", err))
|
|
}
|
|
|
|
if err := ttrpcClient.UserOnCloseWait(ctx); err != nil {
|
|
result = append(result, fmt.Errorf("close wait error: %w", err))
|
|
}
|
|
}
|
|
|
|
if grpcClient, ok := s.client.(*grpcConn); ok {
|
|
if err := grpcClient.Close(); err != nil {
|
|
result = append(result, fmt.Errorf("failed to close grpc client: %w", err))
|
|
}
|
|
|
|
if err := grpcClient.UserOnCloseWait(ctx); err != nil {
|
|
result = append(result, fmt.Errorf("close wait error: %w", err))
|
|
}
|
|
}
|
|
|
|
if err := s.bundle.Delete(); err != nil {
|
|
log.G(ctx).WithField("id", s.ID()).WithError(err).Error("failed to delete bundle")
|
|
result = append(result, fmt.Errorf("failed to delete bundle: %w", err))
|
|
}
|
|
|
|
return errors.Join(result...)
|
|
}
|
|
|
|
var _ runtime.Task = &shimTask{}
|
|
|
|
// shimTask wraps shim process and adds task service client for compatibility with existing shim manager.
|
|
type shimTask struct {
|
|
ShimInstance
|
|
task TaskServiceClient
|
|
}
|
|
|
|
func newShimTask(shim ShimInstance) (*shimTask, error) {
|
|
_, version := shim.Endpoint()
|
|
taskClient, err := NewTaskClient(shim.Client(), version)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &shimTask{
|
|
ShimInstance: shim,
|
|
task: taskClient,
|
|
}, nil
|
|
}
|
|
|
|
func (s *shimTask) Shutdown(ctx context.Context) error {
|
|
_, err := s.task.Shutdown(ctx, &task.ShutdownRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil && !errors.Is(err, ttrpc.ErrClosed) {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) waitShutdown(ctx context.Context) error {
|
|
ctx, cancel := timeout.WithContext(ctx, shutdownTimeout)
|
|
defer cancel()
|
|
return s.Shutdown(ctx)
|
|
}
|
|
|
|
// PID of the task
|
|
func (s *shimTask) PID(ctx context.Context) (uint32, error) {
|
|
response, err := s.task.Connect(ctx, &task.ConnectRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
return 0, errdefs.FromGRPC(err)
|
|
}
|
|
|
|
return response.TaskPid, nil
|
|
}
|
|
|
|
func (s *shimTask) delete(ctx context.Context, sandboxed bool, removeTask func(ctx context.Context, id string)) (*runtime.Exit, error) {
|
|
response, shimErr := s.task.Delete(ctx, &task.DeleteRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if shimErr != nil {
|
|
log.G(ctx).WithField("id", s.ID()).WithError(shimErr).Debug("failed to delete task")
|
|
if !errors.Is(shimErr, ttrpc.ErrClosed) {
|
|
shimErr = errdefs.FromGRPC(shimErr)
|
|
if !errdefs.IsNotFound(shimErr) {
|
|
return nil, shimErr
|
|
}
|
|
}
|
|
}
|
|
|
|
// NOTE: If the shim has been killed and ttrpc connection has been
|
|
// closed, the shimErr will not be nil. For this case, the event
|
|
// subscriber, like moby/moby, might have received the exit or delete
|
|
// events. Just in case, we should allow ttrpc-callback-on-close to
|
|
// send the exit and delete events again. And the exit status will
|
|
// depend on result of shimV2.Delete.
|
|
//
|
|
// If not, the shim has been delivered the exit and delete events.
|
|
// So we should remove the record and prevent duplicate events from
|
|
// ttrpc-callback-on-close.
|
|
//
|
|
// TODO: It's hard to guarantee that the event is unique and sent only
|
|
// once. The moby/moby should not rely on that assumption that there is
|
|
// only one exit event. The moby/moby should handle the duplicate events.
|
|
//
|
|
// REF: https://github.com/containerd/containerd/issues/4769
|
|
if shimErr == nil {
|
|
removeTask(ctx, s.ID())
|
|
}
|
|
|
|
// Don't shutdown sandbox as there may be other containers running.
|
|
// Let controller decide when to shutdown.
|
|
if !sandboxed {
|
|
if err := s.waitShutdown(ctx); err != nil {
|
|
// FIXME(fuweid):
|
|
//
|
|
// If the error is context canceled, should we use context.TODO()
|
|
// to wait for it?
|
|
log.G(ctx).WithField("id", s.ID()).WithError(err).Error("failed to shutdown shim task and the shim might be leaked")
|
|
}
|
|
}
|
|
|
|
if err := s.ShimInstance.Delete(ctx); err != nil {
|
|
log.G(ctx).WithField("id", s.ID()).WithError(err).Error("failed to delete shim")
|
|
}
|
|
|
|
// remove self from the runtime task list
|
|
// this seems dirty but it cleans up the API across runtimes, tasks, and the service
|
|
removeTask(ctx, s.ID())
|
|
|
|
if shimErr != nil {
|
|
return nil, shimErr
|
|
}
|
|
|
|
return &runtime.Exit{
|
|
Status: response.ExitStatus,
|
|
Timestamp: protobuf.FromTimestamp(response.ExitedAt),
|
|
Pid: response.Pid,
|
|
}, nil
|
|
}
|
|
|
|
func (s *shimTask) Create(ctx context.Context, opts runtime.CreateOpts) (runtime.Task, error) {
|
|
topts := opts.TaskOptions
|
|
if topts == nil || topts.GetValue() == nil {
|
|
topts = opts.RuntimeOptions
|
|
}
|
|
request := &task.CreateTaskRequest{
|
|
ID: s.ID(),
|
|
Bundle: s.Bundle(),
|
|
Stdin: opts.IO.Stdin,
|
|
Stdout: opts.IO.Stdout,
|
|
Stderr: opts.IO.Stderr,
|
|
Terminal: opts.IO.Terminal,
|
|
Checkpoint: opts.Checkpoint,
|
|
Options: typeurl.MarshalProto(topts),
|
|
}
|
|
for _, m := range opts.Rootfs {
|
|
request.Rootfs = append(request.Rootfs, &types.Mount{
|
|
Type: m.Type,
|
|
Source: m.Source,
|
|
Target: m.Target,
|
|
Options: m.Options,
|
|
})
|
|
}
|
|
|
|
_, err := s.task.Create(ctx, request)
|
|
if err != nil {
|
|
return nil, errdefs.FromGRPC(err)
|
|
}
|
|
|
|
return s, nil
|
|
}
|
|
|
|
func (s *shimTask) Pause(ctx context.Context) error {
|
|
if _, err := s.task.Pause(ctx, &task.PauseRequest{
|
|
ID: s.ID(),
|
|
}); err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Resume(ctx context.Context) error {
|
|
if _, err := s.task.Resume(ctx, &task.ResumeRequest{
|
|
ID: s.ID(),
|
|
}); err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Start(ctx context.Context) error {
|
|
_, err := s.task.Start(ctx, &task.StartRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Kill(ctx context.Context, signal uint32, all bool) error {
|
|
if _, err := s.task.Kill(ctx, &task.KillRequest{
|
|
ID: s.ID(),
|
|
Signal: signal,
|
|
All: all,
|
|
}); err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Exec(ctx context.Context, id string, opts runtime.ExecOpts) (runtime.ExecProcess, error) {
|
|
if err := identifiers.Validate(id); err != nil {
|
|
return nil, fmt.Errorf("invalid exec id %s: %w", id, err)
|
|
}
|
|
request := &task.ExecProcessRequest{
|
|
ID: s.ID(),
|
|
ExecID: id,
|
|
Stdin: opts.IO.Stdin,
|
|
Stdout: opts.IO.Stdout,
|
|
Stderr: opts.IO.Stderr,
|
|
Terminal: opts.IO.Terminal,
|
|
Spec: opts.Spec,
|
|
}
|
|
if _, err := s.task.Exec(ctx, request); err != nil {
|
|
return nil, errdefs.FromGRPC(err)
|
|
}
|
|
return &process{
|
|
id: id,
|
|
shim: s,
|
|
}, nil
|
|
}
|
|
|
|
func (s *shimTask) Pids(ctx context.Context) ([]runtime.ProcessInfo, error) {
|
|
resp, err := s.task.Pids(ctx, &task.PidsRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
return nil, errdefs.FromGRPC(err)
|
|
}
|
|
var processList []runtime.ProcessInfo
|
|
for _, p := range resp.Processes {
|
|
processList = append(processList, runtime.ProcessInfo{
|
|
Pid: p.Pid,
|
|
Info: p.Info,
|
|
})
|
|
}
|
|
return processList, nil
|
|
}
|
|
|
|
func (s *shimTask) ResizePty(ctx context.Context, size runtime.ConsoleSize) error {
|
|
_, err := s.task.ResizePty(ctx, &task.ResizePtyRequest{
|
|
ID: s.ID(),
|
|
Width: size.Width,
|
|
Height: size.Height,
|
|
})
|
|
if err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) CloseIO(ctx context.Context) error {
|
|
_, err := s.task.CloseIO(ctx, &task.CloseIORequest{
|
|
ID: s.ID(),
|
|
Stdin: true,
|
|
})
|
|
if err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Wait(ctx context.Context) (*runtime.Exit, error) {
|
|
taskPid, err := s.PID(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
response, err := s.task.Wait(ctx, &task.WaitRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
return nil, errdefs.FromGRPC(err)
|
|
}
|
|
return &runtime.Exit{
|
|
Pid: taskPid,
|
|
Timestamp: protobuf.FromTimestamp(response.ExitedAt),
|
|
Status: response.ExitStatus,
|
|
}, nil
|
|
}
|
|
|
|
func (s *shimTask) Checkpoint(ctx context.Context, path string, options *ptypes.Any) error {
|
|
request := &task.CheckpointTaskRequest{
|
|
ID: s.ID(),
|
|
Path: path,
|
|
Options: options,
|
|
}
|
|
if _, err := s.task.Checkpoint(ctx, request); err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Update(ctx context.Context, resources *ptypes.Any, annotations map[string]string) error {
|
|
if _, err := s.task.Update(ctx, &task.UpdateTaskRequest{
|
|
ID: s.ID(),
|
|
Resources: resources,
|
|
Annotations: annotations,
|
|
}); err != nil {
|
|
return errdefs.FromGRPC(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *shimTask) Stats(ctx context.Context) (*ptypes.Any, error) {
|
|
response, err := s.task.Stats(ctx, &task.StatsRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
return nil, errdefs.FromGRPC(err)
|
|
}
|
|
return response.Stats, nil
|
|
}
|
|
|
|
func (s *shimTask) Process(ctx context.Context, id string) (runtime.ExecProcess, error) {
|
|
p := &process{
|
|
id: id,
|
|
shim: s,
|
|
}
|
|
if _, err := p.State(ctx); err != nil {
|
|
return nil, err
|
|
}
|
|
return p, nil
|
|
}
|
|
|
|
func (s *shimTask) State(ctx context.Context) (runtime.State, error) {
|
|
response, err := s.task.State(ctx, &task.StateRequest{
|
|
ID: s.ID(),
|
|
})
|
|
if err != nil {
|
|
if !errors.Is(err, ttrpc.ErrClosed) {
|
|
return runtime.State{}, errdefs.FromGRPC(err)
|
|
}
|
|
return runtime.State{}, errdefs.ErrNotFound
|
|
}
|
|
return runtime.State{
|
|
Pid: response.Pid,
|
|
Status: statusFromProto(response.Status),
|
|
Stdin: response.Stdin,
|
|
Stdout: response.Stdout,
|
|
Stderr: response.Stderr,
|
|
Terminal: response.Terminal,
|
|
ExitStatus: response.ExitStatus,
|
|
ExitedAt: protobuf.FromTimestamp(response.ExitedAt),
|
|
}, nil
|
|
}
|