containerd/lease_test.go
Wei Fu 8a388d6238 leases: support resource management
Add three methods for lease service so that the client can use it to
manage the resource by lease, not just gc.root label. With the following
methods, it is easy for client to maintain their own cache system.

```
 - AddResource(context.Context, Lease, Resource) error
 - RemoveResource(context.Context, Lease, Resource) error
 - ListResources(context.Context, Lease) ([]Resource, error)
```

And the resource is to be

```golang
type Resource {
  ID   string
  Type string
}
```

For the snapshots, the Type field will be formatted by
snapshots/%{type}, like snapshots/overlayfs.

fix: #3295

Signed-off-by: Wei Fu <fuweid89@gmail.com>
2019-05-29 11:00:32 +08:00

142 lines
3.4 KiB
Go

/*
Copyright The containerd Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package containerd
import (
"runtime"
"testing"
"github.com/containerd/containerd/errdefs"
"github.com/containerd/containerd/images"
"github.com/containerd/containerd/leases"
"github.com/opencontainers/image-spec/identity"
"github.com/pkg/errors"
)
func TestLeaseResources(t *testing.T) {
if runtime.GOOS == "windows" {
t.Skip()
}
ctx, cancel := testContext()
defer cancel()
client, err := newClient(t, address)
if err != nil {
t.Fatal(err)
}
defer client.Close()
var (
ls = client.LeasesService()
cs = client.ContentStore()
imgSrv = client.ImageService()
sn = client.SnapshotService("native")
)
l, err := ls.Create(ctx, leases.WithRandomID())
if err != nil {
t.Fatal(err)
}
defer ls.Delete(ctx, l, leases.SynchronousDelete)
// step 1: download image
imageName := "docker.io/library/busybox:1.25"
image, err := client.Pull(ctx, imageName, WithPullUnpack, WithPullSnapshotter("native"))
if err != nil {
t.Fatal(err)
}
defer imgSrv.Delete(ctx, imageName)
// both the config and snapshotter should exist
cfgDesc, err := image.Config(ctx)
if err != nil {
t.Fatal(err)
}
if _, err := cs.Info(ctx, cfgDesc.Digest); err != nil {
t.Fatal(err)
}
dgsts, err := image.RootFS(ctx)
if err != nil {
t.Fatal(err)
}
chainID := identity.ChainID(dgsts)
if _, err := sn.Stat(ctx, chainID.String()); err != nil {
t.Fatal(err)
}
// step 2: reference snapshotter with lease
r := leases.Resource{
ID: chainID.String(),
Type: "snapshots/native",
}
if err := ls.AddResource(ctx, l, r); err != nil {
t.Fatal(err)
}
list, err := ls.ListResources(ctx, l)
if err != nil {
t.Fatal(err)
}
if len(list) != 1 || list[0] != r {
t.Fatalf("expected (%v), but got (%v)", []leases.Resource{r}, list)
}
// step 3: remove image and check the status of snapshotter and content
if err := imgSrv.Delete(ctx, imageName, images.SynchronousDelete()); err != nil {
t.Fatal(err)
}
// config should be removed but the snapshotter should exist
if _, err := cs.Info(ctx, cfgDesc.Digest); errors.Cause(err) != errdefs.ErrNotFound {
t.Fatalf("expected error(%v), but got(%v)", errdefs.ErrNotFound, err)
}
if _, err := sn.Stat(ctx, chainID.String()); err != nil {
t.Fatal(err)
}
// step 4: remove resource from the lease and check the list API
if err := ls.DeleteResource(ctx, l, r); err != nil {
t.Fatal(err)
}
list, err = ls.ListResources(ctx, l)
if err != nil {
t.Fatal(err)
}
if len(list) != 0 {
t.Fatalf("expected nothing, but got (%v)", list)
}
// step 5: remove the lease to check the status of snapshotter
if err := ls.Delete(ctx, l, leases.SynchronousDelete); err != nil {
t.Fatal(err)
}
if _, err := sn.Stat(ctx, chainID.String()); errors.Cause(err) != errdefs.ErrNotFound {
t.Fatalf("expected error(%v), but got(%v)", errdefs.ErrNotFound, err)
}
}