2019-01-12 04:58:27 +00:00
|
|
|
/*
|
|
|
|
Copyright 2017 The Kubernetes Authors.
|
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License.
|
|
|
|
*/
|
|
|
|
|
|
|
|
package server
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"strings"
|
|
|
|
|
|
|
|
"github.com/containerd/containerd"
|
|
|
|
containerdio "github.com/containerd/containerd/cio"
|
|
|
|
"github.com/containerd/containerd/errdefs"
|
|
|
|
"github.com/containerd/containerd/oci"
|
|
|
|
cni "github.com/containerd/go-cni"
|
|
|
|
"github.com/containerd/typeurl"
|
2019-04-07 17:07:55 +00:00
|
|
|
"github.com/davecgh/go-spew/spew"
|
2019-01-12 04:58:27 +00:00
|
|
|
imagespec "github.com/opencontainers/image-spec/specs-go/v1"
|
|
|
|
runtimespec "github.com/opencontainers/runtime-spec/specs-go"
|
|
|
|
"github.com/pkg/errors"
|
|
|
|
"github.com/sirupsen/logrus"
|
|
|
|
"golang.org/x/net/context"
|
|
|
|
"golang.org/x/sys/unix"
|
|
|
|
runtime "k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2"
|
|
|
|
|
|
|
|
"github.com/containerd/cri/pkg/annotations"
|
|
|
|
criconfig "github.com/containerd/cri/pkg/config"
|
|
|
|
customopts "github.com/containerd/cri/pkg/containerd/opts"
|
|
|
|
ctrdutil "github.com/containerd/cri/pkg/containerd/util"
|
|
|
|
"github.com/containerd/cri/pkg/log"
|
|
|
|
"github.com/containerd/cri/pkg/netns"
|
|
|
|
sandboxstore "github.com/containerd/cri/pkg/store/sandbox"
|
|
|
|
"github.com/containerd/cri/pkg/util"
|
|
|
|
)
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
typeurl.Register(&sandboxstore.Metadata{},
|
|
|
|
"github.com/containerd/cri/pkg/store/sandbox", "Metadata")
|
|
|
|
}
|
|
|
|
|
|
|
|
// RunPodSandbox creates and starts a pod-level sandbox. Runtimes should ensure
|
|
|
|
// the sandbox is in ready state.
|
|
|
|
func (c *criService) RunPodSandbox(ctx context.Context, r *runtime.RunPodSandboxRequest) (_ *runtime.RunPodSandboxResponse, retErr error) {
|
|
|
|
config := r.GetConfig()
|
2019-04-07 17:07:55 +00:00
|
|
|
logrus.Debugf("Sandbox config %+v", config)
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
// Generate unique id and name for the sandbox and reserve the name.
|
|
|
|
id := util.GenerateID()
|
|
|
|
name := makeSandboxName(config.GetMetadata())
|
|
|
|
logrus.Debugf("Generated id %q for sandbox %q", id, name)
|
|
|
|
// Reserve the sandbox name to avoid concurrent `RunPodSandbox` request starting the
|
|
|
|
// same sandbox.
|
|
|
|
if err := c.sandboxNameIndex.Reserve(name, id); err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to reserve sandbox name %q", name)
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
// Release the name if the function returns with an error.
|
|
|
|
if retErr != nil {
|
|
|
|
c.sandboxNameIndex.ReleaseByName(name)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Create initial internal sandbox object.
|
|
|
|
sandbox := sandboxstore.NewSandbox(
|
|
|
|
sandboxstore.Metadata{
|
|
|
|
ID: id,
|
|
|
|
Name: name,
|
|
|
|
Config: config,
|
|
|
|
RuntimeHandler: r.GetRuntimeHandler(),
|
|
|
|
},
|
|
|
|
sandboxstore.Status{
|
2019-02-08 04:04:22 +00:00
|
|
|
State: sandboxstore.StateInit,
|
2019-01-12 04:58:27 +00:00
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Ensure sandbox container image snapshot.
|
|
|
|
image, err := c.ensureImageExists(ctx, c.config.SandboxImage)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to get sandbox image %q", c.config.SandboxImage)
|
|
|
|
}
|
|
|
|
securityContext := config.GetLinux().GetSecurityContext()
|
|
|
|
//Create Network Namespace if it is not in host network
|
|
|
|
hostNet := securityContext.GetNamespaceOptions().GetNetwork() == runtime.NamespaceMode_NODE
|
|
|
|
if !hostNet {
|
|
|
|
// If it is not in host network namespace then create a namespace and set the sandbox
|
|
|
|
// handle. NetNSPath in sandbox metadata and NetNS is non empty only for non host network
|
|
|
|
// namespaces. If the pod is in host network namespace then both are empty and should not
|
|
|
|
// be used.
|
|
|
|
sandbox.NetNS, err = netns.NewNetNS()
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to create network namespace for sandbox %q", id)
|
|
|
|
}
|
|
|
|
sandbox.NetNSPath = sandbox.NetNS.GetPath()
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
if err := sandbox.NetNS.Remove(); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to remove network namespace %s for sandbox %q", sandbox.NetNSPath, id)
|
|
|
|
}
|
|
|
|
sandbox.NetNSPath = ""
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
// Setup network for sandbox.
|
|
|
|
// Certain VM based solutions like clear containers (Issue containerd/cri-containerd#524)
|
|
|
|
// rely on the assumption that CRI shim will not be querying the network namespace to check the
|
|
|
|
// network states such as IP.
|
|
|
|
// In future runtime implementation should avoid relying on CRI shim implementation details.
|
|
|
|
// In this case however caching the IP will add a subtle performance enhancement by avoiding
|
|
|
|
// calls to network namespace of the pod to query the IP of the veth interface on every
|
|
|
|
// SandboxStatus request.
|
|
|
|
sandbox.IP, err = c.setupPod(id, sandbox.NetNSPath, config)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to setup network for sandbox %q", id)
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
// Teardown network if an error is returned.
|
|
|
|
if err := c.teardownPod(id, sandbox.NetNSPath, config); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to destroy network for sandbox %q", id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
ociRuntime, err := c.getSandboxRuntime(config, r.GetRuntimeHandler())
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to get sandbox runtime")
|
|
|
|
}
|
|
|
|
logrus.Debugf("Use OCI %+v for sandbox %q", ociRuntime, id)
|
|
|
|
|
|
|
|
// Create sandbox container.
|
|
|
|
spec, err := c.generateSandboxContainerSpec(id, config, &image.ImageSpec.Config, sandbox.NetNSPath)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to generate sandbox container spec")
|
|
|
|
}
|
2019-04-07 17:07:55 +00:00
|
|
|
logrus.Debugf("Sandbox container %q spec: %#+v", id, spew.NewFormatter(spec))
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
var specOpts []oci.SpecOpts
|
|
|
|
userstr, err := generateUserString(
|
|
|
|
"",
|
|
|
|
securityContext.GetRunAsUser(),
|
|
|
|
securityContext.GetRunAsGroup(),
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to generate user string")
|
|
|
|
}
|
|
|
|
if userstr != "" {
|
|
|
|
specOpts = append(specOpts, oci.WithUser(userstr))
|
|
|
|
}
|
|
|
|
|
|
|
|
seccompSpecOpts, err := generateSeccompSpecOpts(
|
|
|
|
securityContext.GetSeccompProfilePath(),
|
|
|
|
securityContext.GetPrivileged(),
|
|
|
|
c.seccompEnabled)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to generate seccomp spec opts")
|
|
|
|
}
|
|
|
|
if seccompSpecOpts != nil {
|
|
|
|
specOpts = append(specOpts, seccompSpecOpts)
|
|
|
|
}
|
|
|
|
|
|
|
|
sandboxLabels := buildLabels(config.Labels, containerKindSandbox)
|
|
|
|
|
|
|
|
runtimeOpts, err := generateRuntimeOptions(ociRuntime, c.config)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to generate runtime options")
|
|
|
|
}
|
|
|
|
opts := []containerd.NewContainerOpts{
|
|
|
|
containerd.WithSnapshotter(c.config.ContainerdConfig.Snapshotter),
|
|
|
|
customopts.WithNewSnapshot(id, image.Image),
|
|
|
|
containerd.WithSpec(spec, specOpts...),
|
|
|
|
containerd.WithContainerLabels(sandboxLabels),
|
|
|
|
containerd.WithContainerExtension(sandboxMetadataExtension, &sandbox.Metadata),
|
|
|
|
containerd.WithRuntime(ociRuntime.Type, runtimeOpts)}
|
|
|
|
|
|
|
|
container, err := c.client.NewContainer(ctx, id, opts...)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to create containerd container")
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
|
|
defer deferCancel()
|
|
|
|
if err := container.Delete(deferCtx, containerd.WithSnapshotCleanup); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to delete containerd container %q", id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Create sandbox container root directories.
|
|
|
|
sandboxRootDir := c.getSandboxRootDir(id)
|
|
|
|
if err := c.os.MkdirAll(sandboxRootDir, 0755); err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to create sandbox root directory %q",
|
|
|
|
sandboxRootDir)
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
// Cleanup the sandbox root directory.
|
|
|
|
if err := c.os.RemoveAll(sandboxRootDir); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to remove sandbox root directory %q",
|
|
|
|
sandboxRootDir)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
volatileSandboxRootDir := c.getVolatileSandboxRootDir(id)
|
|
|
|
if err := c.os.MkdirAll(volatileSandboxRootDir, 0755); err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to create volatile sandbox root directory %q",
|
|
|
|
volatileSandboxRootDir)
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
// Cleanup the volatile sandbox root directory.
|
|
|
|
if err := c.os.RemoveAll(volatileSandboxRootDir); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to remove volatile sandbox root directory %q",
|
|
|
|
volatileSandboxRootDir)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
2019-04-07 17:07:55 +00:00
|
|
|
// Setup sandbox /dev/shm, /etc/hosts, /etc/resolv.conf and /etc/hostname.
|
2019-01-12 04:58:27 +00:00
|
|
|
if err = c.setupSandboxFiles(id, config); err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to setup sandbox files")
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
if err = c.unmountSandboxFiles(id, config); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to unmount sandbox files in %q",
|
|
|
|
sandboxRootDir)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Update sandbox created timestamp.
|
|
|
|
info, err := container.Info(ctx)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to get sandbox container info")
|
|
|
|
}
|
|
|
|
if err := sandbox.Status.Update(func(status sandboxstore.Status) (sandboxstore.Status, error) {
|
|
|
|
status.CreatedAt = info.CreatedAt
|
|
|
|
return status, nil
|
|
|
|
}); err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to update sandbox created timestamp")
|
|
|
|
}
|
|
|
|
|
2019-02-08 04:04:22 +00:00
|
|
|
// Add sandbox into sandbox store in INIT state.
|
2019-01-12 04:58:27 +00:00
|
|
|
sandbox.Container = container
|
|
|
|
if err := c.sandboxStore.Add(sandbox); err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to add sandbox %+v into store", sandbox)
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
// Delete sandbox from sandbox store if there is an error.
|
|
|
|
if retErr != nil {
|
|
|
|
c.sandboxStore.Delete(id)
|
|
|
|
}
|
|
|
|
}()
|
2019-02-08 04:04:22 +00:00
|
|
|
// NOTE(random-liu): Sandbox state only stay in INIT state after this point
|
2019-01-12 04:58:27 +00:00
|
|
|
// and before the end of this function.
|
|
|
|
// * If `Update` succeeds, sandbox state will become READY in one transaction.
|
|
|
|
// * If `Update` fails, sandbox will be removed from the store in the defer above.
|
|
|
|
// * If containerd stops at any point before `Update` finishes, because sandbox
|
|
|
|
// state is not checkpointed, it will be recovered from corresponding containerd task
|
|
|
|
// status during restart:
|
|
|
|
// * If the task is running, sandbox state will be READY,
|
|
|
|
// * Or else, sandbox state will be NOTREADY.
|
|
|
|
//
|
2019-02-08 04:04:22 +00:00
|
|
|
// In any case, sandbox will leave INIT state, so it's safe to ignore sandbox
|
|
|
|
// in INIT state in other functions.
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
// Start sandbox container in one transaction to avoid race condition with
|
|
|
|
// event monitor.
|
|
|
|
if err := sandbox.Status.Update(func(status sandboxstore.Status) (_ sandboxstore.Status, retErr error) {
|
|
|
|
// NOTE(random-liu): We should not change the sandbox state to NOTREADY
|
|
|
|
// if `Update` fails.
|
|
|
|
//
|
|
|
|
// If `Update` fails, the sandbox will be cleaned up by all the defers
|
|
|
|
// above. We should not let user see this sandbox, or else they will
|
|
|
|
// see the sandbox disappear after the defer clean up, which may confuse
|
|
|
|
// them.
|
|
|
|
//
|
2019-02-08 04:04:22 +00:00
|
|
|
// Given so, we should keep the sandbox in INIT state if `Update` fails,
|
|
|
|
// and ignore sandbox in INIT state in all the inspection functions.
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
// Create sandbox task in containerd.
|
|
|
|
log.Tracef("Create sandbox container (id=%q, name=%q).",
|
|
|
|
id, name)
|
|
|
|
|
|
|
|
var taskOpts []containerd.NewTaskOpts
|
|
|
|
// TODO(random-liu): Remove this after shim v1 is deprecated.
|
|
|
|
if c.config.NoPivot && ociRuntime.Type == linuxRuntime {
|
|
|
|
taskOpts = append(taskOpts, containerd.WithNoPivotRoot)
|
|
|
|
}
|
|
|
|
// We don't need stdio for sandbox container.
|
|
|
|
task, err := container.NewTask(ctx, containerdio.NullIO, taskOpts...)
|
|
|
|
if err != nil {
|
|
|
|
return status, errors.Wrap(err, "failed to create containerd task")
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if retErr != nil {
|
|
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
|
|
defer deferCancel()
|
|
|
|
// Cleanup the sandbox container if an error is returned.
|
|
|
|
// It's possible that task is deleted by event monitor.
|
|
|
|
if _, err := task.Delete(deferCtx, containerd.WithProcessKill); err != nil && !errdefs.IsNotFound(err) {
|
|
|
|
logrus.WithError(err).Errorf("Failed to delete sandbox container %q", id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
if err := task.Start(ctx); err != nil {
|
|
|
|
return status, errors.Wrapf(err, "failed to start sandbox container task %q", id)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the pod sandbox as ready after successfully start sandbox container.
|
|
|
|
status.Pid = task.Pid()
|
|
|
|
status.State = sandboxstore.StateReady
|
|
|
|
return status, nil
|
|
|
|
}); err != nil {
|
|
|
|
return nil, errors.Wrap(err, "failed to start sandbox container")
|
|
|
|
}
|
|
|
|
|
|
|
|
return &runtime.RunPodSandboxResponse{PodSandboxId: id}, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *criService) generateSandboxContainerSpec(id string, config *runtime.PodSandboxConfig,
|
|
|
|
imageConfig *imagespec.ImageConfig, nsPath string) (*runtimespec.Spec, error) {
|
|
|
|
// Creates a spec Generator with the default spec.
|
|
|
|
// TODO(random-liu): [P1] Compare the default settings with docker and containerd default.
|
|
|
|
spec, err := defaultRuntimeSpec(id)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
g := newSpecGenerator(spec)
|
|
|
|
|
|
|
|
// Apply default config from image config.
|
|
|
|
if err := addImageEnvs(&g, imageConfig.Env); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
if imageConfig.WorkingDir != "" {
|
|
|
|
g.SetProcessCwd(imageConfig.WorkingDir)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(imageConfig.Entrypoint) == 0 && len(imageConfig.Cmd) == 0 {
|
|
|
|
// Pause image must have entrypoint or cmd.
|
|
|
|
return nil, errors.Errorf("invalid empty entrypoint and cmd in image config %+v", imageConfig)
|
|
|
|
}
|
|
|
|
// Set process commands.
|
|
|
|
g.SetProcessArgs(append(imageConfig.Entrypoint, imageConfig.Cmd...))
|
|
|
|
|
|
|
|
// Set relative root path.
|
|
|
|
g.SetRootPath(relativeRootfsPath)
|
|
|
|
|
|
|
|
// Make root of sandbox container read-only.
|
|
|
|
g.SetRootReadonly(true)
|
|
|
|
|
|
|
|
// Set hostname.
|
|
|
|
g.SetHostname(config.GetHostname())
|
|
|
|
|
|
|
|
// TODO(random-liu): [P2] Consider whether to add labels and annotations to the container.
|
|
|
|
|
|
|
|
// Set cgroups parent.
|
2019-04-09 17:34:40 +00:00
|
|
|
if c.config.DisableCgroup {
|
|
|
|
g.SetLinuxCgroupsPath("")
|
|
|
|
} else {
|
|
|
|
if config.GetLinux().GetCgroupParent() != "" {
|
|
|
|
cgroupsPath := getCgroupsPath(config.GetLinux().GetCgroupParent(), id,
|
|
|
|
c.config.SystemdCgroup)
|
|
|
|
g.SetLinuxCgroupsPath(cgroupsPath)
|
|
|
|
}
|
2019-01-12 04:58:27 +00:00
|
|
|
}
|
|
|
|
// When cgroup parent is not set, containerd-shim will create container in a child cgroup
|
|
|
|
// of the cgroup itself is in.
|
|
|
|
// TODO(random-liu): [P2] Set default cgroup path if cgroup parent is not specified.
|
|
|
|
|
|
|
|
// Set namespace options.
|
|
|
|
securityContext := config.GetLinux().GetSecurityContext()
|
|
|
|
nsOptions := securityContext.GetNamespaceOptions()
|
|
|
|
if nsOptions.GetNetwork() == runtime.NamespaceMode_NODE {
|
|
|
|
g.RemoveLinuxNamespace(string(runtimespec.NetworkNamespace)) // nolint: errcheck
|
2019-04-07 17:07:55 +00:00
|
|
|
g.RemoveLinuxNamespace(string(runtimespec.UTSNamespace)) // nolint: errcheck
|
2019-01-12 04:58:27 +00:00
|
|
|
} else {
|
|
|
|
//TODO(Abhi): May be move this to containerd spec opts (WithLinuxSpaceOption)
|
|
|
|
g.AddOrReplaceLinuxNamespace(string(runtimespec.NetworkNamespace), nsPath) // nolint: errcheck
|
|
|
|
}
|
|
|
|
if nsOptions.GetPid() == runtime.NamespaceMode_NODE {
|
|
|
|
g.RemoveLinuxNamespace(string(runtimespec.PIDNamespace)) // nolint: errcheck
|
|
|
|
}
|
|
|
|
if nsOptions.GetIpc() == runtime.NamespaceMode_NODE {
|
|
|
|
g.RemoveLinuxNamespace(string(runtimespec.IPCNamespace)) // nolint: errcheck
|
|
|
|
}
|
|
|
|
|
|
|
|
// It's fine to generate the spec before the sandbox /dev/shm
|
|
|
|
// is actually created.
|
|
|
|
sandboxDevShm := c.getSandboxDevShm(id)
|
|
|
|
if nsOptions.GetIpc() == runtime.NamespaceMode_NODE {
|
|
|
|
sandboxDevShm = devShm
|
|
|
|
}
|
|
|
|
g.AddMount(runtimespec.Mount{
|
|
|
|
Source: sandboxDevShm,
|
|
|
|
Destination: devShm,
|
|
|
|
Type: "bind",
|
|
|
|
Options: []string{"rbind", "ro"},
|
|
|
|
})
|
|
|
|
|
|
|
|
selinuxOpt := securityContext.GetSelinuxOptions()
|
|
|
|
processLabel, mountLabel, err := initSelinuxOpts(selinuxOpt)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrapf(err, "failed to init selinux options %+v", securityContext.GetSelinuxOptions())
|
|
|
|
}
|
|
|
|
g.SetProcessSelinuxLabel(processLabel)
|
|
|
|
g.SetLinuxMountLabel(mountLabel)
|
|
|
|
|
|
|
|
supplementalGroups := securityContext.GetSupplementalGroups()
|
|
|
|
for _, group := range supplementalGroups {
|
|
|
|
g.AddProcessAdditionalGid(uint32(group))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add sysctls
|
|
|
|
sysctls := config.GetLinux().GetSysctls()
|
|
|
|
for key, value := range sysctls {
|
|
|
|
g.AddLinuxSysctl(key, value)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Note: LinuxSandboxSecurityContext does not currently provide an apparmor profile
|
|
|
|
|
2019-04-09 17:34:40 +00:00
|
|
|
if !c.config.DisableCgroup {
|
|
|
|
g.SetLinuxResourcesCPUShares(uint64(defaultSandboxCPUshares))
|
|
|
|
}
|
|
|
|
adj := int(defaultSandboxOOMAdj)
|
|
|
|
if c.config.RestrictOOMScoreAdj {
|
|
|
|
adj, err = restrictOOMScoreAdj(adj)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
g.SetProcessOOMScoreAdj(adj)
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
g.AddAnnotation(annotations.ContainerType, annotations.ContainerTypeSandbox)
|
|
|
|
g.AddAnnotation(annotations.SandboxID, id)
|
2019-04-07 17:07:55 +00:00
|
|
|
g.AddAnnotation(annotations.SandboxLogDir, config.GetLogDirectory())
|
2019-01-12 04:58:27 +00:00
|
|
|
|
|
|
|
return g.Config, nil
|
|
|
|
}
|
|
|
|
|
2019-04-07 17:07:55 +00:00
|
|
|
// setupSandboxFiles sets up necessary sandbox files including /dev/shm, /etc/hosts,
|
|
|
|
// /etc/resolv.conf and /etc/hostname.
|
2019-01-12 04:58:27 +00:00
|
|
|
func (c *criService) setupSandboxFiles(id string, config *runtime.PodSandboxConfig) error {
|
2019-04-07 17:07:55 +00:00
|
|
|
sandboxEtcHostname := c.getSandboxHostname(id)
|
|
|
|
hostname := config.GetHostname()
|
|
|
|
if hostname == "" {
|
|
|
|
var err error
|
|
|
|
hostname, err = c.os.Hostname()
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrap(err, "failed to get hostname")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if err := c.os.WriteFile(sandboxEtcHostname, []byte(hostname+"\n"), 0644); err != nil {
|
|
|
|
return errors.Wrapf(err, "failed to write hostname to %q", sandboxEtcHostname)
|
|
|
|
}
|
|
|
|
|
2019-01-12 04:58:27 +00:00
|
|
|
// TODO(random-liu): Consider whether we should maintain /etc/hosts and /etc/resolv.conf in kubelet.
|
|
|
|
sandboxEtcHosts := c.getSandboxHosts(id)
|
|
|
|
if err := c.os.CopyFile(etcHosts, sandboxEtcHosts, 0644); err != nil {
|
|
|
|
return errors.Wrapf(err, "failed to generate sandbox hosts file %q", sandboxEtcHosts)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set DNS options. Maintain a resolv.conf for the sandbox.
|
|
|
|
var err error
|
|
|
|
resolvContent := ""
|
|
|
|
if dnsConfig := config.GetDnsConfig(); dnsConfig != nil {
|
|
|
|
resolvContent, err = parseDNSOptions(dnsConfig.Servers, dnsConfig.Searches, dnsConfig.Options)
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrapf(err, "failed to parse sandbox DNSConfig %+v", dnsConfig)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
resolvPath := c.getResolvPath(id)
|
|
|
|
if resolvContent == "" {
|
|
|
|
// copy host's resolv.conf to resolvPath
|
|
|
|
err = c.os.CopyFile(resolvConfPath, resolvPath, 0644)
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrapf(err, "failed to copy host's resolv.conf to %q", resolvPath)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
err = c.os.WriteFile(resolvPath, []byte(resolvContent), 0644)
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrapf(err, "failed to write resolv content to %q", resolvPath)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Setup sandbox /dev/shm.
|
|
|
|
if config.GetLinux().GetSecurityContext().GetNamespaceOptions().GetIpc() == runtime.NamespaceMode_NODE {
|
|
|
|
if _, err := c.os.Stat(devShm); err != nil {
|
|
|
|
return errors.Wrapf(err, "host %q is not available for host ipc", devShm)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
sandboxDevShm := c.getSandboxDevShm(id)
|
|
|
|
if err := c.os.MkdirAll(sandboxDevShm, 0700); err != nil {
|
|
|
|
return errors.Wrap(err, "failed to create sandbox shm")
|
|
|
|
}
|
|
|
|
shmproperty := fmt.Sprintf("mode=1777,size=%d", defaultShmSize)
|
|
|
|
if err := c.os.Mount("shm", sandboxDevShm, "tmpfs", uintptr(unix.MS_NOEXEC|unix.MS_NOSUID|unix.MS_NODEV), shmproperty); err != nil {
|
|
|
|
return errors.Wrap(err, "failed to mount sandbox shm")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// parseDNSOptions parse DNS options into resolv.conf format content,
|
|
|
|
// if none option is specified, will return empty with no error.
|
|
|
|
func parseDNSOptions(servers, searches, options []string) (string, error) {
|
|
|
|
resolvContent := ""
|
|
|
|
|
|
|
|
if len(searches) > maxDNSSearches {
|
|
|
|
return "", errors.New("DNSOption.Searches has more than 6 domains")
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(searches) > 0 {
|
|
|
|
resolvContent += fmt.Sprintf("search %s\n", strings.Join(searches, " "))
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(servers) > 0 {
|
|
|
|
resolvContent += fmt.Sprintf("nameserver %s\n", strings.Join(servers, "\nnameserver "))
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(options) > 0 {
|
|
|
|
resolvContent += fmt.Sprintf("options %s\n", strings.Join(options, " "))
|
|
|
|
}
|
|
|
|
|
|
|
|
return resolvContent, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// unmountSandboxFiles unmount some sandbox files, we rely on the removal of sandbox root directory to
|
|
|
|
// remove these files. Unmount should *NOT* return error if the mount point is already unmounted.
|
|
|
|
func (c *criService) unmountSandboxFiles(id string, config *runtime.PodSandboxConfig) error {
|
|
|
|
if config.GetLinux().GetSecurityContext().GetNamespaceOptions().GetIpc() != runtime.NamespaceMode_NODE {
|
|
|
|
path, err := c.os.FollowSymlinkInScope(c.getSandboxDevShm(id), "/")
|
|
|
|
if err != nil {
|
|
|
|
return errors.Wrap(err, "failed to follow symlink")
|
|
|
|
}
|
|
|
|
if err := c.os.Unmount(path); err != nil && !os.IsNotExist(err) {
|
|
|
|
return errors.Wrapf(err, "failed to unmount %q", path)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// setupPod setups up the network for a pod
|
|
|
|
func (c *criService) setupPod(id string, path string, config *runtime.PodSandboxConfig) (string, error) {
|
|
|
|
if c.netPlugin == nil {
|
|
|
|
return "", errors.New("cni config not intialized")
|
|
|
|
}
|
|
|
|
|
|
|
|
labels := getPodCNILabels(id, config)
|
|
|
|
result, err := c.netPlugin.Setup(id,
|
|
|
|
path,
|
|
|
|
cni.WithLabels(labels),
|
|
|
|
cni.WithCapabilityPortMap(toCNIPortMappings(config.GetPortMappings())))
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
// Check if the default interface has IP config
|
|
|
|
if configs, ok := result.Interfaces[defaultIfName]; ok && len(configs.IPConfigs) > 0 {
|
|
|
|
return selectPodIP(configs.IPConfigs), nil
|
|
|
|
}
|
|
|
|
// If it comes here then the result was invalid so destroy the pod network and return error
|
|
|
|
if err := c.teardownPod(id, path, config); err != nil {
|
|
|
|
logrus.WithError(err).Errorf("Failed to destroy network for sandbox %q", id)
|
|
|
|
}
|
|
|
|
return "", errors.Errorf("failed to find network info for sandbox %q", id)
|
|
|
|
}
|
|
|
|
|
|
|
|
// toCNIPortMappings converts CRI port mappings to CNI.
|
|
|
|
func toCNIPortMappings(criPortMappings []*runtime.PortMapping) []cni.PortMapping {
|
|
|
|
var portMappings []cni.PortMapping
|
|
|
|
for _, mapping := range criPortMappings {
|
|
|
|
if mapping.HostPort <= 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if mapping.Protocol != runtime.Protocol_TCP && mapping.Protocol != runtime.Protocol_UDP {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
portMappings = append(portMappings, cni.PortMapping{
|
|
|
|
HostPort: mapping.HostPort,
|
|
|
|
ContainerPort: mapping.ContainerPort,
|
|
|
|
Protocol: strings.ToLower(mapping.Protocol.String()),
|
|
|
|
HostIP: mapping.HostIp,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
return portMappings
|
|
|
|
}
|
|
|
|
|
|
|
|
// selectPodIP select an ip from the ip list. It prefers ipv4 more than ipv6.
|
|
|
|
func selectPodIP(ipConfigs []*cni.IPConfig) string {
|
|
|
|
for _, c := range ipConfigs {
|
|
|
|
if c.IP.To4() != nil {
|
|
|
|
return c.IP.String()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ipConfigs[0].IP.String()
|
|
|
|
}
|
|
|
|
|
|
|
|
// untrustedWorkload returns true if the sandbox contains untrusted workload.
|
|
|
|
func untrustedWorkload(config *runtime.PodSandboxConfig) bool {
|
|
|
|
return config.GetAnnotations()[annotations.UntrustedWorkload] == "true"
|
|
|
|
}
|
|
|
|
|
|
|
|
// hostAccessingSandbox returns true if the sandbox configuration
|
|
|
|
// requires additional host access for the sandbox.
|
|
|
|
func hostAccessingSandbox(config *runtime.PodSandboxConfig) bool {
|
|
|
|
securityContext := config.GetLinux().GetSecurityContext()
|
|
|
|
|
|
|
|
namespaceOptions := securityContext.GetNamespaceOptions()
|
|
|
|
if namespaceOptions.GetNetwork() == runtime.NamespaceMode_NODE ||
|
|
|
|
namespaceOptions.GetPid() == runtime.NamespaceMode_NODE ||
|
|
|
|
namespaceOptions.GetIpc() == runtime.NamespaceMode_NODE {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// getSandboxRuntime returns the runtime configuration for sandbox.
|
|
|
|
// If the sandbox contains untrusted workload, runtime for untrusted workload will be returned,
|
|
|
|
// or else default runtime will be returned.
|
|
|
|
func (c *criService) getSandboxRuntime(config *runtime.PodSandboxConfig, runtimeHandler string) (criconfig.Runtime, error) {
|
|
|
|
if untrustedWorkload(config) {
|
|
|
|
// If the untrusted annotation is provided, runtimeHandler MUST be empty.
|
|
|
|
if runtimeHandler != "" && runtimeHandler != criconfig.RuntimeUntrusted {
|
|
|
|
return criconfig.Runtime{}, errors.New("untrusted workload with explicit runtime handler is not allowed")
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the untrusted workload is requesting access to the host/node, this request will fail.
|
|
|
|
//
|
|
|
|
// Note: If the workload is marked untrusted but requests privileged, this can be granted, as the
|
|
|
|
// runtime may support this. For example, in a virtual-machine isolated runtime, privileged
|
|
|
|
// is a supported option, granting the workload to access the entire guest VM instead of host.
|
|
|
|
if hostAccessingSandbox(config) {
|
|
|
|
return criconfig.Runtime{}, errors.New("untrusted workload with host access is not allowed")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Handle the deprecated UntrustedWorkloadRuntime.
|
|
|
|
if c.config.ContainerdConfig.UntrustedWorkloadRuntime.Type != "" {
|
|
|
|
return c.config.ContainerdConfig.UntrustedWorkloadRuntime, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
runtimeHandler = criconfig.RuntimeUntrusted
|
|
|
|
}
|
|
|
|
|
|
|
|
if runtimeHandler == "" {
|
|
|
|
return c.config.ContainerdConfig.DefaultRuntime, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
handler, ok := c.config.ContainerdConfig.Runtimes[runtimeHandler]
|
|
|
|
if !ok {
|
|
|
|
return criconfig.Runtime{}, errors.Errorf("no runtime for %q is configured", runtimeHandler)
|
|
|
|
}
|
|
|
|
return handler, nil
|
|
|
|
}
|