package service import ( "context" "encoding/json" "errors" "fmt" "io" "os/exec" "sort" "strconv" "strings" "sync" "time" "github.com/1Panel-dev/1Panel/backend/app/dto" "github.com/1Panel-dev/1Panel/backend/buserr" "github.com/1Panel-dev/1Panel/backend/constant" "github.com/1Panel-dev/1Panel/backend/global" "github.com/1Panel-dev/1Panel/backend/utils/common" "github.com/1Panel-dev/1Panel/backend/utils/docker" "github.com/docker/docker/api/types" "github.com/docker/docker/api/types/container" "github.com/docker/docker/api/types/filters" "github.com/docker/docker/api/types/network" "github.com/docker/docker/client" "github.com/docker/go-connections/nat" v1 "github.com/opencontainers/image-spec/specs-go/v1" ) type ContainerService struct{} type IContainerService interface { Page(req dto.PageContainer) (int64, interface{}, error) PageNetwork(req dto.SearchWithPage) (int64, interface{}, error) PageVolume(req dto.SearchWithPage) (int64, interface{}, error) ListVolume() ([]dto.Options, error) PageCompose(req dto.SearchWithPage) (int64, interface{}, error) CreateCompose(req dto.ComposeCreate) (string, error) ComposeOperation(req dto.ComposeOperation) error ContainerCreate(req dto.ContainerCreate) error ContainerOperation(req dto.ContainerOperation) error ContainerLogs(param dto.ContainerLog) (string, error) ContainerStats(id string) (*dto.ContainterStats, error) Inspect(req dto.InspectReq) (string, error) DeleteNetwork(req dto.BatchDelete) error CreateNetwork(req dto.NetworkCreat) error DeleteVolume(req dto.BatchDelete) error CreateVolume(req dto.VolumeCreat) error TestCompose(req dto.ComposeCreate) (bool, error) ComposeUpdate(req dto.ComposeUpdate) error } func NewIContainerService() IContainerService { return &ContainerService{} } func (u *ContainerService) Page(req dto.PageContainer) (int64, interface{}, error) { var ( records []types.Container list []types.Container backDatas []dto.ContainerInfo ) client, err := docker.NewDockerClient() if err != nil { return 0, nil, err } options := types.ContainerListOptions{All: true} if len(req.Filters) != 0 { options.Filters = filters.NewArgs() options.Filters.Add("label", req.Filters) } list, err = client.ContainerList(context.Background(), options) if err != nil { return 0, nil, err } if len(req.Name) != 0 { lenth, count := len(list), 0 for count < lenth { if !strings.Contains(list[count].Names[0][1:], req.Name) { list = append(list[:count], list[(count+1):]...) lenth-- } else { count++ } } } sort.Slice(list, func(i, j int) bool { return list[i].Created > list[j].Created }) total, start, end := len(list), (req.Page-1)*req.PageSize, req.Page*req.PageSize if start > total { records = make([]types.Container, 0) } else { if end >= total { end = total } records = list[start:end] } var wg sync.WaitGroup wg.Add(len(records)) for _, container := range records { go func(item types.Container) { IsFromCompose := false if _, ok := item.Labels[composeProjectLabel]; ok { IsFromCompose = true } IsFromApp := false if created, ok := item.Labels[composeCreatedBy]; ok && created == "Apps" { IsFromApp = true } var ports []string for _, port := range item.Ports { if port.IP == "::" || port.PublicPort == 0 { continue } ports = append(ports, fmt.Sprintf("%v:%v/%s", port.PublicPort, port.PrivatePort, port.Type)) } cpu, mem := loadCpuAndMem(client, item.ID) backDatas = append(backDatas, dto.ContainerInfo{ ContainerID: item.ID, CreateTime: time.Unix(item.Created, 0).Format("2006-01-02 15:04:05"), Name: item.Names[0][1:], ImageId: strings.Split(item.ImageID, ":")[1], ImageName: item.Image, State: item.State, RunTime: item.Status, CPUPercent: cpu, MemoryPercent: mem, Ports: ports, IsFromApp: IsFromApp, IsFromCompose: IsFromCompose, }) wg.Done() }(container) } wg.Wait() return int64(total), backDatas, nil } func (u *ContainerService) Inspect(req dto.InspectReq) (string, error) { client, err := docker.NewDockerClient() if err != nil { return "", err } var inspectInfo interface{} switch req.Type { case "container": inspectInfo, err = client.ContainerInspect(context.Background(), req.ID) case "network": inspectInfo, err = client.NetworkInspect(context.TODO(), req.ID, types.NetworkInspectOptions{}) case "volume": inspectInfo, err = client.VolumeInspect(context.TODO(), req.ID) } if err != nil { return "", err } bytes, err := json.Marshal(inspectInfo) if err != nil { return "", err } return string(bytes), nil } func (u *ContainerService) ContainerCreate(req dto.ContainerCreate) error { if len(req.ExposedPorts) != 0 { for _, port := range req.ExposedPorts { if strings.Contains(port.HostPort, "-") { portStart, _ := strconv.Atoi(strings.Split(port.HostPort, "-")[0]) portEnd, _ := strconv.Atoi(strings.Split(port.HostPort, "-")[1]) for i := portStart; i <= portEnd; i++ { if common.ScanPort(i) { return buserr.WithDetail(constant.ErrPortInUsed, i, nil) } } } else { portItem, _ := strconv.Atoi(port.HostPort) if common.ScanPort(portItem) { return buserr.WithDetail(constant.ErrPortInUsed, portItem, nil) } } } } client, err := docker.NewDockerClient() if err != nil { return err } config := &container.Config{ Image: req.Image, Cmd: req.Cmd, Env: req.Env, Labels: stringsToMap(req.Labels), Tty: true, OpenStdin: true, } hostConf := &container.HostConfig{ AutoRemove: req.AutoRemove, PublishAllPorts: req.PublishAllPorts, RestartPolicy: container.RestartPolicy{Name: req.RestartPolicy}, } if req.RestartPolicy == "on-failure" { hostConf.RestartPolicy.MaximumRetryCount = 5 } if req.NanoCPUs != 0 { hostConf.NanoCPUs = req.NanoCPUs * 1000000000 } if req.Memory != 0 { hostConf.Memory = req.Memory } if len(req.ExposedPorts) != 0 { hostConf.PortBindings = make(nat.PortMap) for _, port := range req.ExposedPorts { bindItem := nat.PortBinding{HostPort: port.HostPort, HostIP: port.HostIP} hostConf.PortBindings[nat.Port(fmt.Sprintf("%s/%s", port.ContainerPort, port.Protocol))] = []nat.PortBinding{bindItem} } } if len(req.Volumes) != 0 { config.Volumes = make(map[string]struct{}) for _, volume := range req.Volumes { config.Volumes[volume.ContainerDir] = struct{}{} hostConf.Binds = append(hostConf.Binds, fmt.Sprintf("%s:%s:%s", volume.SourceDir, volume.ContainerDir, volume.Mode)) } } global.LOG.Infof("new container info %s has been made, now start to create", req.Name) ctx := context.Background() if !checkImageExist(client, req.Image) { if err := pullImages(ctx, client, req.Image); err != nil { return err } } container, err := client.ContainerCreate(ctx, config, hostConf, &network.NetworkingConfig{}, &v1.Platform{}, req.Name) if err != nil { _ = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true}) return err } global.LOG.Infof("create container %s successful! now check if the container is started and delete the container information if it is not.", req.Name) if err := client.ContainerStart(ctx, container.ID, types.ContainerStartOptions{}); err != nil { _ = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true}) return fmt.Errorf("create successful but start failed, err: %v", err) } return nil } func (u *ContainerService) ContainerOperation(req dto.ContainerOperation) error { var err error ctx := context.Background() client, err := docker.NewDockerClient() if err != nil { return err } global.LOG.Infof("start container %s operation %s", req.Name, req.Operation) switch req.Operation { case constant.ContainerOpStart: err = client.ContainerStart(ctx, req.Name, types.ContainerStartOptions{}) case constant.ContainerOpStop: err = client.ContainerStop(ctx, req.Name, container.StopOptions{}) case constant.ContainerOpRestart: err = client.ContainerRestart(ctx, req.Name, container.StopOptions{}) case constant.ContainerOpKill: err = client.ContainerKill(ctx, req.Name, "SIGKILL") case constant.ContainerOpPause: err = client.ContainerPause(ctx, req.Name) case constant.ContainerOpUnpause: err = client.ContainerUnpause(ctx, req.Name) case constant.ContainerOpRename: err = client.ContainerRename(ctx, req.Name, req.NewName) case constant.ContainerOpRemove: err = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true}) } return err } func (u *ContainerService) ContainerLogs(req dto.ContainerLog) (string, error) { cmd := exec.Command("docker", "logs", req.ContainerID) if req.Mode != "all" { cmd = exec.Command("docker", "logs", req.ContainerID, "--since", req.Mode) } stdout, err := cmd.CombinedOutput() if err != nil { return "", errors.New(string(stdout)) } return string(stdout), nil } func (u *ContainerService) ContainerStats(id string) (*dto.ContainterStats, error) { client, err := docker.NewDockerClient() if err != nil { return nil, err } res, err := client.ContainerStats(context.TODO(), id, false) if err != nil { return nil, err } defer res.Body.Close() body, err := io.ReadAll(res.Body) if err != nil { res.Body.Close() return nil, err } res.Body.Close() var stats *types.StatsJSON if err := json.Unmarshal(body, &stats); err != nil { return nil, err } var data dto.ContainterStats data.CPUPercent = calculateCPUPercentUnix(stats) data.IORead, data.IOWrite = calculateBlockIO(stats.BlkioStats) data.Memory = float64(stats.MemoryStats.Usage) / 1024 / 1024 if cache, ok := stats.MemoryStats.Stats["cache"]; ok { data.Cache = float64(cache) / 1024 / 1024 } data.Memory = data.Memory - data.Cache data.NetworkRX, data.NetworkTX = calculateNetwork(stats.Networks) data.ShotTime = stats.Read return &data, nil } func stringsToMap(list []string) map[string]string { var lableMap = make(map[string]string) for _, label := range list { sps := strings.Split(label, "=") if len(sps) > 1 { lableMap[sps[0]] = sps[1] } } return lableMap } func calculateCPUPercentUnix(stats *types.StatsJSON) float64 { cpuPercent := 0.0 cpuDelta := float64(stats.CPUStats.CPUUsage.TotalUsage) - float64(stats.PreCPUStats.CPUUsage.TotalUsage) systemDelta := float64(stats.CPUStats.SystemUsage) - float64(stats.PreCPUStats.SystemUsage) if systemDelta > 0.0 && cpuDelta > 0.0 { cpuPercent = (cpuDelta / systemDelta) * float64(len(stats.CPUStats.CPUUsage.PercpuUsage)) * 100.0 } return cpuPercent } func calculateMemPercentUnix(memStats types.MemoryStats) float64 { memPercent := 0.0 memUsage := float64(memStats.Usage - memStats.Stats["cache"]) memLimit := float64(memStats.Limit) if memUsage > 0.0 && memLimit > 0.0 { memPercent = (memUsage / memLimit) * 100.0 } return memPercent } func calculateBlockIO(blkio types.BlkioStats) (blkRead float64, blkWrite float64) { for _, bioEntry := range blkio.IoServiceBytesRecursive { switch strings.ToLower(bioEntry.Op) { case "read": blkRead = (blkRead + float64(bioEntry.Value)) / 1024 / 1024 case "write": blkWrite = (blkWrite + float64(bioEntry.Value)) / 1024 / 1024 } } return } func calculateNetwork(network map[string]types.NetworkStats) (float64, float64) { var rx, tx float64 for _, v := range network { rx += float64(v.RxBytes) / 1024 tx += float64(v.TxBytes) / 1024 } return rx, tx } func checkImageExist(client *client.Client, image string) bool { images, err := client.ImageList(context.Background(), types.ImageListOptions{}) if err != nil { fmt.Println(err) return false } for _, img := range images { for _, tag := range img.RepoTags { if tag == image || tag == image+":latest" { return true } } } return false } func pullImages(ctx context.Context, client *client.Client, image string) error { out, err := client.ImagePull(ctx, image, types.ImagePullOptions{}) if err != nil { return err } defer out.Close() _, err = io.Copy(io.Discard, out) if err != nil { return err } return nil } func loadCpuAndMem(client *client.Client, container string) (float64, float64) { res, err := client.ContainerStats(context.Background(), container, false) if err != nil { return 0, 0 } body, err := io.ReadAll(res.Body) if err != nil { res.Body.Close() return 0, 0 } res.Body.Close() var stats *types.StatsJSON if err := json.Unmarshal(body, &stats); err != nil { return 0, 0 } CPUPercent := calculateCPUPercentUnix(stats) MemPercent := calculateMemPercentUnix(stats.MemoryStats) return CPUPercent, MemPercent }