mirror of
https://github.com/1Panel-dev/1Panel.git
synced 2024-11-11 18:05:59 +08:00
531 lines
16 KiB
Go
531 lines
16 KiB
Go
package service
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"os/exec"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/1Panel-dev/1Panel/backend/app/dto"
|
|
"github.com/1Panel-dev/1Panel/backend/buserr"
|
|
"github.com/1Panel-dev/1Panel/backend/constant"
|
|
"github.com/1Panel-dev/1Panel/backend/global"
|
|
"github.com/1Panel-dev/1Panel/backend/utils/common"
|
|
"github.com/1Panel-dev/1Panel/backend/utils/docker"
|
|
"github.com/docker/docker/api/types"
|
|
"github.com/docker/docker/api/types/container"
|
|
"github.com/docker/docker/api/types/filters"
|
|
"github.com/docker/docker/api/types/network"
|
|
"github.com/docker/docker/client"
|
|
"github.com/docker/go-connections/nat"
|
|
v1 "github.com/opencontainers/image-spec/specs-go/v1"
|
|
)
|
|
|
|
type ContainerService struct{}
|
|
|
|
type IContainerService interface {
|
|
Page(req dto.PageContainer) (int64, interface{}, error)
|
|
PageNetwork(req dto.SearchWithPage) (int64, interface{}, error)
|
|
PageVolume(req dto.SearchWithPage) (int64, interface{}, error)
|
|
ListVolume() ([]dto.Options, error)
|
|
PageCompose(req dto.SearchWithPage) (int64, interface{}, error)
|
|
CreateCompose(req dto.ComposeCreate) (string, error)
|
|
ComposeOperation(req dto.ComposeOperation) error
|
|
ContainerCreate(req dto.ContainerCreate) error
|
|
ContainerLogClean(req dto.OperationWithName) error
|
|
ContainerOperation(req dto.ContainerOperation) error
|
|
ContainerLogs(param dto.ContainerLog) (string, error)
|
|
ContainerStats(id string) (*dto.ContainterStats, error)
|
|
Inspect(req dto.InspectReq) (string, error)
|
|
DeleteNetwork(req dto.BatchDelete) error
|
|
CreateNetwork(req dto.NetworkCreate) error
|
|
DeleteVolume(req dto.BatchDelete) error
|
|
CreateVolume(req dto.VolumeCreate) error
|
|
TestCompose(req dto.ComposeCreate) (bool, error)
|
|
ComposeUpdate(req dto.ComposeUpdate) error
|
|
Prune(req dto.ContainerPrune) (dto.ContainerPruneReport, error)
|
|
}
|
|
|
|
func NewIContainerService() IContainerService {
|
|
return &ContainerService{}
|
|
}
|
|
|
|
func (u *ContainerService) Page(req dto.PageContainer) (int64, interface{}, error) {
|
|
var (
|
|
records []types.Container
|
|
list []types.Container
|
|
backDatas []dto.ContainerInfo
|
|
)
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return 0, nil, err
|
|
}
|
|
options := types.ContainerListOptions{All: true}
|
|
if len(req.Filters) != 0 {
|
|
options.Filters = filters.NewArgs()
|
|
options.Filters.Add("label", req.Filters)
|
|
}
|
|
list, err = client.ContainerList(context.Background(), options)
|
|
if err != nil {
|
|
return 0, nil, err
|
|
}
|
|
if len(req.Name) != 0 {
|
|
length, count := len(list), 0
|
|
for count < length {
|
|
if !strings.Contains(list[count].Names[0][1:], req.Name) {
|
|
list = append(list[:count], list[(count+1):]...)
|
|
length--
|
|
} else {
|
|
count++
|
|
}
|
|
}
|
|
}
|
|
sort.Slice(list, func(i, j int) bool {
|
|
return list[i].Created > list[j].Created
|
|
})
|
|
total, start, end := len(list), (req.Page-1)*req.PageSize, req.Page*req.PageSize
|
|
if start > total {
|
|
records = make([]types.Container, 0)
|
|
} else {
|
|
if end >= total {
|
|
end = total
|
|
}
|
|
records = list[start:end]
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(len(records))
|
|
for _, container := range records {
|
|
go func(item types.Container) {
|
|
IsFromCompose := false
|
|
if _, ok := item.Labels[composeProjectLabel]; ok {
|
|
IsFromCompose = true
|
|
}
|
|
IsFromApp := false
|
|
if created, ok := item.Labels[composeCreatedBy]; ok && created == "Apps" {
|
|
IsFromApp = true
|
|
}
|
|
|
|
var ports []string
|
|
for _, port := range item.Ports {
|
|
if port.IP == "::" || port.PublicPort == 0 {
|
|
continue
|
|
}
|
|
ports = append(ports, fmt.Sprintf("%v:%v/%s", port.PublicPort, port.PrivatePort, port.Type))
|
|
}
|
|
cpu, mem := loadCpuAndMem(client, item.ID)
|
|
backDatas = append(backDatas, dto.ContainerInfo{
|
|
ContainerID: item.ID,
|
|
CreateTime: time.Unix(item.Created, 0).Format("2006-01-02 15:04:05"),
|
|
Name: item.Names[0][1:],
|
|
ImageId: strings.Split(item.ImageID, ":")[1],
|
|
ImageName: item.Image,
|
|
State: item.State,
|
|
RunTime: item.Status,
|
|
CPUPercent: cpu,
|
|
MemoryPercent: mem,
|
|
Ports: ports,
|
|
IsFromApp: IsFromApp,
|
|
IsFromCompose: IsFromCompose,
|
|
})
|
|
wg.Done()
|
|
}(container)
|
|
}
|
|
wg.Wait()
|
|
|
|
return int64(total), backDatas, nil
|
|
}
|
|
|
|
func (u *ContainerService) Inspect(req dto.InspectReq) (string, error) {
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
var inspectInfo interface{}
|
|
switch req.Type {
|
|
case "container":
|
|
inspectInfo, err = client.ContainerInspect(context.Background(), req.ID)
|
|
case "network":
|
|
inspectInfo, err = client.NetworkInspect(context.TODO(), req.ID, types.NetworkInspectOptions{})
|
|
case "volume":
|
|
inspectInfo, err = client.VolumeInspect(context.TODO(), req.ID)
|
|
}
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
bytes, err := json.Marshal(inspectInfo)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
return string(bytes), nil
|
|
}
|
|
|
|
func (u *ContainerService) Prune(req dto.ContainerPrune) (dto.ContainerPruneReport, error) {
|
|
report := dto.ContainerPruneReport{}
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return report, err
|
|
}
|
|
pruneFilters := filters.NewArgs()
|
|
if req.WithTagAll {
|
|
pruneFilters.Add("dangling", "false")
|
|
if req.PruneType != "image" {
|
|
pruneFilters.Add("until", "24h")
|
|
}
|
|
}
|
|
switch req.PruneType {
|
|
case "container":
|
|
rep, err := client.ContainersPrune(context.Background(), pruneFilters)
|
|
if err != nil {
|
|
return report, err
|
|
}
|
|
report.DeletedNumber = len(rep.ContainersDeleted)
|
|
report.SpaceReclaimed = int(rep.SpaceReclaimed)
|
|
case "image":
|
|
rep, err := client.ImagesPrune(context.Background(), pruneFilters)
|
|
if err != nil {
|
|
return report, err
|
|
}
|
|
report.DeletedNumber = len(rep.ImagesDeleted)
|
|
report.SpaceReclaimed = int(rep.SpaceReclaimed)
|
|
case "network":
|
|
rep, err := client.NetworksPrune(context.Background(), pruneFilters)
|
|
if err != nil {
|
|
return report, err
|
|
}
|
|
report.DeletedNumber = len(rep.NetworksDeleted)
|
|
case "volume":
|
|
rep, err := client.VolumesPrune(context.Background(), pruneFilters)
|
|
if err != nil {
|
|
return report, err
|
|
}
|
|
report.DeletedNumber = len(rep.VolumesDeleted)
|
|
report.SpaceReclaimed = int(rep.SpaceReclaimed)
|
|
}
|
|
return report, nil
|
|
}
|
|
|
|
func (u *ContainerService) ContainerCreate(req dto.ContainerCreate) error {
|
|
portMap, err := checkPortStats(req.ExposedPorts)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
exposeds := make(nat.PortSet)
|
|
for port := range portMap {
|
|
exposeds[port] = struct{}{}
|
|
}
|
|
config := &container.Config{
|
|
Image: req.Image,
|
|
Cmd: req.Cmd,
|
|
Env: req.Env,
|
|
Labels: stringsToMap(req.Labels),
|
|
Tty: true,
|
|
OpenStdin: true,
|
|
ExposedPorts: exposeds,
|
|
}
|
|
hostConf := &container.HostConfig{
|
|
AutoRemove: req.AutoRemove,
|
|
PublishAllPorts: req.PublishAllPorts,
|
|
RestartPolicy: container.RestartPolicy{Name: req.RestartPolicy},
|
|
}
|
|
if req.RestartPolicy == "on-failure" {
|
|
hostConf.RestartPolicy.MaximumRetryCount = 5
|
|
}
|
|
if req.NanoCPUs != 0 {
|
|
hostConf.NanoCPUs = req.NanoCPUs * 1000000000
|
|
}
|
|
if req.Memory != 0 {
|
|
hostConf.Memory = req.Memory
|
|
}
|
|
if len(req.ExposedPorts) != 0 {
|
|
hostConf.PortBindings = portMap
|
|
}
|
|
if len(req.Volumes) != 0 {
|
|
config.Volumes = make(map[string]struct{})
|
|
for _, volume := range req.Volumes {
|
|
config.Volumes[volume.ContainerDir] = struct{}{}
|
|
hostConf.Binds = append(hostConf.Binds, fmt.Sprintf("%s:%s:%s", volume.SourceDir, volume.ContainerDir, volume.Mode))
|
|
}
|
|
}
|
|
|
|
global.LOG.Infof("new container info %s has been made, now start to create", req.Name)
|
|
|
|
ctx := context.Background()
|
|
if !checkImageExist(client, req.Image) {
|
|
if err := pullImages(ctx, client, req.Image); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
container, err := client.ContainerCreate(ctx, config, hostConf, &network.NetworkingConfig{}, &v1.Platform{}, req.Name)
|
|
if err != nil {
|
|
_ = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true})
|
|
return err
|
|
}
|
|
global.LOG.Infof("create container %s successful! now check if the container is started and delete the container information if it is not.", req.Name)
|
|
if err := client.ContainerStart(ctx, container.ID, types.ContainerStartOptions{}); err != nil {
|
|
_ = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true})
|
|
return fmt.Errorf("create successful but start failed, err: %v", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (u *ContainerService) ContainerOperation(req dto.ContainerOperation) error {
|
|
var err error
|
|
ctx := context.Background()
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
global.LOG.Infof("start container %s operation %s", req.Name, req.Operation)
|
|
switch req.Operation {
|
|
case constant.ContainerOpStart:
|
|
err = client.ContainerStart(ctx, req.Name, types.ContainerStartOptions{})
|
|
case constant.ContainerOpStop:
|
|
err = client.ContainerStop(ctx, req.Name, container.StopOptions{})
|
|
case constant.ContainerOpRestart:
|
|
err = client.ContainerRestart(ctx, req.Name, container.StopOptions{})
|
|
case constant.ContainerOpKill:
|
|
err = client.ContainerKill(ctx, req.Name, "SIGKILL")
|
|
case constant.ContainerOpPause:
|
|
err = client.ContainerPause(ctx, req.Name)
|
|
case constant.ContainerOpUnpause:
|
|
err = client.ContainerUnpause(ctx, req.Name)
|
|
case constant.ContainerOpRename:
|
|
err = client.ContainerRename(ctx, req.Name, req.NewName)
|
|
case constant.ContainerOpRemove:
|
|
err = client.ContainerRemove(ctx, req.Name, types.ContainerRemoveOptions{RemoveVolumes: true, Force: true})
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (u *ContainerService) ContainerLogClean(req dto.OperationWithName) error {
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
container, err := client.ContainerInspect(context.Background(), req.Name)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
file, err := os.OpenFile(container.LogPath, os.O_RDWR|os.O_CREATE, 0666)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer file.Close()
|
|
if err = file.Truncate(0); err != nil {
|
|
return err
|
|
}
|
|
_, _ = file.Seek(0, 0)
|
|
return nil
|
|
}
|
|
|
|
func (u *ContainerService) ContainerLogs(req dto.ContainerLog) (string, error) {
|
|
cmd := exec.Command("docker", "logs", req.ContainerID)
|
|
if req.Mode != "all" {
|
|
cmd = exec.Command("docker", "logs", req.ContainerID, "--since", req.Mode)
|
|
}
|
|
stdout, err := cmd.CombinedOutput()
|
|
if err != nil {
|
|
return "", errors.New(string(stdout))
|
|
}
|
|
return string(stdout), nil
|
|
}
|
|
|
|
func (u *ContainerService) ContainerStats(id string) (*dto.ContainterStats, error) {
|
|
client, err := docker.NewDockerClient()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
res, err := client.ContainerStats(context.TODO(), id, false)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer res.Body.Close()
|
|
|
|
body, err := io.ReadAll(res.Body)
|
|
if err != nil {
|
|
res.Body.Close()
|
|
return nil, err
|
|
}
|
|
res.Body.Close()
|
|
var stats *types.StatsJSON
|
|
if err := json.Unmarshal(body, &stats); err != nil {
|
|
return nil, err
|
|
}
|
|
var data dto.ContainterStats
|
|
data.CPUPercent = calculateCPUPercentUnix(stats)
|
|
data.IORead, data.IOWrite = calculateBlockIO(stats.BlkioStats)
|
|
data.Memory = float64(stats.MemoryStats.Usage) / 1024 / 1024
|
|
if cache, ok := stats.MemoryStats.Stats["cache"]; ok {
|
|
data.Cache = float64(cache) / 1024 / 1024
|
|
}
|
|
data.Memory = data.Memory - data.Cache
|
|
data.NetworkRX, data.NetworkTX = calculateNetwork(stats.Networks)
|
|
data.ShotTime = stats.Read
|
|
return &data, nil
|
|
}
|
|
|
|
func stringsToMap(list []string) map[string]string {
|
|
var lableMap = make(map[string]string)
|
|
for _, label := range list {
|
|
if strings.Contains(label, "=") {
|
|
sps := strings.SplitN(label, "=", 2)
|
|
lableMap[sps[0]] = sps[1]
|
|
}
|
|
}
|
|
return lableMap
|
|
}
|
|
|
|
func calculateCPUPercentUnix(stats *types.StatsJSON) float64 {
|
|
cpuPercent := 0.0
|
|
cpuDelta := float64(stats.CPUStats.CPUUsage.TotalUsage) - float64(stats.PreCPUStats.CPUUsage.TotalUsage)
|
|
systemDelta := float64(stats.CPUStats.SystemUsage) - float64(stats.PreCPUStats.SystemUsage)
|
|
|
|
if systemDelta > 0.0 && cpuDelta > 0.0 {
|
|
cpuPercent = (cpuDelta / systemDelta) * 100.0
|
|
if len(stats.CPUStats.CPUUsage.PercpuUsage) != 0 {
|
|
cpuPercent = cpuPercent * float64(len(stats.CPUStats.CPUUsage.PercpuUsage))
|
|
}
|
|
}
|
|
return cpuPercent
|
|
}
|
|
func calculateMemPercentUnix(memStats types.MemoryStats) float64 {
|
|
memPercent := 0.0
|
|
memUsage := float64(memStats.Usage - memStats.Stats["cache"])
|
|
memLimit := float64(memStats.Limit)
|
|
if memUsage > 0.0 && memLimit > 0.0 {
|
|
memPercent = (memUsage / memLimit) * 100.0
|
|
}
|
|
return memPercent
|
|
}
|
|
func calculateBlockIO(blkio types.BlkioStats) (blkRead float64, blkWrite float64) {
|
|
for _, bioEntry := range blkio.IoServiceBytesRecursive {
|
|
switch strings.ToLower(bioEntry.Op) {
|
|
case "read":
|
|
blkRead = (blkRead + float64(bioEntry.Value)) / 1024 / 1024
|
|
case "write":
|
|
blkWrite = (blkWrite + float64(bioEntry.Value)) / 1024 / 1024
|
|
}
|
|
}
|
|
return
|
|
}
|
|
func calculateNetwork(network map[string]types.NetworkStats) (float64, float64) {
|
|
var rx, tx float64
|
|
|
|
for _, v := range network {
|
|
rx += float64(v.RxBytes) / 1024
|
|
tx += float64(v.TxBytes) / 1024
|
|
}
|
|
return rx, tx
|
|
}
|
|
|
|
func checkImageExist(client *client.Client, image string) bool {
|
|
images, err := client.ImageList(context.Background(), types.ImageListOptions{})
|
|
if err != nil {
|
|
fmt.Println(err)
|
|
return false
|
|
}
|
|
|
|
for _, img := range images {
|
|
for _, tag := range img.RepoTags {
|
|
if tag == image || tag == image+":latest" {
|
|
return true
|
|
}
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func pullImages(ctx context.Context, client *client.Client, image string) error {
|
|
out, err := client.ImagePull(ctx, image, types.ImagePullOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer out.Close()
|
|
_, err = io.Copy(io.Discard, out)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func loadCpuAndMem(client *client.Client, container string) (float64, float64) {
|
|
res, err := client.ContainerStats(context.Background(), container, false)
|
|
if err != nil {
|
|
return 0, 0
|
|
}
|
|
|
|
body, err := io.ReadAll(res.Body)
|
|
if err != nil {
|
|
res.Body.Close()
|
|
return 0, 0
|
|
}
|
|
res.Body.Close()
|
|
var stats *types.StatsJSON
|
|
if err := json.Unmarshal(body, &stats); err != nil {
|
|
return 0, 0
|
|
}
|
|
|
|
CPUPercent := calculateCPUPercentUnix(stats)
|
|
MemPercent := calculateMemPercentUnix(stats.MemoryStats)
|
|
return CPUPercent, MemPercent
|
|
}
|
|
|
|
func checkPortStats(ports []dto.PortHelper) (nat.PortMap, error) {
|
|
portMap := make(nat.PortMap)
|
|
if len(ports) == 0 {
|
|
return portMap, nil
|
|
}
|
|
for _, port := range ports {
|
|
if strings.Contains(port.ContainerPort, "-") {
|
|
if !strings.Contains(port.HostPort, "-") {
|
|
return portMap, buserr.New(constant.ErrPortRules)
|
|
}
|
|
hostStart, _ := strconv.Atoi(strings.Split(port.HostPort, "-")[0])
|
|
hostEnd, _ := strconv.Atoi(strings.Split(port.HostPort, "-")[1])
|
|
containerStart, _ := strconv.Atoi(strings.Split(port.ContainerPort, "-")[0])
|
|
containerEnd, _ := strconv.Atoi(strings.Split(port.ContainerPort, "-")[1])
|
|
if (hostEnd-hostStart) <= 0 || (containerEnd-containerStart) <= 0 {
|
|
return portMap, buserr.New(constant.ErrPortRules)
|
|
}
|
|
if (containerEnd - containerStart) != (hostEnd - hostStart) {
|
|
return portMap, buserr.New(constant.ErrPortRules)
|
|
}
|
|
for i := 0; i <= hostEnd-hostStart; i++ {
|
|
bindItem := nat.PortBinding{HostPort: strconv.Itoa(hostStart + i), HostIP: port.HostIP}
|
|
portMap[nat.Port(fmt.Sprintf("%d/%s", containerStart+i, port.Protocol))] = []nat.PortBinding{bindItem}
|
|
}
|
|
for i := hostStart; i <= hostEnd; i++ {
|
|
if common.ScanPort(i) {
|
|
return portMap, buserr.WithDetail(constant.ErrPortInUsed, i, nil)
|
|
}
|
|
}
|
|
} else {
|
|
portItem := 0
|
|
if strings.Contains(port.HostPort, "-") {
|
|
portItem, _ = strconv.Atoi(strings.Split(port.HostPort, "-")[0])
|
|
} else {
|
|
portItem, _ = strconv.Atoi(port.HostPort)
|
|
}
|
|
if common.ScanPort(portItem) {
|
|
return portMap, buserr.WithDetail(constant.ErrPortInUsed, portItem, nil)
|
|
}
|
|
bindItem := nat.PortBinding{HostPort: strconv.Itoa(portItem), HostIP: port.HostIP}
|
|
portMap[nat.Port(fmt.Sprintf("%s/%s", port.ContainerPort, port.Protocol))] = []nat.PortBinding{bindItem}
|
|
}
|
|
}
|
|
return portMap, nil
|
|
}
|