mirror of
https://github.com/bcicen/ctop.git
synced 2024-08-30 18:23:19 +00:00
222a7d77b2
commit9968c504a5
Author: Florian Vahl <florian@flova.de> Date: Tue Feb 1 00:37:31 2022 +0100 Fix formating (Spaces -> Tabs) commit6b8840647e
Author: Florian <florian@flova.de> Date: Tue Feb 1 00:27:44 2022 +0100 Use better duration formating commit815dc6ec84
Author: Florian <florian@flova.de> Date: Tue Feb 1 00:27:04 2022 +0100 Fix date formating in 'created' col commit05b17798f8
Author: Florian <florian@flova.de> Date: Tue Feb 1 00:26:08 2022 +0100 Add fix from #275 for uptime of running containers
318 lines
7.7 KiB
Go
318 lines
7.7 KiB
Go
package connector
|
|
|
|
import (
|
|
"fmt"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/op/go-logging"
|
|
"github.com/hako/durafmt"
|
|
|
|
"github.com/bcicen/ctop/connector/collector"
|
|
"github.com/bcicen/ctop/connector/manager"
|
|
"github.com/bcicen/ctop/container"
|
|
api "github.com/fsouza/go-dockerclient"
|
|
)
|
|
|
|
func init() { enabled["docker"] = NewDocker }
|
|
|
|
var actionToStatus = map[string]string{
|
|
"start": "running",
|
|
"die": "exited",
|
|
"stop": "exited",
|
|
"pause": "paused",
|
|
"unpause": "running",
|
|
}
|
|
|
|
type StatusUpdate struct {
|
|
Cid string
|
|
Field string // "status" or "health"
|
|
Status string
|
|
}
|
|
|
|
type Docker struct {
|
|
client *api.Client
|
|
containers map[string]*container.Container
|
|
needsRefresh chan string // container IDs requiring refresh
|
|
statuses chan StatusUpdate
|
|
closed chan struct{}
|
|
lock sync.RWMutex
|
|
}
|
|
|
|
func NewDocker() (Connector, error) {
|
|
// init docker client
|
|
client, err := api.NewClientFromEnv()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
cm := &Docker{
|
|
client: client,
|
|
containers: make(map[string]*container.Container),
|
|
needsRefresh: make(chan string, 60),
|
|
statuses: make(chan StatusUpdate, 60),
|
|
closed: make(chan struct{}),
|
|
lock: sync.RWMutex{},
|
|
}
|
|
|
|
// query info as pre-flight healthcheck
|
|
info, err := client.Info()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugf("docker-connector ID: %s", info.ID)
|
|
log.Debugf("docker-connector Driver: %s", info.Driver)
|
|
log.Debugf("docker-connector Images: %d", info.Images)
|
|
log.Debugf("docker-connector Name: %s", info.Name)
|
|
log.Debugf("docker-connector ServerVersion: %s", info.ServerVersion)
|
|
|
|
go cm.Loop()
|
|
go cm.LoopStatuses()
|
|
cm.refreshAll()
|
|
go cm.watchEvents()
|
|
return cm, nil
|
|
}
|
|
|
|
// Docker implements Connector
|
|
func (cm *Docker) Wait() struct{} { return <-cm.closed }
|
|
|
|
// Docker events watcher
|
|
func (cm *Docker) watchEvents() {
|
|
log.Info("docker event listener starting")
|
|
events := make(chan *api.APIEvents)
|
|
opts := api.EventsOptions{Filters: map[string][]string{
|
|
"type": {"container"},
|
|
"event": {"create", "start", "health_status", "pause", "unpause", "stop", "die", "destroy"},
|
|
},
|
|
}
|
|
cm.client.AddEventListenerWithOptions(opts, events)
|
|
|
|
for e := range events {
|
|
actionName := e.Action
|
|
switch actionName {
|
|
// most frequent event is a health checks
|
|
case "health_status: healthy", "health_status: unhealthy":
|
|
sepIdx := strings.Index(actionName, ": ")
|
|
healthStatus := e.Action[sepIdx+2:]
|
|
if log.IsEnabledFor(logging.DEBUG) {
|
|
log.Debugf("handling docker event: action=health_status id=%s %s", e.ID, healthStatus)
|
|
}
|
|
cm.statuses <- StatusUpdate{e.ID, "health", healthStatus}
|
|
case "create":
|
|
if log.IsEnabledFor(logging.DEBUG) {
|
|
log.Debugf("handling docker event: action=create id=%s", e.ID)
|
|
}
|
|
cm.needsRefresh <- e.ID
|
|
case "destroy":
|
|
if log.IsEnabledFor(logging.DEBUG) {
|
|
log.Debugf("handling docker event: action=destroy id=%s", e.ID)
|
|
}
|
|
cm.delByID(e.ID)
|
|
default:
|
|
// check if this action changes status e.g. start -> running
|
|
status := actionToStatus[actionName]
|
|
if status != "" {
|
|
if log.IsEnabledFor(logging.DEBUG) {
|
|
log.Debugf("handling docker event: action=%s id=%s %s", actionName, e.ID, status)
|
|
}
|
|
cm.statuses <- StatusUpdate{e.ID, "status", status}
|
|
}
|
|
}
|
|
}
|
|
log.Info("docker event listener exited")
|
|
close(cm.closed)
|
|
}
|
|
|
|
func portsFormat(ports map[api.Port][]api.PortBinding) string {
|
|
var exposed []string
|
|
var published []string
|
|
|
|
for k, v := range ports {
|
|
if len(v) == 0 {
|
|
exposed = append(exposed, string(k))
|
|
continue
|
|
}
|
|
for _, binding := range v {
|
|
s := fmt.Sprintf("%s:%s -> %s", binding.HostIP, binding.HostPort, k)
|
|
published = append(published, s)
|
|
}
|
|
}
|
|
|
|
return strings.Join(append(exposed, published...), "\n")
|
|
}
|
|
|
|
func webPort(ports map[api.Port][]api.PortBinding) string {
|
|
for _, v := range ports {
|
|
if len(v) == 0 {
|
|
continue
|
|
}
|
|
for _, binding := range v {
|
|
publishedIp := binding.HostIP
|
|
if publishedIp == "0.0.0.0" {
|
|
publishedIp = "localhost"
|
|
}
|
|
publishedWebPort := fmt.Sprintf("%s:%s", publishedIp, binding.HostPort)
|
|
return publishedWebPort
|
|
}
|
|
}
|
|
return ""
|
|
}
|
|
|
|
func ipsFormat(networks map[string]api.ContainerNetwork) string {
|
|
var ips []string
|
|
|
|
for k, v := range networks {
|
|
s := fmt.Sprintf("%s:%s", k, v.IPAddress)
|
|
ips = append(ips, s)
|
|
}
|
|
|
|
return strings.Join(ips, "\n")
|
|
}
|
|
|
|
func (cm *Docker) refresh(c *container.Container) {
|
|
insp, found, failed := cm.inspect(c.Id)
|
|
if failed {
|
|
return
|
|
}
|
|
// remove container if no longer exists
|
|
if !found {
|
|
cm.delByID(c.Id)
|
|
return
|
|
}
|
|
c.SetMeta("name", shortName(insp.Name))
|
|
c.SetMeta("image", insp.Config.Image)
|
|
c.SetMeta("IPs", ipsFormat(insp.NetworkSettings.Networks))
|
|
c.SetMeta("ports", portsFormat(insp.NetworkSettings.Ports))
|
|
webPort := webPort(insp.NetworkSettings.Ports)
|
|
if webPort != "" {
|
|
c.SetMeta("Web Port", webPort)
|
|
}
|
|
c.SetMeta("created", insp.Created.Format("Mon Jan 02 15:04:05 2006"))
|
|
c.SetMeta("uptime", calcUptime(insp))
|
|
c.SetMeta("health", insp.State.Health.Status)
|
|
c.SetMeta("[ENV-VAR]", strings.Join(insp.Config.Env, ";"))
|
|
c.SetState(insp.State.Status)
|
|
}
|
|
|
|
func (cm *Docker) inspect(id string) (insp *api.Container, found bool, failed bool) {
|
|
c, err := cm.client.InspectContainer(id)
|
|
if err != nil {
|
|
if _, notFound := err.(*api.NoSuchContainer); notFound {
|
|
return c, false, false
|
|
}
|
|
// other error e.g. connection failed
|
|
log.Errorf("%s (%T)", err.Error(), err)
|
|
return c, false, true
|
|
}
|
|
return c, true, false
|
|
}
|
|
|
|
func calcUptime(insp *api.Container) string {
|
|
endTime := insp.State.FinishedAt
|
|
if endTime.IsZero() || insp.State.Running {
|
|
endTime = time.Now()
|
|
}
|
|
uptime := endTime.Sub(insp.State.StartedAt)
|
|
return durafmt.Parse(uptime).LimitFirstN(1).String()
|
|
}
|
|
|
|
// Mark all container IDs for refresh
|
|
func (cm *Docker) refreshAll() {
|
|
opts := api.ListContainersOptions{All: true}
|
|
allContainers, err := cm.client.ListContainers(opts)
|
|
if err != nil {
|
|
log.Errorf("%s (%T)", err.Error(), err)
|
|
return
|
|
}
|
|
|
|
for _, i := range allContainers {
|
|
c := cm.MustGet(i.ID)
|
|
c.SetMeta("name", shortName(i.Names[0]))
|
|
c.SetState(i.State)
|
|
cm.needsRefresh <- c.Id
|
|
}
|
|
}
|
|
|
|
func (cm *Docker) Loop() {
|
|
for {
|
|
select {
|
|
case id := <-cm.needsRefresh:
|
|
c := cm.MustGet(id)
|
|
cm.refresh(c)
|
|
case <-cm.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
func (cm *Docker) LoopStatuses() {
|
|
for {
|
|
select {
|
|
case statusUpdate := <-cm.statuses:
|
|
c, _ := cm.Get(statusUpdate.Cid)
|
|
if c != nil {
|
|
if statusUpdate.Field == "health" {
|
|
c.SetMeta("health", statusUpdate.Status)
|
|
} else {
|
|
c.SetState(statusUpdate.Status)
|
|
}
|
|
}
|
|
case <-cm.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// MustGet gets a single container, creating one anew if not existing
|
|
func (cm *Docker) MustGet(id string) *container.Container {
|
|
c, ok := cm.Get(id)
|
|
// append container struct for new containers
|
|
if !ok {
|
|
// create collector
|
|
collector := collector.NewDocker(cm.client, id)
|
|
// create manager
|
|
manager := manager.NewDocker(cm.client, id)
|
|
// create container
|
|
c = container.New(id, collector, manager)
|
|
cm.lock.Lock()
|
|
cm.containers[id] = c
|
|
cm.lock.Unlock()
|
|
}
|
|
return c
|
|
}
|
|
|
|
// Docker implements Connector
|
|
func (cm *Docker) Get(id string) (*container.Container, bool) {
|
|
cm.lock.Lock()
|
|
c, ok := cm.containers[id]
|
|
cm.lock.Unlock()
|
|
return c, ok
|
|
}
|
|
|
|
// Remove containers by ID
|
|
func (cm *Docker) delByID(id string) {
|
|
cm.lock.Lock()
|
|
delete(cm.containers, id)
|
|
cm.lock.Unlock()
|
|
log.Infof("removed dead container: %s", id)
|
|
}
|
|
|
|
// Docker implements Connector
|
|
func (cm *Docker) All() (containers container.Containers) {
|
|
cm.lock.Lock()
|
|
for _, c := range cm.containers {
|
|
containers = append(containers, c)
|
|
}
|
|
|
|
containers.Sort()
|
|
containers.Filter()
|
|
cm.lock.Unlock()
|
|
return containers
|
|
}
|
|
|
|
// use primary container name
|
|
func shortName(name string) string {
|
|
return strings.TrimPrefix(name, "/")
|
|
}
|