mailcow/data/Dockerfiles/dockerapi/dockerapi.py

548 lines
25 KiB
Python
Raw Normal View History

2022-08-23 11:57:05 +02:00
from fastapi import FastAPI, Response, Request
import aiodocker
2023-01-25 09:31:22 +01:00
import docker
2022-08-23 11:57:05 +02:00
import psutil
import sys
import re
import time
2017-12-09 13:15:24 +01:00
import os
2022-08-23 11:57:05 +02:00
import json
import asyncio
import redis
from datetime import datetime
2023-01-25 09:31:22 +01:00
import logging
from logging.config import dictConfig
log_config = {
"version": 1,
"disable_existing_loggers": False,
"formatters": {
"default": {
"()": "uvicorn.logging.DefaultFormatter",
"fmt": "%(levelprefix)s %(asctime)s %(message)s",
"datefmt": "%Y-%m-%d %H:%M:%S",
},
},
"handlers": {
"default": {
"formatter": "default",
"class": "logging.StreamHandler",
"stream": "ext://sys.stderr",
},
},
"loggers": {
"api-logger": {"handlers": ["default"], "level": "INFO"},
},
}
dictConfig(log_config)
2022-08-23 11:57:05 +02:00
containerIds_to_update = []
host_stats_isUpdating = False
app = FastAPI()
2023-01-25 09:31:22 +01:00
logger = logging.getLogger('api-logger')
2022-08-23 11:57:05 +02:00
@app.get("/host/stats")
async def get_host_update_stats():
global host_stats_isUpdating
if host_stats_isUpdating == False:
asyncio.create_task(get_host_stats())
host_stats_isUpdating = True
while True:
if redis_client.exists('host_stats'):
break
await asyncio.sleep(1.5)
stats = json.loads(redis_client.get('host_stats'))
return Response(content=json.dumps(stats, indent=4), media_type="application/json")
@app.get("/containers/{container_id}/json")
async def get_container(container_id : str):
if container_id and container_id.isalnum():
try:
for container in (await async_docker_client.containers.list()):
if container._id == container_id:
container_info = await container.show()
return Response(content=json.dumps(container_info, indent=4), media_type="application/json")
res = {
"type": "danger",
"msg": "no container found"
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
except Exception as e:
res = {
"type": "danger",
"msg": str(e)
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
res = {
"type": "danger",
"msg": "no or invalid id defined"
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
@app.get("/containers/json")
async def get_containers():
containers = {}
try:
for container in (await async_docker_client.containers.list()):
container_info = await container.show()
containers.update({container_info['Id']: container_info})
return Response(content=json.dumps(containers, indent=4), media_type="application/json")
except Exception as e:
res = {
"type": "danger",
"msg": str(e)
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
@app.post("/containers/{container_id}/{post_action}")
async def post_containers(container_id : str, post_action : str, request: Request):
try :
request_json = await request.json()
except Exception as err:
request_json = {}
if container_id and container_id.isalnum() and post_action:
2017-11-03 20:26:09 +01:00
try:
2022-08-23 11:57:05 +02:00
"""Dispatch container_post api call"""
if post_action == 'exec':
if not request_json or not 'cmd' in request_json:
res = {
"type": "danger",
"msg": "cmd is missing"
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
if not request_json or not 'task' in request_json:
res = {
"type": "danger",
"msg": "task is missing"
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
api_call_method_name = '__'.join(['container_post', str(post_action), str(request_json['cmd']), str(request_json['task']) ])
else:
api_call_method_name = '__'.join(['container_post', str(post_action) ])
2023-01-25 09:31:22 +01:00
docker_utils = DockerUtils(sync_docker_client)
2022-08-23 11:57:05 +02:00
api_call_method = getattr(docker_utils, api_call_method_name, lambda container_id: Response(content=json.dumps({'type': 'danger', 'msg':'container_post - unknown api call' }, indent=4), media_type="application/json"))
2023-01-25 09:31:22 +01:00
logger.info("api call: %s, container_id: %s" % (api_call_method_name, container_id))
return api_call_method(container_id, request_json)
2017-11-03 20:26:09 +01:00
except Exception as e:
2023-01-25 09:31:22 +01:00
logger.error("error - container_post: %s" % str(e))
2022-08-23 11:57:05 +02:00
res = {
"type": "danger",
"msg": str(e)
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2017-11-03 20:26:09 +01:00
2022-08-23 11:57:05 +02:00
else:
res = {
"type": "danger",
"msg": "invalid container id or missing action"
}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2017-11-03 20:26:09 +01:00
2022-08-23 11:57:05 +02:00
@app.post("/container/{container_id}/stats/update")
async def post_container_update_stats(container_id : str):
global containerIds_to_update
2018-10-27 13:25:18 +02:00
2022-08-23 11:57:05 +02:00
# start update task for container if no task is running
if container_id not in containerIds_to_update:
asyncio.create_task(get_container_stats(container_id))
containerIds_to_update.append(container_id)
while True:
if redis_client.exists(container_id + '_stats'):
break
await asyncio.sleep(1.5)
stats = json.loads(redis_client.get(container_id + '_stats'))
return Response(content=json.dumps(stats, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
class DockerUtils:
def __init__(self, docker_client):
self.docker_client = docker_client
# api call: container_post - post_action: stop
2023-01-25 09:31:22 +01:00
def container_post__stop(self, container_id, request_json):
for container in self.docker_client.containers.list(all=True, filters={"id": container_id}):
container.stop()
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { 'type': 'success', 'msg': 'command completed successfully'}
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2023-01-25 09:31:22 +01:00
# api call: container_post - post_action: start
def container_post__start(self, container_id, request_json):
for container in self.docker_client.containers.list(all=True, filters={"id": container_id}):
container.start()
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { 'type': 'success', 'msg': 'command completed successfully'}
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2023-01-25 09:31:22 +01:00
# api call: container_post - post_action: restart
def container_post__restart(self, container_id, request_json):
for container in self.docker_client.containers.list(all=True, filters={"id": container_id}):
container.restart()
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { 'type': 'success', 'msg': 'command completed successfully'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: top
2023-01-25 09:31:22 +01:00
def container_post__top(self, container_id, request_json):
for container in self.docker_client.containers.list(all=True, filters={"id": container_id}):
res = { 'type': 'success', 'msg': container.top()}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
# api call: container_post - post_action: stats
def container_post__stats(self, container_id, request_json):
for container in self.docker_client.containers.list(all=True, filters={"id": container_id}):
for stat in container.stats(decode=True, stream=True):
res = { 'type': 'success', 'msg': stat}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: delete
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__delete(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'items' in request_json:
2019-05-20 21:30:40 +02:00
r = re.compile("^[0-9a-fA-F]+$")
2022-08-23 11:57:05 +02:00
filtered_qids = filter(r.match, request_json['items'])
2019-05-20 21:30:40 +02:00
if filtered_qids:
flagged_qids = ['-d %s' % i for i in filtered_qids]
2023-01-25 09:31:22 +01:00
sanitized_string = str(' '.join(flagged_qids));
for container in self.docker_client.containers.list(filters={"id": container_id}):
postsuper_r = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postsuper " + sanitized_string])
return exec_run_handler('generic', postsuper_r)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: hold
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__hold(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'items' in request_json:
2019-05-20 21:30:40 +02:00
r = re.compile("^[0-9a-fA-F]+$")
2022-08-23 11:57:05 +02:00
filtered_qids = filter(r.match, request_json['items'])
2019-05-20 21:30:40 +02:00
if filtered_qids:
flagged_qids = ['-h %s' % i for i in filtered_qids]
2023-01-25 09:31:22 +01:00
sanitized_string = str(' '.join(flagged_qids));
for container in self.docker_client.containers.list(filters={"id": container_id}):
postsuper_r = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postsuper " + sanitized_string])
return exec_run_handler('generic', postsuper_r)
2017-12-09 13:15:24 +01:00
# api call: container_post - post_action: exec - cmd: mailq - task: cat
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__cat(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'items' in request_json:
r = re.compile("^[0-9a-fA-F]+$")
2022-08-23 11:57:05 +02:00
filtered_qids = filter(r.match, request_json['items'])
if filtered_qids:
2023-01-25 09:31:22 +01:00
sanitized_string = str(' '.join(filtered_qids));
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
postcat_return = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postcat -q " + sanitized_string], user='postfix')
if not postcat_return:
postcat_return = 'err: invalid'
return exec_run_handler('utf8_text_only', postcat_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: unhold
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__unhold(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'items' in request_json:
2019-05-20 21:30:40 +02:00
r = re.compile("^[0-9a-fA-F]+$")
2022-08-23 11:57:05 +02:00
filtered_qids = filter(r.match, request_json['items'])
2019-05-20 21:30:40 +02:00
if filtered_qids:
flagged_qids = ['-H %s' % i for i in filtered_qids]
2023-01-25 09:31:22 +01:00
sanitized_string = str(' '.join(flagged_qids));
for container in self.docker_client.containers.list(filters={"id": container_id}):
postsuper_r = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postsuper " + sanitized_string])
return exec_run_handler('generic', postsuper_r)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: deliver
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__deliver(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'items' in request_json:
2019-05-20 21:30:40 +02:00
r = re.compile("^[0-9a-fA-F]+$")
2022-08-23 11:57:05 +02:00
filtered_qids = filter(r.match, request_json['items'])
2019-05-20 21:30:40 +02:00
if filtered_qids:
flagged_qids = ['-i %s' % i for i in filtered_qids]
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
for i in flagged_qids:
postqueue_r = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postqueue " + i], user='postfix')
# todo: check each exit code
res = { 'type': 'success', 'msg': 'Scheduled immediate delivery'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: list
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__list(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
mailq_return = container.exec_run(["/usr/sbin/postqueue", "-j"], user='postfix')
return exec_run_handler('utf8_text_only', mailq_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: flush
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__flush(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
postqueue_r = container.exec_run(["/usr/sbin/postqueue", "-f"], user='postfix')
return exec_run_handler('generic', postqueue_r)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: super_delete
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__super_delete(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
postsuper_r = container.exec_run(["/usr/sbin/postsuper", "-d", "ALL"])
return exec_run_handler('generic', postsuper_r)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: system - task: fts_rescan
2023-01-25 09:31:22 +01:00
def container_post__exec__system__fts_rescan(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'username' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
rescan_return = container.exec_run(["/bin/bash", "-c", "/usr/bin/doveadm fts rescan -u '" + request_json['username'].replace("'", "'\\''") + "'"], user='vmail')
if rescan_return.exit_code == 0:
res = { 'type': 'success', 'msg': 'fts_rescan: rescan triggered'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
res = { 'type': 'warning', 'msg': 'fts_rescan error'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2022-08-23 11:57:05 +02:00
if 'all' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
rescan_return = container.exec_run(["/bin/bash", "-c", "/usr/bin/doveadm fts rescan -A"], user='vmail')
if rescan_return.exit_code == 0:
res = { 'type': 'success', 'msg': 'fts_rescan: rescan triggered'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
res = { 'type': 'warning', 'msg': 'fts_rescan error'}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: system - task: df
2023-01-25 09:31:22 +01:00
def container_post__exec__system__df(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'dir' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
df_return = container.exec_run(["/bin/bash", "-c", "/bin/df -H '" + request_json['dir'].replace("'", "'\\''") + "' | /usr/bin/tail -n1 | /usr/bin/tr -s [:blank:] | /usr/bin/tr ' ' ','"], user='nobody')
if df_return.exit_code == 0:
return df_return.output.decode('utf-8').rstrip()
else:
2023-01-25 09:31:22 +01:00
return "0,0,0,0,0,0"
# api call: container_post - post_action: exec - cmd: system - task: mysql_upgrade
def container_post__exec__system__mysql_upgrade(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
sql_return = container.exec_run(["/bin/bash", "-c", "/usr/bin/mysql_upgrade -uroot -p'" + os.environ['DBROOT'].replace("'", "'\\''") + "'\n"], user='mysql')
if sql_return.exit_code == 0:
matched = False
for line in sql_return.output.decode('utf-8').split("\n"):
if 'is already upgraded to' in line:
matched = True
if matched:
res = { 'type': 'success', 'msg':'mysql_upgrade: already upgraded', 'text': sql_return.output.decode('utf-8')}
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
2023-01-25 09:31:22 +01:00
container.restart()
res = { 'type': 'warning', 'msg':'mysql_upgrade: upgrade was applied', 'text': sql_return.output.decode('utf-8')}
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2023-01-25 09:31:22 +01:00
else:
res = { 'type': 'error', 'msg': 'mysql_upgrade: error running command', 'text': sql_return.output.decode('utf-8')}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
# api call: container_post - post_action: exec - cmd: system - task: mysql_tzinfo_to_sql
def container_post__exec__system__mysql_tzinfo_to_sql(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
sql_return = container.exec_run(["/bin/bash", "-c", "/usr/bin/mysql_tzinfo_to_sql /usr/share/zoneinfo | /bin/sed 's/Local time zone must be set--see zic manual page/FCTY/' | /usr/bin/mysql -uroot -p'" + os.environ['DBROOT'].replace("'", "'\\''") + "' mysql \n"], user='mysql')
if sql_return.exit_code == 0:
res = { 'type': 'info', 'msg': 'mysql_tzinfo_to_sql: command completed successfully', 'text': sql_return.output.decode('utf-8')}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
res = { 'type': 'error', 'msg': 'mysql_tzinfo_to_sql: error running command', 'text': sql_return.output.decode('utf-8')}
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: dovecot
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__dovecot(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
reload_return = container.exec_run(["/bin/bash", "-c", "/usr/sbin/dovecot reload"])
return exec_run_handler('generic', reload_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: postfix
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__postfix(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
reload_return = container.exec_run(["/bin/bash", "-c", "/usr/sbin/postfix reload"])
return exec_run_handler('generic', reload_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: nginx
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__nginx(self, container_id, request_json):
for container in self.docker_client.containers.list(filters={"id": container_id}):
reload_return = container.exec_run(["/bin/sh", "-c", "/usr/sbin/nginx -s reload"])
return exec_run_handler('generic', reload_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: sieve - task: list
2023-01-25 09:31:22 +01:00
def container_post__exec__sieve__list(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'username' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
sieve_return = container.exec_run(["/bin/bash", "-c", "/usr/bin/doveadm sieve list -u '" + request_json['username'].replace("'", "'\\''") + "'"])
return exec_run_handler('utf8_text_only', sieve_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: sieve - task: print
2023-01-25 09:31:22 +01:00
def container_post__exec__sieve__print(self, container_id, request_json):
if 'username' in request.json and 'script_name' in request_json:
for container in self.docker_client.containers.list(filters={"id": container_id}):
cmd = ["/bin/bash", "-c", "/usr/bin/doveadm sieve get -u '" + request_json['username'].replace("'", "'\\''") + "' '" + request_json['script_name'].replace("'", "'\\''") + "'"]
sieve_return = container.exec_run(cmd)
return exec_run_handler('utf8_text_only', sieve_return)
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: maildir - task: cleanup
2023-01-25 09:31:22 +01:00
def container_post__exec__maildir__cleanup(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'maildir' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
sane_name = re.sub(r'\W+', '', request_json['maildir'])
vmail_name = request_json['maildir'].replace("'", "'\\''")
cmd_vmail = "if [[ -d '/var/vmail/" + vmail_name + "' ]]; then /bin/mv '/var/vmail/" + vmail_name + "' '/var/vmail/_garbage/" + str(int(time.time())) + "_" + sane_name + "'; fi"
index_name = request_json['maildir'].split("/")
if len(index_name) > 1:
index_name = index_name[1].replace("'", "'\\''") + "@" + index_name[0].replace("'", "'\\''")
cmd_vmail_index = "if [[ -d '/var/vmail_index/" + index_name + "' ]]; then /bin/mv '/var/vmail_index/" + index_name + "' '/var/vmail/_garbage/" + str(int(time.time())) + "_" + sane_name + "_index'; fi"
cmd = ["/bin/bash", "-c", cmd_vmail + " && " + cmd_vmail_index]
else:
cmd = ["/bin/bash", "-c", cmd_vmail]
2023-01-25 09:31:22 +01:00
maildir_cleanup = container.exec_run(cmd, user='vmail')
return exec_run_handler('generic', maildir_cleanup)
2022-08-23 11:57:05 +02:00
# api call: container_post - post_action: exec - cmd: rspamd - task: worker_password
2023-01-25 09:31:22 +01:00
def container_post__exec__rspamd__worker_password(self, container_id, request_json):
2022-08-23 11:57:05 +02:00
if 'raw' in request_json:
2023-01-25 09:31:22 +01:00
for container in self.docker_client.containers.list(filters={"id": container_id}):
cmd = "/usr/bin/rspamadm pw -e -p '" + request_json['raw'].replace("'", "'\\''") + "' 2> /dev/null"
cmd_response = exec_cmd_container(container, cmd, user="_rspamd")
matched = False
for line in cmd_response.split("\n"):
if '$2$' in line:
hash = line.strip()
hash_out = re.search('\$2\$.+$', hash).group(0)
rspamd_passphrase_hash = re.sub('[^0-9a-zA-Z\$]+', '', hash_out.rstrip())
rspamd_password_filename = "/etc/rspamd/override.d/worker-controller-password.inc"
cmd = '''/bin/echo 'enable_password = "%s";' > %s && cat %s''' % (rspamd_passphrase_hash, rspamd_password_filename, rspamd_password_filename)
cmd_response = exec_cmd_container(container, cmd, user="_rspamd")
if rspamd_passphrase_hash.startswith("$2$") and rspamd_passphrase_hash in cmd_response:
container.restart()
matched = True
if matched:
res = { 'type': 'success', 'msg': 'command completed successfully' }
logger.info('success changing Rspamd password')
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
logger.error('failed changing Rspamd password')
res = { 'type': 'danger', 'msg': 'command did not complete' }
return Response(content=json.dumps(res, indent=4), media_type="application/json")
2022-08-23 11:57:05 +02:00
2023-01-25 09:31:22 +01:00
def exec_cmd_container(container, cmd, user, timeout=2, shell_cmd="/bin/bash"):
def recv_socket_data(c_socket, timeout):
c_socket.setblocking(0)
total_data=[]
data=''
begin=time.time()
while True:
if total_data and time.time()-begin > timeout:
break
elif time.time()-begin > timeout*2:
break
try:
data = c_socket.recv(8192)
if data:
total_data.append(data.decode('utf-8'))
#change the beginning time for measurement
begin=time.time()
else:
#sleep for sometime to indicate a gap
time.sleep(0.1)
break
except:
pass
return ''.join(total_data)
try :
socket = container.exec_run([shell_cmd], stdin=True, socket=True, user=user).output._sock
if not cmd.endswith("\n"):
cmd = cmd + "\n"
socket.send(cmd.encode('utf-8'))
data = recv_socket_data(socket, timeout)
socket.close()
return data
except Exception as e:
logger.error("error - exec_cmd_container: %s" % str(e))
traceback.print_exc(file=sys.stdout)
def exec_run_handler(type, output):
if type == 'generic':
if output.exit_code == 0:
res = { 'type': 'success', 'msg': 'command completed successfully' }
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
else:
2023-01-25 09:31:22 +01:00
res = { 'type': 'danger', 'msg': 'command failed: ' + output.output.decode('utf-8') }
2022-08-23 11:57:05 +02:00
return Response(content=json.dumps(res, indent=4), media_type="application/json")
if type == 'utf8_text_only':
2023-01-25 09:31:22 +01:00
return Response(content=output.output.decode('utf-8'), media_type="text/plain")
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
async def get_host_stats(wait=5):
global host_stats_isUpdating
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
try:
system_time = datetime.now()
host_stats = {
"cpu": {
"cores": psutil.cpu_count(),
"usage": psutil.cpu_percent()
},
"memory": {
"total": psutil.virtual_memory().total,
"usage": psutil.virtual_memory().percent,
"swap": psutil.swap_memory()
},
"uptime": time.time() - psutil.boot_time(),
"system_time": system_time.strftime("%d.%m.%Y %H:%M:%S")
}
redis_client.set('host_stats', json.dumps(host_stats), ex=10)
except Exception as e:
res = {
"type": "danger",
"msg": str(e)
}
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
await asyncio.sleep(wait)
host_stats_isUpdating = False
async def get_container_stats(container_id, wait=5, stop=False):
global containerIds_to_update
if container_id and container_id.isalnum():
2022-08-09 20:29:33 +02:00
try:
2022-08-23 11:57:05 +02:00
for container in (await async_docker_client.containers.list()):
if container._id == container_id:
res = await container.stats(stream=False)
if redis_client.exists(container_id + '_stats'):
stats = json.loads(redis_client.get(container_id + '_stats'))
else:
stats = []
stats.append(res[0])
if len(stats) > 3:
del stats[0]
redis_client.set(container_id + '_stats', json.dumps(stats), ex=60)
2022-08-09 20:29:33 +02:00
except Exception as e:
2022-08-23 11:57:05 +02:00
res = {
"type": "danger",
"msg": str(e)
}
else:
res = {
"type": "danger",
"msg": "no or invalid id defined"
}
await asyncio.sleep(wait)
if stop == True:
# update task was called second time, stop
containerIds_to_update.remove(container_id)
else:
# call update task a second time
await get_container_stats(container_id, wait=0, stop=True)
2023-01-25 09:31:22 +01:00
2022-08-23 11:57:05 +02:00
if os.environ['REDIS_SLAVEOF_IP'] != "":
redis_client = redis.Redis(host=os.environ['REDIS_SLAVEOF_IP'], port=os.environ['REDIS_SLAVEOF_PORT'], db=0)
else:
redis_client = redis.Redis(host='redis-mailcow', port=6379, db=0)
2023-01-25 09:31:22 +01:00
sync_docker_client = docker.DockerClient(base_url='unix://var/run/docker.sock', version='auto')
2022-08-23 11:57:05 +02:00
async_docker_client = aiodocker.Docker(url='unix:///var/run/docker.sock')
2023-01-25 09:31:22 +01:00
logger.info('DockerApi started')