2022-08-23 11:57:05 +02:00
from fastapi import FastAPI , Response , Request
import aiodocker
2023-01-25 09:31:22 +01:00
import docker
2022-08-23 11:57:05 +02:00
import psutil
import sys
import re
2017-10-27 11:22:39 +02:00
import time
2017-12-09 13:15:24 +01:00
import os
2022-08-23 11:57:05 +02:00
import json
import asyncio
import redis
from datetime import datetime
2023-01-25 09:31:22 +01:00
import logging
from logging . config import dictConfig
log_config = {
" version " : 1 ,
" disable_existing_loggers " : False ,
" formatters " : {
" default " : {
" () " : " uvicorn.logging.DefaultFormatter " ,
" fmt " : " %(levelprefix)s %(asctime)s %(message)s " ,
" datefmt " : " % Y- % m- %d % H: % M: % S " ,
} ,
} ,
" handlers " : {
" default " : {
" formatter " : " default " ,
" class " : " logging.StreamHandler " ,
" stream " : " ext://sys.stderr " ,
} ,
} ,
" loggers " : {
" api-logger " : { " handlers " : [ " default " ] , " level " : " INFO " } ,
} ,
}
dictConfig ( log_config )
2017-10-05 23:38:33 +02:00
2022-08-23 11:57:05 +02:00
containerIds_to_update = [ ]
host_stats_isUpdating = False
app = FastAPI ( )
2023-01-25 09:31:22 +01:00
logger = logging . getLogger ( ' api-logger ' )
2017-10-05 23:38:33 +02:00
2022-08-23 11:57:05 +02:00
@app.get ( " /host/stats " )
async def get_host_update_stats ( ) :
global host_stats_isUpdating
if host_stats_isUpdating == False :
asyncio . create_task ( get_host_stats ( ) )
host_stats_isUpdating = True
while True :
if redis_client . exists ( ' host_stats ' ) :
break
await asyncio . sleep ( 1.5 )
stats = json . loads ( redis_client . get ( ' host_stats ' ) )
return Response ( content = json . dumps ( stats , indent = 4 ) , media_type = " application/json " )
@app.get ( " /containers/ {container_id} /json " )
async def get_container ( container_id : str ) :
if container_id and container_id . isalnum ( ) :
try :
for container in ( await async_docker_client . containers . list ( ) ) :
if container . _id == container_id :
container_info = await container . show ( )
return Response ( content = json . dumps ( container_info , indent = 4 ) , media_type = " application/json " )
res = {
" type " : " danger " ,
" msg " : " no container found "
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
except Exception as e :
res = {
" type " : " danger " ,
" msg " : str ( e )
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
res = {
" type " : " danger " ,
" msg " : " no or invalid id defined "
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
@app.get ( " /containers/json " )
async def get_containers ( ) :
containers = { }
try :
for container in ( await async_docker_client . containers . list ( ) ) :
container_info = await container . show ( )
containers . update ( { container_info [ ' Id ' ] : container_info } )
return Response ( content = json . dumps ( containers , indent = 4 ) , media_type = " application/json " )
except Exception as e :
res = {
" type " : " danger " ,
" msg " : str ( e )
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
@app.post ( " /containers/ {container_id} / {post_action} " )
async def post_containers ( container_id : str , post_action : str , request : Request ) :
try :
request_json = await request . json ( )
except Exception as err :
request_json = { }
if container_id and container_id . isalnum ( ) and post_action :
2017-11-03 20:26:09 +01:00
try :
2022-08-23 11:57:05 +02:00
""" Dispatch container_post api call """
if post_action == ' exec ' :
if not request_json or not ' cmd ' in request_json :
res = {
" type " : " danger " ,
" msg " : " cmd is missing "
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
if not request_json or not ' task ' in request_json :
res = {
" type " : " danger " ,
" msg " : " task is missing "
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
api_call_method_name = ' __ ' . join ( [ ' container_post ' , str ( post_action ) , str ( request_json [ ' cmd ' ] ) , str ( request_json [ ' task ' ] ) ] )
else :
api_call_method_name = ' __ ' . join ( [ ' container_post ' , str ( post_action ) ] )
2023-01-25 09:31:22 +01:00
docker_utils = DockerUtils ( sync_docker_client )
2022-08-23 11:57:05 +02:00
api_call_method = getattr ( docker_utils , api_call_method_name , lambda container_id : Response ( content = json . dumps ( { ' type ' : ' danger ' , ' msg ' : ' container_post - unknown api call ' } , indent = 4 ) , media_type = " application/json " ) )
2023-01-25 09:31:22 +01:00
logger . info ( " api call: %s , container_id: %s " % ( api_call_method_name , container_id ) )
return api_call_method ( container_id , request_json )
2017-11-03 20:26:09 +01:00
except Exception as e :
2023-01-25 09:31:22 +01:00
logger . error ( " error - container_post: %s " % str ( e ) )
2022-08-23 11:57:05 +02:00
res = {
" type " : " danger " ,
" msg " : str ( e )
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2017-11-03 20:26:09 +01:00
2022-08-23 11:57:05 +02:00
else :
res = {
" type " : " danger " ,
" msg " : " invalid container id or missing action "
}
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2017-11-03 20:26:09 +01:00
2022-08-23 11:57:05 +02:00
@app.post ( " /container/ {container_id} /stats/update " )
async def post_container_update_stats ( container_id : str ) :
global containerIds_to_update
2018-10-27 13:25:18 +02:00
2022-08-23 11:57:05 +02:00
# start update task for container if no task is running
if container_id not in containerIds_to_update :
asyncio . create_task ( get_container_stats ( container_id ) )
containerIds_to_update . append ( container_id )
while True :
if redis_client . exists ( container_id + ' _stats ' ) :
break
await asyncio . sleep ( 1.5 )
stats = json . loads ( redis_client . get ( container_id + ' _stats ' ) )
return Response ( content = json . dumps ( stats , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
class DockerUtils :
def __init__ ( self , docker_client ) :
self . docker_client = docker_client
# api call: container_post - post_action: stop
2023-01-25 09:31:22 +01:00
def container_post__stop ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( all = True , filters = { " id " : container_id } ) :
container . stop ( )
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { ' type ' : ' success ' , ' msg ' : ' command completed successfully ' }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2023-01-25 09:31:22 +01:00
# api call: container_post - post_action: start
def container_post__start ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( all = True , filters = { " id " : container_id } ) :
container . start ( )
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { ' type ' : ' success ' , ' msg ' : ' command completed successfully ' }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2023-01-25 09:31:22 +01:00
# api call: container_post - post_action: restart
def container_post__restart ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( all = True , filters = { " id " : container_id } ) :
container . restart ( )
2019-05-20 21:30:40 +02:00
2023-01-25 09:31:22 +01:00
res = { ' type ' : ' success ' , ' msg ' : ' command completed successfully ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: top
2023-01-25 09:31:22 +01:00
def container_post__top ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( all = True , filters = { " id " : container_id } ) :
res = { ' type ' : ' success ' , ' msg ' : container . top ( ) }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
# api call: container_post - post_action: stats
def container_post__stats ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( all = True , filters = { " id " : container_id } ) :
for stat in container . stats ( decode = True , stream = True ) :
res = { ' type ' : ' success ' , ' msg ' : stat }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: delete
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__delete ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' items ' in request_json :
2019-05-20 21:30:40 +02:00
r = re . compile ( " ^[0-9a-fA-F]+$ " )
2022-08-23 11:57:05 +02:00
filtered_qids = filter ( r . match , request_json [ ' items ' ] )
2019-05-20 21:30:40 +02:00
if filtered_qids :
flagged_qids = [ ' -d %s ' % i for i in filtered_qids ]
2023-01-25 09:31:22 +01:00
sanitized_string = str ( ' ' . join ( flagged_qids ) ) ;
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postsuper_r = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postsuper " + sanitized_string ] )
return exec_run_handler ( ' generic ' , postsuper_r )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: hold
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__hold ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' items ' in request_json :
2019-05-20 21:30:40 +02:00
r = re . compile ( " ^[0-9a-fA-F]+$ " )
2022-08-23 11:57:05 +02:00
filtered_qids = filter ( r . match , request_json [ ' items ' ] )
2019-05-20 21:30:40 +02:00
if filtered_qids :
flagged_qids = [ ' -h %s ' % i for i in filtered_qids ]
2023-01-25 09:31:22 +01:00
sanitized_string = str ( ' ' . join ( flagged_qids ) ) ;
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postsuper_r = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postsuper " + sanitized_string ] )
return exec_run_handler ( ' generic ' , postsuper_r )
2017-12-09 13:15:24 +01:00
2020-05-19 20:15:01 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: cat
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__cat ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' items ' in request_json :
2020-05-19 20:15:01 +02:00
r = re . compile ( " ^[0-9a-fA-F]+$ " )
2022-08-23 11:57:05 +02:00
filtered_qids = filter ( r . match , request_json [ ' items ' ] )
2020-05-19 20:15:01 +02:00
if filtered_qids :
2023-01-25 09:31:22 +01:00
sanitized_string = str ( ' ' . join ( filtered_qids ) ) ;
2020-05-19 20:15:01 +02:00
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postcat_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postcat -q " + sanitized_string ] , user = ' postfix ' )
if not postcat_return :
postcat_return = ' err: invalid '
return exec_run_handler ( ' utf8_text_only ' , postcat_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: unhold
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__unhold ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' items ' in request_json :
2019-05-20 21:30:40 +02:00
r = re . compile ( " ^[0-9a-fA-F]+$ " )
2022-08-23 11:57:05 +02:00
filtered_qids = filter ( r . match , request_json [ ' items ' ] )
2019-05-20 21:30:40 +02:00
if filtered_qids :
flagged_qids = [ ' -H %s ' % i for i in filtered_qids ]
2023-01-25 09:31:22 +01:00
sanitized_string = str ( ' ' . join ( flagged_qids ) ) ;
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postsuper_r = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postsuper " + sanitized_string ] )
return exec_run_handler ( ' generic ' , postsuper_r )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: deliver
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__deliver ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' items ' in request_json :
2019-05-20 21:30:40 +02:00
r = re . compile ( " ^[0-9a-fA-F]+$ " )
2022-08-23 11:57:05 +02:00
filtered_qids = filter ( r . match , request_json [ ' items ' ] )
2019-05-20 21:30:40 +02:00
if filtered_qids :
flagged_qids = [ ' -i %s ' % i for i in filtered_qids ]
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
for i in flagged_qids :
postqueue_r = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postqueue " + i ] , user = ' postfix ' )
# todo: check each exit code
res = { ' type ' : ' success ' , ' msg ' : ' Scheduled immediate delivery ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: list
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__list ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
mailq_return = container . exec_run ( [ " /usr/sbin/postqueue " , " -j " ] , user = ' postfix ' )
return exec_run_handler ( ' utf8_text_only ' , mailq_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: flush
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__flush ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postqueue_r = container . exec_run ( [ " /usr/sbin/postqueue " , " -f " ] , user = ' postfix ' )
return exec_run_handler ( ' generic ' , postqueue_r )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: mailq - task: super_delete
2023-01-25 09:31:22 +01:00
def container_post__exec__mailq__super_delete ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
postsuper_r = container . exec_run ( [ " /usr/sbin/postsuper " , " -d " , " ALL " ] )
return exec_run_handler ( ' generic ' , postsuper_r )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: system - task: fts_rescan
2023-01-25 09:31:22 +01:00
def container_post__exec__system__fts_rescan ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' username ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
rescan_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/bin/doveadm fts rescan -u ' " + request_json [ ' username ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' " ] , user = ' vmail ' )
if rescan_return . exit_code == 0 :
res = { ' type ' : ' success ' , ' msg ' : ' fts_rescan: rescan triggered ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
res = { ' type ' : ' warning ' , ' msg ' : ' fts_rescan error ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2022-08-23 11:57:05 +02:00
if ' all ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
rescan_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/bin/doveadm fts rescan -A " ] , user = ' vmail ' )
if rescan_return . exit_code == 0 :
res = { ' type ' : ' success ' , ' msg ' : ' fts_rescan: rescan triggered ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
res = { ' type ' : ' warning ' , ' msg ' : ' fts_rescan error ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: system - task: df
2023-01-25 09:31:22 +01:00
def container_post__exec__system__df ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' dir ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
df_return = container . exec_run ( [ " /bin/bash " , " -c " , " /bin/df -H ' " + request_json [ ' dir ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' | /usr/bin/tail -n1 | /usr/bin/tr -s [:blank:] | /usr/bin/tr ' ' ' , ' " ] , user = ' nobody ' )
if df_return . exit_code == 0 :
return df_return . output . decode ( ' utf-8 ' ) . rstrip ( )
2019-09-23 21:42:47 +02:00
else :
2023-01-25 09:31:22 +01:00
return " 0,0,0,0,0,0 "
# api call: container_post - post_action: exec - cmd: system - task: mysql_upgrade
def container_post__exec__system__mysql_upgrade ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
sql_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/bin/mysql_upgrade -uroot -p ' " + os . environ [ ' DBROOT ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' \n " ] , user = ' mysql ' )
if sql_return . exit_code == 0 :
matched = False
for line in sql_return . output . decode ( ' utf-8 ' ) . split ( " \n " ) :
if ' is already upgraded to ' in line :
matched = True
if matched :
res = { ' type ' : ' success ' , ' msg ' : ' mysql_upgrade: already upgraded ' , ' text ' : sql_return . output . decode ( ' utf-8 ' ) }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
2023-01-25 09:31:22 +01:00
container . restart ( )
res = { ' type ' : ' warning ' , ' msg ' : ' mysql_upgrade: upgrade was applied ' , ' text ' : sql_return . output . decode ( ' utf-8 ' ) }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2023-01-25 09:31:22 +01:00
else :
res = { ' type ' : ' error ' , ' msg ' : ' mysql_upgrade: error running command ' , ' text ' : sql_return . output . decode ( ' utf-8 ' ) }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
# api call: container_post - post_action: exec - cmd: system - task: mysql_tzinfo_to_sql
def container_post__exec__system__mysql_tzinfo_to_sql ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
sql_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/bin/mysql_tzinfo_to_sql /usr/share/zoneinfo | /bin/sed ' s/Local time zone must be set--see zic manual page/FCTY/ ' | /usr/bin/mysql -uroot -p ' " + os . environ [ ' DBROOT ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' mysql \n " ] , user = ' mysql ' )
if sql_return . exit_code == 0 :
res = { ' type ' : ' info ' , ' msg ' : ' mysql_tzinfo_to_sql: command completed successfully ' , ' text ' : sql_return . output . decode ( ' utf-8 ' ) }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
res = { ' type ' : ' error ' , ' msg ' : ' mysql_tzinfo_to_sql: error running command ' , ' text ' : sql_return . output . decode ( ' utf-8 ' ) }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: dovecot
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__dovecot ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
reload_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/dovecot reload " ] )
return exec_run_handler ( ' generic ' , reload_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: postfix
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__postfix ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
reload_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/sbin/postfix reload " ] )
return exec_run_handler ( ' generic ' , reload_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: reload - task: nginx
2023-01-25 09:31:22 +01:00
def container_post__exec__reload__nginx ( self , container_id , request_json ) :
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
reload_return = container . exec_run ( [ " /bin/sh " , " -c " , " /usr/sbin/nginx -s reload " ] )
return exec_run_handler ( ' generic ' , reload_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: sieve - task: list
2023-01-25 09:31:22 +01:00
def container_post__exec__sieve__list ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' username ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
sieve_return = container . exec_run ( [ " /bin/bash " , " -c " , " /usr/bin/doveadm sieve list -u ' " + request_json [ ' username ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' " ] )
return exec_run_handler ( ' utf8_text_only ' , sieve_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: sieve - task: print
2023-01-25 09:31:22 +01:00
def container_post__exec__sieve__print ( self , container_id , request_json ) :
2023-04-21 16:15:16 +02:00
if ' username ' in request_json and ' script_name ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
cmd = [ " /bin/bash " , " -c " , " /usr/bin/doveadm sieve get -u ' " + request_json [ ' username ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' ' " + request_json [ ' script_name ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' " ]
sieve_return = container . exec_run ( cmd )
return exec_run_handler ( ' utf8_text_only ' , sieve_return )
2019-05-20 21:30:40 +02:00
# api call: container_post - post_action: exec - cmd: maildir - task: cleanup
2023-01-25 09:31:22 +01:00
def container_post__exec__maildir__cleanup ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' maildir ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
sane_name = re . sub ( r ' \ W+ ' , ' ' , request_json [ ' maildir ' ] )
2023-03-28 10:59:08 +02:00
vmail_name = request_json [ ' maildir ' ] . replace ( " ' " , " ' \\ ' ' " )
cmd_vmail = " if [[ -d ' /var/vmail/ " + vmail_name + " ' ]]; then /bin/mv ' /var/vmail/ " + vmail_name + " ' ' /var/vmail/_garbage/ " + str ( int ( time . time ( ) ) ) + " _ " + sane_name + " ' ; fi "
2023-04-20 13:41:11 +02:00
index_name = request_json [ ' maildir ' ] . split ( " / " )
if len ( index_name ) > 1 :
index_name = index_name [ 1 ] . replace ( " ' " , " ' \\ ' ' " ) + " @ " + index_name [ 0 ] . replace ( " ' " , " ' \\ ' ' " )
cmd_vmail_index = " if [[ -d ' /var/vmail_index/ " + index_name + " ' ]]; then /bin/mv ' /var/vmail_index/ " + index_name + " ' ' /var/vmail/_garbage/ " + str ( int ( time . time ( ) ) ) + " _ " + sane_name + " _index ' ; fi "
cmd = [ " /bin/bash " , " -c " , cmd_vmail + " && " + cmd_vmail_index ]
else :
cmd = [ " /bin/bash " , " -c " , cmd_vmail ]
2023-01-25 09:31:22 +01:00
maildir_cleanup = container . exec_run ( cmd , user = ' vmail ' )
return exec_run_handler ( ' generic ' , maildir_cleanup )
2022-08-23 11:57:05 +02:00
# api call: container_post - post_action: exec - cmd: rspamd - task: worker_password
2023-01-25 09:31:22 +01:00
def container_post__exec__rspamd__worker_password ( self , container_id , request_json ) :
2022-08-23 11:57:05 +02:00
if ' raw ' in request_json :
2023-01-25 09:31:22 +01:00
for container in self . docker_client . containers . list ( filters = { " id " : container_id } ) :
cmd = " /usr/bin/rspamadm pw -e -p ' " + request_json [ ' raw ' ] . replace ( " ' " , " ' \\ ' ' " ) + " ' 2> /dev/null "
cmd_response = exec_cmd_container ( container , cmd , user = " _rspamd " )
matched = False
for line in cmd_response . split ( " \n " ) :
if ' $2$ ' in line :
hash = line . strip ( )
hash_out = re . search ( ' \ $2 \ $.+$ ' , hash ) . group ( 0 )
rspamd_passphrase_hash = re . sub ( ' [^0-9a-zA-Z \ $]+ ' , ' ' , hash_out . rstrip ( ) )
rspamd_password_filename = " /etc/rspamd/override.d/worker-controller-password.inc "
cmd = ''' /bin/echo ' enable_password = " %s " ; ' > %s && cat %s ''' % ( rspamd_passphrase_hash , rspamd_password_filename , rspamd_password_filename )
cmd_response = exec_cmd_container ( container , cmd , user = " _rspamd " )
if rspamd_passphrase_hash . startswith ( " $2$ " ) and rspamd_passphrase_hash in cmd_response :
container . restart ( )
matched = True
if matched :
res = { ' type ' : ' success ' , ' msg ' : ' command completed successfully ' }
logger . info ( ' success changing Rspamd password ' )
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
logger . error ( ' failed changing Rspamd password ' )
res = { ' type ' : ' danger ' , ' msg ' : ' command did not complete ' }
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
2022-08-23 11:57:05 +02:00
2023-01-25 09:31:22 +01:00
def exec_cmd_container ( container , cmd , user , timeout = 2 , shell_cmd = " /bin/bash " ) :
def recv_socket_data ( c_socket , timeout ) :
c_socket . setblocking ( 0 )
total_data = [ ]
data = ' '
begin = time . time ( )
while True :
if total_data and time . time ( ) - begin > timeout :
break
elif time . time ( ) - begin > timeout * 2 :
break
try :
data = c_socket . recv ( 8192 )
if data :
total_data . append ( data . decode ( ' utf-8 ' ) )
#change the beginning time for measurement
begin = time . time ( )
else :
#sleep for sometime to indicate a gap
time . sleep ( 0.1 )
break
except :
pass
return ' ' . join ( total_data )
try :
socket = container . exec_run ( [ shell_cmd ] , stdin = True , socket = True , user = user ) . output . _sock
if not cmd . endswith ( " \n " ) :
cmd = cmd + " \n "
socket . send ( cmd . encode ( ' utf-8 ' ) )
data = recv_socket_data ( socket , timeout )
socket . close ( )
return data
except Exception as e :
logger . error ( " error - exec_cmd_container: %s " % str ( e ) )
traceback . print_exc ( file = sys . stdout )
def exec_run_handler ( type , output ) :
if type == ' generic ' :
if output . exit_code == 0 :
res = { ' type ' : ' success ' , ' msg ' : ' command completed successfully ' }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
else :
2023-01-25 09:31:22 +01:00
res = { ' type ' : ' danger ' , ' msg ' : ' command failed: ' + output . output . decode ( ' utf-8 ' ) }
2022-08-23 11:57:05 +02:00
return Response ( content = json . dumps ( res , indent = 4 ) , media_type = " application/json " )
if type == ' utf8_text_only ' :
2023-01-25 09:31:22 +01:00
return Response ( content = output . output . decode ( ' utf-8 ' ) , media_type = " text/plain " )
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
async def get_host_stats ( wait = 5 ) :
global host_stats_isUpdating
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
try :
system_time = datetime . now ( )
host_stats = {
" cpu " : {
" cores " : psutil . cpu_count ( ) ,
" usage " : psutil . cpu_percent ( )
} ,
" memory " : {
" total " : psutil . virtual_memory ( ) . total ,
" usage " : psutil . virtual_memory ( ) . percent ,
" swap " : psutil . swap_memory ( )
} ,
" uptime " : time . time ( ) - psutil . boot_time ( ) ,
" system_time " : system_time . strftime ( " %d . % m. % Y % H: % M: % S " )
}
redis_client . set ( ' host_stats ' , json . dumps ( host_stats ) , ex = 10 )
except Exception as e :
res = {
" type " : " danger " ,
" msg " : str ( e )
}
2019-05-20 21:30:40 +02:00
2022-08-23 11:57:05 +02:00
await asyncio . sleep ( wait )
host_stats_isUpdating = False
async def get_container_stats ( container_id , wait = 5 , stop = False ) :
global containerIds_to_update
if container_id and container_id . isalnum ( ) :
2022-08-09 20:29:33 +02:00
try :
2022-08-23 11:57:05 +02:00
for container in ( await async_docker_client . containers . list ( ) ) :
if container . _id == container_id :
res = await container . stats ( stream = False )
if redis_client . exists ( container_id + ' _stats ' ) :
stats = json . loads ( redis_client . get ( container_id + ' _stats ' ) )
else :
stats = [ ]
stats . append ( res [ 0 ] )
if len ( stats ) > 3 :
del stats [ 0 ]
redis_client . set ( container_id + ' _stats ' , json . dumps ( stats ) , ex = 60 )
2022-08-09 20:29:33 +02:00
except Exception as e :
2022-08-23 11:57:05 +02:00
res = {
" type " : " danger " ,
" msg " : str ( e )
}
else :
res = {
" type " : " danger " ,
" msg " : " no or invalid id defined "
}
await asyncio . sleep ( wait )
if stop == True :
# update task was called second time, stop
containerIds_to_update . remove ( container_id )
else :
# call update task a second time
await get_container_stats ( container_id , wait = 0 , stop = True )
2023-01-25 09:31:22 +01:00
2022-08-23 11:57:05 +02:00
if os . environ [ ' REDIS_SLAVEOF_IP ' ] != " " :
redis_client = redis . Redis ( host = os . environ [ ' REDIS_SLAVEOF_IP ' ] , port = os . environ [ ' REDIS_SLAVEOF_PORT ' ] , db = 0 )
else :
redis_client = redis . Redis ( host = ' redis-mailcow ' , port = 6379 , db = 0 )
2023-01-25 09:31:22 +01:00
sync_docker_client = docker . DockerClient ( base_url = ' unix://var/run/docker.sock ' , version = ' auto ' )
2022-08-23 11:57:05 +02:00
async_docker_client = aiodocker . Docker ( url = ' unix:///var/run/docker.sock ' )
2023-01-25 09:31:22 +01:00
logger . info ( ' DockerApi started ' )