mirror of
https://github.com/openappsec/attachment.git
synced 2025-06-28 16:41:03 +03:00
adding the webhook injector image
This commit is contained in:
parent
490436e2e7
commit
a1dc2bf807
26
docker/cloudguard-waf-webhook/Dockerfile
Executable file
26
docker/cloudguard-waf-webhook/Dockerfile
Executable file
@ -0,0 +1,26 @@
|
||||
# Use Python 3.9 slim image as the base
|
||||
FROM python:3.9-slim
|
||||
|
||||
# Install dependencies
|
||||
RUN pip install kubernetes cryptography flask --progress-bar off
|
||||
|
||||
# Create directory for certs
|
||||
RUN mkdir -p /certs
|
||||
|
||||
# Copy the Python scripts
|
||||
COPY keygen.py /app/keygen.py
|
||||
COPY webhook_server.py /app/webhook_server.py
|
||||
COPY secretgen.py /app/secretgen.py
|
||||
COPY run.sh /app/run.sh
|
||||
|
||||
# Make the script executable
|
||||
RUN chmod +x /app/run.sh
|
||||
|
||||
# Set the working directory
|
||||
WORKDIR /app
|
||||
|
||||
# Expose port 443
|
||||
EXPOSE 443
|
||||
|
||||
# Run the webhook server
|
||||
CMD ["/app/run.sh"]
|
85
docker/cloudguard-waf-webhook/keygen.py
Executable file
85
docker/cloudguard-waf-webhook/keygen.py
Executable file
@ -0,0 +1,85 @@
|
||||
#!/usr/bin/python3
|
||||
import base64
|
||||
import os
|
||||
import argparse
|
||||
import subprocess
|
||||
from shutil import which
|
||||
from pathlib import Path
|
||||
|
||||
def generate_keys(service, namespace, directory="generated"):
|
||||
"Generate key material and configuration for Kubernetes admission controllers"
|
||||
|
||||
if not which("openssl"):
|
||||
raise click.UsageError("Unable to detect the openssl CLI tool on the path")
|
||||
|
||||
if not os.path.exists(directory):
|
||||
os.makedirs(directory)
|
||||
|
||||
print("==> Generating CA")
|
||||
|
||||
command = """openssl genrsa -out ca.key 2048
|
||||
openssl req -x509 -new -nodes -key ca.key -days 100000 -out ca.crt -subj '/CN=admission_ca'"""
|
||||
|
||||
subprocess.run(command, cwd=directory, shell=True, stderr=subprocess.DEVNULL)
|
||||
|
||||
print("==> Creating configuration")
|
||||
|
||||
with open(os.path.sep.join((directory, "server.conf")), "w") as f:
|
||||
f.write(
|
||||
"""[req]
|
||||
default_bits = 2048
|
||||
req_extensions = v3_req
|
||||
distinguished_name = req_distinguished_name
|
||||
req_extensions = req_ext
|
||||
prompt = no
|
||||
[ req_distinguished_name ]
|
||||
CN = {service}.{namespace}.svc
|
||||
[ v3_req ]
|
||||
basicConstraints = CA:FALSE
|
||||
keyUsage = nonRepudiation, digitalSignature, keyEncipherment
|
||||
extendedKeyUsage = clientAuth, serverAuth
|
||||
[ req_ext ]
|
||||
subjectAltName = @alt_names
|
||||
[alt_names]
|
||||
DNS.1 = {service}
|
||||
DNS.2 = {service}.{namespace}
|
||||
DNS.3 = {service}.{namespace}.svc
|
||||
""".format(service = service, namespace = namespace)
|
||||
)
|
||||
|
||||
print("==> Generating private key and certificate")
|
||||
|
||||
address = "{}.{}.svc".format(service, namespace)
|
||||
|
||||
command = """openssl genrsa -out server.key 2048
|
||||
openssl req -out server.csr -newkey rsa:2048 -nodes -keyout server.key -config server.conf
|
||||
openssl x509 -req -in server.csr -CA ca.crt -CAkey ca.key -CAcreateserial -out server.crt -days 100000 -extensions req_ext -extfile server.conf""".format(
|
||||
ADDRESS = address
|
||||
)
|
||||
|
||||
subprocess.run(command, cwd=directory, shell=True, stderr=subprocess.DEVNULL)
|
||||
|
||||
print("==> Key material generated")
|
||||
|
||||
with open(os.path.sep.join((directory, "ca.crt")), "rb") as f:
|
||||
ca_cert = f.read()
|
||||
print("Use this as the caBundle:")
|
||||
print(base64.b64encode(ca_cert).decode("ascii"))
|
||||
|
||||
print("==> Command to create secret")
|
||||
print("Run this to upload the key material to a Kubernetes secret")
|
||||
print()
|
||||
|
||||
print(
|
||||
"kubectl --namespace={0} create secret tls {1}-certs --cert={2}/server.crt --key={2}/server.key".format(
|
||||
namespace, service, directory
|
||||
)
|
||||
)
|
||||
|
||||
if __name__ == "__main__":
|
||||
parser = argparse.ArgumentParser(description = 'Check Point Webhooks K8s Webhook Keygen')
|
||||
parser.add_argument("namespace", help = "Destination namespace")
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
generate_keys("envoy-injector", args.namespace)
|
10
docker/cloudguard-waf-webhook/run.sh
Executable file
10
docker/cloudguard-waf-webhook/run.sh
Executable file
@ -0,0 +1,10 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Start certificate manager in the background
|
||||
# Start the webhook server
|
||||
while [ true ]; do
|
||||
echo "Starting Webhook Server"
|
||||
python3 webhook_server.py
|
||||
echo "Webhook Server crashed, restarting..."
|
||||
sleep 5 # Pause before restarting
|
||||
done
|
65
docker/cloudguard-waf-webhook/secretgen.py
Executable file
65
docker/cloudguard-waf-webhook/secretgen.py
Executable file
@ -0,0 +1,65 @@
|
||||
#!/usr/bin/python
|
||||
|
||||
import os
|
||||
import base64
|
||||
import kubernetes.client
|
||||
from kubernetes.client.rest import ApiException
|
||||
from kubernetes import client, config
|
||||
from pprint import pprint
|
||||
|
||||
# Key generation script
|
||||
import keygen
|
||||
|
||||
SERVICE_PATH = "/var/run/secrets/kubernetes.io/serviceaccount/"
|
||||
API_SERVER = "https://kubernetes.default.svc"
|
||||
|
||||
GENERATED_CERTS_FOLDER = "/certs/"
|
||||
|
||||
config.load_incluster_config()
|
||||
|
||||
def getToken():
|
||||
with open(os.path.sep.join((SERVICE_PATH, "token")), "r") as f:
|
||||
return f.read()
|
||||
|
||||
def main():
|
||||
# First, generate keys
|
||||
keygen.generate_keys("openappsec-waf-webhook-svc", os.environ["K8S_NAMESPACE"], GENERATED_CERTS_FOLDER)
|
||||
|
||||
found = None
|
||||
|
||||
api_instance = client.AdmissionregistrationV1Api()
|
||||
|
||||
try:
|
||||
api_response = api_instance.list_mutating_webhook_configuration()
|
||||
|
||||
for result in api_response.items:
|
||||
print(result.metadata.name)
|
||||
if "openappsec-waf.injector" in result.metadata.name:
|
||||
pprint(result)
|
||||
found = result
|
||||
break
|
||||
|
||||
if found is None:
|
||||
raise Exception("Could not find webhook")
|
||||
|
||||
# Change the CA file
|
||||
with open(os.path.sep.join((GENERATED_CERTS_FOLDER, "ca.crt")), "rb") as f:
|
||||
cert = base64.b64encode(f.read()).decode("utf-8")
|
||||
|
||||
print("CA Cert:", cert)
|
||||
|
||||
# Update cert
|
||||
for webhook in found.webhooks:
|
||||
if "openappsec-waf.injector" in webhook.name:
|
||||
webhook.client_config.ca_bundle = cert;
|
||||
|
||||
|
||||
# Patch
|
||||
response = api_instance.patch_mutating_webhook_configuration(found.metadata.name, found, pretty = "true")
|
||||
|
||||
pprint(response)
|
||||
except ApiException as e:
|
||||
print("Exception when calling AdmissionregistrationApi->get_api_group: %s\n" % e)
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
649
docker/cloudguard-waf-webhook/webhook_server.py
Executable file
649
docker/cloudguard-waf-webhook/webhook_server.py
Executable file
@ -0,0 +1,649 @@
|
||||
import os
|
||||
import json
|
||||
import logging
|
||||
import base64
|
||||
import secretgen
|
||||
import sys
|
||||
from kubernetes import client, config
|
||||
from flask import Flask, request, jsonify, Response
|
||||
|
||||
app = Flask(__name__)
|
||||
|
||||
# Read agent image and tag from environment variables
|
||||
AGENT_IMAGE = os.getenv('AGENT_IMAGE', 'ghcr.io/openappsec/agent')
|
||||
AGENT_TAG = os.getenv('AGENT_TAG', 'latest')
|
||||
FULL_AGENT_IMAGE = f"{AGENT_IMAGE}:{AGENT_TAG}"
|
||||
|
||||
config.load_incluster_config()
|
||||
|
||||
def configure_logging():
|
||||
# Read the DEBUG_LEVEL from environment variables, defaulting to WARNING
|
||||
DEBUG_LEVEL = os.getenv('DEBUG_LEVEL', 'WARNING').upper()
|
||||
|
||||
# Map the string value of DEBUG_LEVEL to actual logging level
|
||||
logging_levels = {
|
||||
'DEBUG': logging.DEBUG,
|
||||
'INFO': logging.INFO,
|
||||
'WARNING': logging.WARNING,
|
||||
'ERROR': logging.ERROR,
|
||||
'CRITICAL': logging.CRITICAL
|
||||
}
|
||||
|
||||
# Set the logging level based on the environment variable
|
||||
log_level = logging_levels.get(DEBUG_LEVEL, logging.INFO)
|
||||
|
||||
# Configure Flask's logger to handle the specified logging level
|
||||
handler = logging.StreamHandler()
|
||||
handler.setLevel(log_level)
|
||||
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
|
||||
handler.setFormatter(formatter)
|
||||
|
||||
# Remove any existing handlers
|
||||
if app.logger.hasHandlers():
|
||||
app.logger.handlers.clear()
|
||||
|
||||
app.logger.addHandler(handler)
|
||||
app.logger.setLevel(log_level)
|
||||
|
||||
# The sidecar container spec with configurable image
|
||||
def get_sidecar_container():
|
||||
app.logger.debug("Entering get_sidecar_container()")
|
||||
token = os.getenv("TOKEN")
|
||||
sidecar = {
|
||||
"name": "infinity-next-nano-agent",
|
||||
"image": FULL_AGENT_IMAGE,
|
||||
"imagePullPolicy": "Always",
|
||||
"command": ["/cp-nano-agent"],
|
||||
"args": [
|
||||
"--token",
|
||||
token
|
||||
],
|
||||
"env": [
|
||||
{"name": "registered_server", "value": "NGINX Server"}
|
||||
],
|
||||
"volumeMounts": [
|
||||
{"name": "envoy-attachment-shared", "mountPath": "/envoy/attachment/shared/"}
|
||||
],
|
||||
"resources": {},
|
||||
"securityContext": {
|
||||
"runAsNonRoot": False,
|
||||
"runAsUser": 0
|
||||
},
|
||||
"terminationMessagePath": "/dev/termination-log",
|
||||
"terminationMessagePolicy": "File"
|
||||
}
|
||||
app.logger.debug(f"Sidecar container spec: {sidecar}")
|
||||
app.logger.debug("Exiting get_sidecar_container()")
|
||||
return sidecar
|
||||
|
||||
def get_init_container():
|
||||
# Define the initContainer you want to inject
|
||||
init_container = {
|
||||
"name": "prepare-attachment",
|
||||
"image": FULL_AGENT_IMAGE,
|
||||
"imagePullPolicy": "Always",
|
||||
"command": [
|
||||
"sh", "-c",
|
||||
"mkdir -p /envoy/attachment/shared && cp -r /envoy/attachment/lib* /envoy/attachment/shared"
|
||||
],
|
||||
"volumeMounts": [
|
||||
{
|
||||
"mountPath": "/envoy/attachment/shared",
|
||||
"name": "envoy-attachment-shared"
|
||||
}
|
||||
]
|
||||
}
|
||||
app.logger.debug(f"Init container spec: {init_container}")
|
||||
app.logger.debug("Exiting get_init_container()")
|
||||
return init_container
|
||||
|
||||
# The volume mount configuration for both the original and sidecar containers
|
||||
def get_volume_mount():
|
||||
app.logger.debug("Entering get_volume_mount()")
|
||||
volume_mount = {
|
||||
"name": "envoy-attachment-shared",
|
||||
"mountPath": "/usr/lib/attachment/"
|
||||
}
|
||||
app.logger.debug(f"Volume mount spec: {volume_mount}")
|
||||
app.logger.debug("Exiting get_volume_mount()")
|
||||
return volume_mount
|
||||
|
||||
# Volume definition for the pod
|
||||
def get_volume_definition():
|
||||
app.logger.debug("Entering get_volume_definition()")
|
||||
volume_def = {
|
||||
"name": "envoy-attachment-shared",
|
||||
"emptyDir": {}
|
||||
}
|
||||
app.logger.debug(f"Volume definition: {volume_def}")
|
||||
app.logger.debug("Exiting get_volume_definition()")
|
||||
return volume_def
|
||||
|
||||
def add_env_if_not_exist(containers, container_name, patches):
|
||||
# Find the container by name
|
||||
container = next((c for c in containers if c.get('name') == container_name), None)
|
||||
|
||||
if container:
|
||||
# Get the existing environment variables (if any)
|
||||
env_vars = container.get('env', None)
|
||||
|
||||
if env_vars is None:
|
||||
# If no env variables exist, add an empty env array first
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": f"/spec/containers/{containers.index(container)}/env",
|
||||
"value": []
|
||||
})
|
||||
|
||||
def add_env_variable_value_from(containers, container_name, env_var_name, env_value, patches, value_from):
|
||||
"""Adds or updates a specified environment variable in a given container."""
|
||||
container_index = next((i for i, container in enumerate(containers) if container['name'] == container_name), None)
|
||||
|
||||
if container_index is not None:
|
||||
env_vars = containers[container_index].get('env', [])
|
||||
existing_env_var = next((env for env in env_vars if env['name'] == env_var_name), None)
|
||||
|
||||
if existing_env_var:
|
||||
env_var_patch = {
|
||||
"op": "replace",
|
||||
"path": f"/spec/containers/{container_index}/env/{env_vars.index(existing_env_var)}",
|
||||
"value": {"name": env_var_name, "valueFrom": value_from}
|
||||
}
|
||||
patches.append(env_var_patch)
|
||||
app.logger.debug(f"Updated {env_var_name} environment variable in {container_name} container to use valueFrom.")
|
||||
else:
|
||||
env_var_patch = {
|
||||
"op": "add",
|
||||
"path": f"/spec/containers/{container_index}/env/-",
|
||||
"value": {"name": env_var_name, "valueFrom": value_from}
|
||||
}
|
||||
patches.append(env_var_patch)
|
||||
app.logger.debug(f"Added {env_var_name} environment variable with valueFrom to {container_name} container.")
|
||||
else:
|
||||
app.logger.warning(f"{container_name} container not found; no environment variable modification applied.")
|
||||
|
||||
def add_env_variable(containers, container_name, env_var_name, env_value, patches):
|
||||
"""Adds or updates a specified environment variable in a given container."""
|
||||
# Find the specified container by name
|
||||
container_index = next((i for i, container in enumerate(containers) if container['name'] == container_name), None)
|
||||
|
||||
if container_index is not None:
|
||||
# Get the list of environment variables for the specified container
|
||||
env_vars = containers[container_index].get('env', [])
|
||||
|
||||
# Find the specified environment variable if it exists
|
||||
existing_env_var = next((env for env in env_vars if env['name'] == env_var_name), None)
|
||||
|
||||
# If the environment variable exists, handle it based on its name
|
||||
if existing_env_var:
|
||||
current_value = existing_env_var['value']
|
||||
|
||||
if env_var_name == 'LD_LIBRARY_PATH':
|
||||
# For LD_LIBRARY_PATH, append env_value if not already present
|
||||
if env_value not in current_value:
|
||||
new_value = f"{current_value}:{env_value}"
|
||||
env_var_patch = {
|
||||
"op": "replace",
|
||||
"path": f"/spec/containers/{container_index}/env/{env_vars.index(existing_env_var)}/value",
|
||||
"value": new_value
|
||||
}
|
||||
patches.append(env_var_patch)
|
||||
app.logger.debug(f"Updated {env_var_name} environment variable in {container_name} container to new value.")
|
||||
else:
|
||||
app.logger.debug(f"{env_var_name} already exists with the correct value; no changes made.")
|
||||
else:
|
||||
# For other environment variables, replace the value directly
|
||||
env_var_patch = {
|
||||
"op": "replace",
|
||||
"path": f"/spec/containers/{container_index}/env/{env_vars.index(existing_env_var)}/value",
|
||||
"value": env_value
|
||||
}
|
||||
patches.append(env_var_patch)
|
||||
app.logger.debug(f"Replaced {env_var_name} environment variable in {container_name} container with new value.")
|
||||
|
||||
else:
|
||||
# Add the environment variable if it does not exist
|
||||
env_var_patch = {
|
||||
"op": "add",
|
||||
"path": f"/spec/containers/{container_index}/env/-",
|
||||
"value": {
|
||||
"name": env_var_name,
|
||||
"value": env_value
|
||||
}
|
||||
}
|
||||
patches.append(env_var_patch)
|
||||
app.logger.debug(f"Added {env_var_name} environment variable to {container_name} container.")
|
||||
else:
|
||||
app.logger.warning(f"{container_name} container not found; no environment variable modification applied.")
|
||||
|
||||
def remove_env_variable(containers, container_name, env_var_name, patches):
|
||||
"""Removes a specified environment variable from a given container if it exists."""
|
||||
# Find the specified container by name
|
||||
container_index = next((i for i, container in enumerate(containers) if container['name'] == container_name), None)
|
||||
|
||||
if container_index is not None:
|
||||
# Get the list of environment variables for the specified container
|
||||
env_vars = containers[container_index].get('env', [])
|
||||
|
||||
# Check if the specified environment variable exists
|
||||
env_var_exists = any(env['name'] == env_var_name for env in env_vars)
|
||||
|
||||
# Remove the environment variable if it exists
|
||||
if env_var_exists:
|
||||
# Find the index of the specified environment variable in the env array
|
||||
env_var_index = next(i for i, env in enumerate(env_vars) if env['name'] == env_var_name)
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": f"/spec/containers/{container_index}/env/{env_var_index}"
|
||||
})
|
||||
app.logger.debug(f"Removed {env_var_name} environment variable from {container_name} container.")
|
||||
else:
|
||||
app.logger.debug(f"{env_var_name} does not exist, nothing to remove.")
|
||||
else:
|
||||
app.logger.warning(f"{container_name} container not found; no environment variable modification applied.")
|
||||
|
||||
def create_or_update_envoy_filter(name, namespace, selector_label_name, selector_label_value):
|
||||
api = client.CustomObjectsApi()
|
||||
# Define the EnvoyFilter specification
|
||||
envoy_filter_spec = {
|
||||
"apiVersion": "networking.istio.io/v1alpha3",
|
||||
"kind": "EnvoyFilter",
|
||||
"metadata": {
|
||||
"name": name,
|
||||
"namespace": namespace,
|
||||
"labels": {
|
||||
"owner": "waf"
|
||||
}
|
||||
},
|
||||
"spec": {
|
||||
"workloadSelector": {
|
||||
"labels": {
|
||||
selector_label_name: selector_label_value
|
||||
}
|
||||
},
|
||||
"configPatches": [
|
||||
{
|
||||
"applyTo": "HTTP_FILTER",
|
||||
"match": {
|
||||
"context": "GATEWAY",
|
||||
"listener": {
|
||||
"filterChain": {
|
||||
"filter": {
|
||||
"name": "envoy.filters.network.http_connection_manager"
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"patch": {
|
||||
"operation": "INSERT_BEFORE",
|
||||
"value": {
|
||||
"name": "envoy.filters.http.golang",
|
||||
"typed_config": {
|
||||
"@type": "type.googleapis.com/envoy.extensions.filters.http.golang.v3alpha.Config",
|
||||
"library_id": "cp_nano_filter",
|
||||
"library_path": "/usr/lib/attachment/libenvoy_attachment.so",
|
||||
"plugin_name": "cp_nano_filter",
|
||||
"plugin_config": {
|
||||
"@type": "type.googleapis.com/xds.type.v3.TypedStruct",
|
||||
"type_url": "type.googleapis.com/envoy.extensions.filters.http.golang.v3alpha.PluginConfig",
|
||||
"value": {
|
||||
"prefix_localreply_body": "Configured local reply from go"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
|
||||
# Check if the EnvoyFilter exists
|
||||
try:
|
||||
existing_envoy_filter = api.get_namespaced_custom_object(
|
||||
group="networking.istio.io",
|
||||
version="v1alpha3",
|
||||
namespace=namespace,
|
||||
plural="envoyfilters",
|
||||
name=name
|
||||
)
|
||||
|
||||
# Compare workloadSelector labels
|
||||
existing_labels = existing_envoy_filter.get("spec", {}).get("workloadSelector", {}).get("labels", {})
|
||||
new_labels = envoy_filter_spec["spec"]["workloadSelector"]["labels"]
|
||||
|
||||
if existing_labels == new_labels:
|
||||
app.logger.info(f"EnvoyFilter '{name}' already exists with matching selector labels.")
|
||||
return
|
||||
else:
|
||||
# Update the existing EnvoyFilter's workloadSelector labels
|
||||
existing_envoy_filter["spec"]["workloadSelector"]["labels"] = new_labels
|
||||
api.replace_namespaced_custom_object(
|
||||
group="networking.istio.io",
|
||||
version="v1alpha3",
|
||||
namespace=namespace,
|
||||
plural="envoyfilters",
|
||||
name=name,
|
||||
body=existing_envoy_filter
|
||||
)
|
||||
app.logger.info(f"EnvoyFilter '{name}' updated successfully with new selector labels.")
|
||||
return
|
||||
|
||||
except client.exceptions.ApiException as e:
|
||||
if e.status == 404:
|
||||
# EnvoyFilter doesn't exist, proceed with creation
|
||||
api.create_namespaced_custom_object(
|
||||
group="networking.istio.io",
|
||||
version="v1alpha3",
|
||||
namespace=namespace,
|
||||
plural="envoyfilters",
|
||||
body=envoy_filter_spec
|
||||
)
|
||||
app.logger.info(f"EnvoyFilter '{name}' created successfully.")
|
||||
|
||||
def remove_envoy_filter_by_selector(namespace, selector_label_name, selector_label_value):
|
||||
api = client.CustomObjectsApi()
|
||||
try:
|
||||
# List all EnvoyFilters in the namespace
|
||||
existing_envoy_filters = api.list_namespaced_custom_object(
|
||||
group="networking.istio.io",
|
||||
version="v1alpha3",
|
||||
namespace=namespace,
|
||||
plural="envoyfilters"
|
||||
)
|
||||
|
||||
# Check if there is any EnvoyFilter with the same selector labels
|
||||
for item in existing_envoy_filters.get("items", []):
|
||||
workload_selector = item["spec"].get("workloadSelector", {}).get("labels", {})
|
||||
if workload_selector.get(selector_label_name) == selector_label_value:
|
||||
# Delete the matching EnvoyFilter
|
||||
api.delete_namespaced_custom_object(
|
||||
group="networking.istio.io",
|
||||
version="v1alpha3",
|
||||
namespace=namespace,
|
||||
plural="envoyfilters",
|
||||
name=item["metadata"]["name"],
|
||||
body=client.V1DeleteOptions()
|
||||
)
|
||||
print(f"EnvoyFilter '{item['metadata']['name']}' with matching selector labels deleted successfully.")
|
||||
return
|
||||
print("No EnvoyFilter found with the specified selector labels.")
|
||||
|
||||
except client.exceptions.ApiException as e:
|
||||
print(f"Failed to delete EnvoyFilter: {e}")
|
||||
|
||||
@app.route('/mutate', methods=['POST'])
|
||||
def mutate():
|
||||
app.logger.debug("Received request to mutate deployment.")
|
||||
|
||||
try:
|
||||
request_data = request.get_json()
|
||||
app.logger.debug("Admission Review Request: %s", json.dumps(request_data, indent=2))
|
||||
except Exception as e:
|
||||
app.logger.error("Failed to parse request JSON: %s", str(e))
|
||||
return Response(status=400)
|
||||
|
||||
# Extract the UID and the object from the request
|
||||
uid = request_data.get('request', {}).get('uid', '')
|
||||
obj = request_data.get('request', {}).get('object', {})
|
||||
namespace = request_data.get("request", {}).get("namespace")
|
||||
app.logger.debug("Extracted UID: %s", uid)
|
||||
app.logger.debug("Extracted Object: %s", json.dumps(obj, indent=2))
|
||||
|
||||
# Initialize patches
|
||||
patches = []
|
||||
|
||||
# Extract deployment annotations and spec
|
||||
annotations = obj.get('metadata', {}).get('annotations', {})
|
||||
spec = obj.get('spec', {})
|
||||
app.logger.debug("Current annotations: %s", json.dumps(annotations, indent=2))
|
||||
app.logger.debug("Deployment spec: %s", json.dumps(spec, indent=2))
|
||||
|
||||
# Check if the 'original-configuration' annotation already exists
|
||||
if 'original-configuration' not in annotations:
|
||||
app.logger.debug("Original configuration annotation not found, storing original spec.")
|
||||
# Store the original spec in an annotation as a JSON string
|
||||
original_spec_json = json.dumps(spec)
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/metadata/annotations/original-configuration",
|
||||
"value": original_spec_json
|
||||
})
|
||||
app.logger.debug("Added original-configuration annotation patch: %s", patches[-1])
|
||||
|
||||
# Extract containers and check if sidecar exists
|
||||
containers = obj.get('spec', {}).get('containers', [])
|
||||
init_containers = obj.get('spec', {}).get('initContainers', [])
|
||||
volumes = obj.get('spec', {}).get('volumes', [])
|
||||
app.logger.debug("Current containers in the pod: %s", json.dumps(containers, indent=2))
|
||||
sidecar_exists = any(container['name'] == 'infinity-next-nano-agent' for container in containers)
|
||||
init_container_exist = any(init_container['name'] == 'prepare-attachment' for init_container in init_containers)
|
||||
volume_exist = any(volume['name'] == 'envoy-attachment-shared' for volume in volumes)
|
||||
app.logger.debug("Does sidecar 'infinity-next-nano-agent' exist? %s", sidecar_exists)
|
||||
|
||||
# Determine if we should remove the injected data
|
||||
REMOVE_WAF = os.getenv('REMOVE_INJECTED_DATA', 'false').lower() == 'true'
|
||||
DEPLOY_FILTER = os.getenv('DEPLOY_ENVOY_FILTER', 'false').lower() == 'true'
|
||||
|
||||
ISTIO_CONTAINER_NAME = os.getenv('ISTIO_CONTAINER_NAME', 'istio-proxy')
|
||||
LIBRARY_PATH_VALUE = os.getenv('LIBRARY_PATH_VALUE', '/usr/lib/attachment')
|
||||
SELECTOR_LABEL_NAME = os.getenv("SELECTOR_LABEL_NAME")
|
||||
SELECTOR_LABEL_VALUE = os.getenv("SELECTOR_LABEL_VALUE")
|
||||
CONCURRENCY_CALC_VALUE = os.getenv('CONCURRENCY_CALC')
|
||||
CONFIG_PORT_VALUE = os.getenv('CONFIG_PORT')
|
||||
CONCURRENCY_NUMBER_VALUE = os.getenv('CONCURRENCY_NUMBER')
|
||||
if REMOVE_WAF:
|
||||
if DEPLOY_FILTER and SELECTOR_LABEL_NAME and SELECTOR_LABEL_VALUE:
|
||||
remove_envoy_filter_by_selector(namespace, SELECTOR_LABEL_NAME, SELECTOR_LABEL_VALUE)
|
||||
|
||||
app.logger.debug("Removing injected sidecar and associated resources.")
|
||||
|
||||
# Remove ld library path env variable
|
||||
if ISTIO_CONTAINER_NAME:
|
||||
if CONCURRENCY_NUMBER_VALUE:
|
||||
remove_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONCURRENCY_NUMBER', patches)
|
||||
if CONFIG_PORT_VALUE:
|
||||
remove_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONFIG_PORT', patches)
|
||||
if CONCURRENCY_CALC_VALUE:
|
||||
remove_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONCURRENCY_CALC', patches)
|
||||
if LIBRARY_PATH_VALUE:
|
||||
remove_env_variable(containers, ISTIO_CONTAINER_NAME, 'LD_LIBRARY_PATH', patches)
|
||||
|
||||
if 'shareProcessNamespace' in obj.get('spec', {}):
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": "/spec/shareProcessNamespace"
|
||||
})
|
||||
app.logger.debug("Removed shareProcessNamespace patch")
|
||||
else:
|
||||
app.logger.debug("shareProcessNamespace not found; no patch to remove it")
|
||||
|
||||
# Remove the init container if it exists
|
||||
if init_container_exist:
|
||||
for idx, init_container in enumerate(init_containers):
|
||||
if init_container['name'] == 'prepare-attachment':
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": f"/spec/initContainers/{idx}"
|
||||
})
|
||||
app.logger.debug(f"Removed init container patch: {patches[-1]}")
|
||||
break # Stop once we find and remove the target container
|
||||
|
||||
# Remove the sidecar container if it exists
|
||||
if sidecar_exists:
|
||||
for idx, container in enumerate(containers):
|
||||
volume_mounts = container.get('volumeMounts', [])
|
||||
for idx_v, volume_mount in enumerate(volume_mounts):
|
||||
if volume_mount['name'] == 'envoy-attachment-shared':
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": f"/spec/containers/{idx}/volumeMounts/{idx_v}"
|
||||
})
|
||||
app.logger.debug(f"Removed volumeMount: {patches[-1]}")
|
||||
if container['name'] == 'infinity-next-nano-agent':
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": f"/spec/containers/{idx}"
|
||||
})
|
||||
app.logger.debug(f"Removed sidecar container patch: {patches[-1]}")
|
||||
|
||||
# Remove the volume if it exists
|
||||
if volume_exist:
|
||||
for idx, volume in enumerate(volumes):
|
||||
if volume['name'] == 'envoy-attachment-shared':
|
||||
patches.append({
|
||||
"op": "remove",
|
||||
"path": f"/spec/volumes/{idx}"
|
||||
})
|
||||
app.logger.debug(f"Removed volume patch: {patches[-1]}")
|
||||
break # Stop once we find and remove the target container
|
||||
|
||||
else:
|
||||
app.logger.debug("Before if: Sidecar 'infinity-next-nano-agent' does not exist. Preparing to add it.")
|
||||
|
||||
# Define the sidecar container
|
||||
sidecar = get_sidecar_container()
|
||||
|
||||
# Define the init container()
|
||||
init_container = get_init_container()
|
||||
|
||||
# Define the volume mount for istio-proxy
|
||||
volume_mount = get_volume_mount()
|
||||
|
||||
# Define the volume
|
||||
volume_def = get_volume_definition()
|
||||
|
||||
if ISTIO_CONTAINER_NAME:
|
||||
add_env_if_not_exist(containers, ISTIO_CONTAINER_NAME, patches)
|
||||
add_env_variable_value_from(containers, ISTIO_CONTAINER_NAME, 'CLOUDGUARD_UID', None, patches, value_from={"fieldRef": {"fieldPath": "metadata.uid"}})
|
||||
if LIBRARY_PATH_VALUE:
|
||||
add_env_variable(containers, ISTIO_CONTAINER_NAME, 'LD_LIBRARY_PATH', LIBRARY_PATH_VALUE, patches)
|
||||
if CONCURRENCY_CALC_VALUE:
|
||||
add_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONCURRENCY_CALC', CONCURRENCY_CALC_VALUE, patches)
|
||||
if CONFIG_PORT_VALUE:
|
||||
add_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONFIG_PORT', CONFIG_PORT_VALUE, patches)
|
||||
if CONCURRENCY_NUMBER_VALUE:
|
||||
add_env_variable(containers, ISTIO_CONTAINER_NAME, 'CONCURRENCY_NUMBER', CONCURRENCY_NUMBER_VALUE, patches)
|
||||
else:
|
||||
app.logger.debug("ISTIO_CONTAINER_NAME skipping environment variable addition")
|
||||
|
||||
# Add the sidecar container
|
||||
if not sidecar_exists:
|
||||
|
||||
# Add shareProcessNamespace if not already set
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/spec/shareProcessNamespace",
|
||||
"value": True
|
||||
})
|
||||
app.logger.debug("Added shareProcessNamespace patch")
|
||||
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/spec/containers/-",
|
||||
"value": sidecar
|
||||
})
|
||||
app.logger.debug("Added sidecar container patch: %s", patches[-1])
|
||||
|
||||
# Add the volume mount to istio-proxy container (assumes istio-proxy is first container)
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/spec/containers/0/volumeMounts/-",
|
||||
"value": volume_mount
|
||||
})
|
||||
app.logger.debug("Added volume mount patch to istio-proxy: %s", patches[-1])
|
||||
|
||||
# Add the new volume definition
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/spec/volumes/-",
|
||||
"value": volume_def
|
||||
})
|
||||
app.logger.debug("Added volume definition patch: %s", patches[-1])
|
||||
|
||||
if DEPLOY_FILTER and SELECTOR_LABEL_NAME and SELECTOR_LABEL_VALUE:
|
||||
RELEASE_NAME = os.getenv('RELEASE_NAME', 'cloudguard-waf-injected')
|
||||
envoy_filter_name = RELEASE_NAME + "-waf-filter"
|
||||
create_or_update_envoy_filter(envoy_filter_name, namespace, SELECTOR_LABEL_NAME, SELECTOR_LABEL_VALUE)
|
||||
else:
|
||||
app.logger.debug("Before else: Sidecar 'infinity-next-nano-agent' already exists. Checking for image updates.")
|
||||
|
||||
# Optionally, update the sidecar image and tag if necessary
|
||||
for idx, container in enumerate(containers):
|
||||
if container['name'] == 'infinity-next-nano-agent':
|
||||
current_image = container.get('image', '')
|
||||
app.logger.debug("Current sidecar image: %s", current_image)
|
||||
app.logger.debug("Desired sidecar image: %s", FULL_AGENT_IMAGE)
|
||||
if current_image != FULL_AGENT_IMAGE:
|
||||
patches.append({
|
||||
"op": "replace",
|
||||
"path": f"/spec/containers/{idx}/image",
|
||||
"value": FULL_AGENT_IMAGE
|
||||
})
|
||||
app.logger.debug(f"Updated sidecar image patch: {patches[-1]}")
|
||||
break # Sidecar found and handled
|
||||
|
||||
if not init_container_exist:
|
||||
# Add the initContainer to the pod spec in the deployment
|
||||
if 'initContainers' in obj['spec']:
|
||||
obj['spec']['initContainers'].append(init_container)
|
||||
else:
|
||||
obj['spec']['initContainers'] = [init_container]
|
||||
|
||||
patches.append({
|
||||
"op": "add",
|
||||
"path": "/spec/initContainers",
|
||||
"value": obj['spec']['initContainers']
|
||||
})
|
||||
else:
|
||||
app.logger.debug("Before else: init-container 'prepare-attachment' already exists. Checking for image updates.")
|
||||
|
||||
# Optionally, update the sidecar image and tag if necessary
|
||||
for idx, container in enumerate(containers):
|
||||
if container['name'] == 'prepare-attachment':
|
||||
current_image = container.get('image', '')
|
||||
app.logger.debug("Current init container image: %s", current_image)
|
||||
app.logger.debug("Desired init container image: %s", FULL_AGENT_IMAGE)
|
||||
if current_image != FULL_AGENT_IMAGE:
|
||||
patches.append({
|
||||
"op": "replace",
|
||||
"path": f"/spec/containers/{idx}/image",
|
||||
"value": FULL_AGENT_IMAGE
|
||||
})
|
||||
app.logger.debug(f"Updated sidecar image patch: {patches[-1]}")
|
||||
break # Sidecar found and handled
|
||||
|
||||
|
||||
app.logger.info("Total patches: %s", json.dumps(patches, indent=2))
|
||||
|
||||
# Prepare the AdmissionReview response
|
||||
admission_response = {
|
||||
"kind": "AdmissionReview",
|
||||
"apiVersion": "admission.k8s.io/v1",
|
||||
"response": {
|
||||
"uid": uid,
|
||||
"allowed": True,
|
||||
"patchType": "JSONPatch",
|
||||
"patch": base64.b64encode(json.dumps(patches).encode('utf-8')).decode('utf-8')
|
||||
}
|
||||
}
|
||||
app.logger.debug("Sending admission response: %s", json.dumps(admission_response, indent=2))
|
||||
|
||||
return jsonify(admission_response)
|
||||
|
||||
if __name__ == '__main__':
|
||||
# Configure logger
|
||||
configure_logging()
|
||||
|
||||
# Ensure certificates exist
|
||||
secretgen.main()
|
||||
if not os.path.exists("/certs/server.crt") or not os.path.exists("/certs/server.key"):
|
||||
print("Error: Certificates not found. Exiting...")
|
||||
exit(1)
|
||||
|
||||
cli = sys.modules['flask.cli']
|
||||
cli.show_server_banner = lambda *x: None
|
||||
|
||||
# Run the Flask app with the generated certificates
|
||||
app.run(host='0.0.0.0', port=443, ssl_context=('/certs/server.crt', '/certs/server.key'))
|
||||
|
Loading…
x
Reference in New Issue
Block a user