Ingress Monitor Controller
An operator to watch ingresses/routes and create liveness alerts for your apps/microservices in Uptime checkers.
Problem Statement
We want to monitor ingresses in a kubernetes cluster and routes in openshift cluster via any uptime checker but the problem is having to manually check for new ingresses or routes / removed ingresses or routes and add them to the checker or remove them.
Solution
This operator will continuously watch ingresses/routes based on defined EndpointMonitor
custom resource, and
automatically add / remove monitors in any of the uptime checkers. With the help of this solution, you can keep a check
on your services and see whether they're up and running and live, without worrying about manually registering them on
the Uptime checker.
Supported Uptime Checkers
Currently we support the following monitors:
- UptimeRobot (Additional Config)
- Pingdom (Additional Config) (Not fully tested)
- StatusCake (Additional Config)
- Uptime (Additional Config)
- Updown (Additional Config)
- Application Insights (Additional Config)
- gcloud (Additional Config)
Usage
Adding configuration
Configure the uptime checker configuration in the config.yaml
based on your uptime provider. Add create a secret
imc-config
that holds config.yaml
key:
kind: Secret
apiVersion: v1
metadata:
name: imc-config
data:
config.yaml: >-
<BASE64_ENCODED_CONFIG.YAML>
type: Opaque
Configuration Parameters
Following are the available options that you can use to customize the controller:
Key | Description |
---|---|
providers | An array of uptime providers that you want to add to your controller |
enableMonitorDeletion | A safeguard flag that is used to enable or disable monitor deletion on ingress deletion (Useful for prod environments where you don't want to remove monitor on ingress deletion) |
resyncPeriod | Resync period in seconds, allows to re-sync periodically the monitors with the Routes. Defaults to 0 (= disabled) |
creationDelay | CreationDelay is a duration string to add a delay before creating new monitor (e.g., to allow DNS to catch up first) |
monitorNameTemplate | Template for monitor name eg, {{.Namespace}}-{{.Name}} |
- Replace
BASE64_ENCODED_CONFIG.YAML
with your config.yaml file that is encoded in base64. - For detailed guide for the configuration refer to Docs and go through configuration guidelines for your uptime provider.
- For sample
config.yaml
files refer to Sample Configs. - Name of secret can be changed by setting environment variable
CONFIG_SECRET_NAME
.
Add EndpointMonitor
EndpointMonitor
resource can be used to manage monitors on static urls or route/ingress references.
- Specifying url:
apiVersion: endpointmonitor.stakater.com/v1alpha1
kind: EndpointMonitor
metadata:
name: stakater
spec:
forceHttps: true
url: https://stakater.com
- Specifying route reference:
apiVersion: endpointmonitor.stakater.com/v1alpha1
kind: EndpointMonitor
metadata:
name: frontend
spec:
forceHttps: true
urlFrom:
routeRef:
name: frontend
- Specifying ingress reference:
apiVersion: endpointmonitor.stakater.com/v1alpha1
kind: EndpointMonitor
metadata:
name: frontend
spec:
forceHttps: true
urlFrom:
ingressRef:
name: frontend
NOTE: For provider specific additional configuration refer to Docs and go through configuration guidelines for your uptime provider.
Deploying the Operator
The following quickstart let's you set up Ingress Monitor Controller to register uptime monitors for endpoints:
Helm Chart
If you have configured helm on your cluster, you can deploy IngressMonitorController via helm using below mentioned commands. For details on chart, see IMC Helm Chart
# Install CRDs
kubectl apply -f https://raw.githubusercontent.com/stakater/IngressMonitorController/master/charts/ingressmonitorcontroller/crds/endpointmonitor.stakater.com_endpointmonitors.yaml
# Install chart
helm repo add stakater https://stakater.github.io/stakater-charts
helm repo update
helm install stakater/ingressmonitorcontroller
Vanilla Manifests
- Clone this repository
$ git clone [email protected]:stakater/IngressMonitorController.git
- Deploy dependencies(crds):
$ make deploy
Environment Variables
Key | Default | Description |
---|---|---|
WATCH_NAMESPACE | Namespace in which operator is deployed | Use comma separated list of namespaces or leave the field empty to watch all namespaces(cluster scope) |
CONFIG_SECRET_NAME | imc-config | Name of secret that holds the configuration |
REQUEUE_TIME | 300 seconds | Integer value to specify number of seconds after which the resource should be reconciled again |
Help
Documentation
You can find more detailed documentation for configuration, extension, and support for other Uptime checkers etc. here
Contributing
If you'd like to contribute any fixes or enhancements, please refer to the documentation here
Have a question?
File a GitHub issue.
Talk to us on Slack
Join and talk to us on the #tools-ingressmonitor channel for discussing the Ingress Monitor Controller
Known Issues
- Latest image of kube-rbac-proxy fails on openshift with permission issues. To resolve use
registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
instead of kube-rbac-proxy. This issue can be tracked here.
License
Apache2 © Stakater
About
The IngressMonitorController
is maintained by Stakater. Like it? Please let us know at [email protected]
See our other projects or contact us in case of professional services and queries on [email protected]
Contributors
Stakater Team and the Open Source community!