Proactively Detect and Resolve Istio Ambient Issues

Istio

As applications grow into distributed systems, secure communication, traffic routing, and observability become just as critical as the business logic itself. That’s where a service mesh comes in. 

Istio is one of the most widely used service meshes in Kubernetes environments, providing a transparent way to handle traffic policies, encryption, and telemetry. 

We’ve seen teams spend days troubleshooting issues that they wish they were able to proactively detect and resolve. The Prequel community has responded by contributing common reliability enumerations (CREs) for Istio failures.  These CRE can be executed by preq (100% open source) or Prequel to discover problems before they turn into incidents.   

As teams cautiously make the shift to Istio’s Ambient mode, we’ve seen intense focus around related problems. Community members have been turning troubleshooting tips into proactive detections. 

Ambient Mode

Traditionally, Istio integrated with your application by injecting sidecar proxies that intercept all inbound and outbound traffic, alongside every pod. While powerful, this sidecar model increases resource usage and operational overhead. 

To reduce that, Istio introduced ambient mode, which moves traffic interception to node-level components, ztunnels, and offloads policy enforcement to centralized waypoints. This new approach simplifies the data plane and lowers costs, but it also brings new classes of failure that teams need to watch for to maintain reliability.Understanding how ambient mesh can fail—and detecting those failures early—is key to reliable adoption of Istio. This post examines several common Istio ambient mode issues, how to detect them proactively, and how to resolve them (before they spiral into incidents).  

We draw on real-world patterns from Istio’s guidance, code, and Prequel’s Reliability Research, showing how CREs can help you automatically pinpoint these issues.

If you want to proactively detect these issues, preq (100% open source) or Prequel are automatically updated with the latest community CREs.

CRE ID Detection
CRE‑2025‑0106 Ambient CNI sandbox creation failures (“FailedCreatePodSandBox” + “no ztunnel connection”)
CRE‑2025‑0111 Ztunnel IPv6 bind failures when binding to [::1]:15053 (“Address family not supported”)
CRE‑2025‑0104 XDS gRPC timeouts waiting for workload metadata (“timed out waiting for workload … from xds”)
CRE‑2025‑0110 HBONE tunnel traffic timeouts on TCP port 15008 (“io error: deadline has elapsed” / “connection timed out…HBONE port 15008”)
CRE‑2025‑0109 Non‑2xx HTTP status codes in Ztunnel access logs (status=4xx)
CRE‑2025‑0108 Ambient mode readiness probe failures due to missing SNAT bypass (Readiness probe failed: + Back-off restarting failed container)

CRE‑2025‑0106 — Ambient  CNI  Sandbox  Creation  Failure

Description

When Istio is running in ambient mode, the CNI plugin must create a specialized network “sandbox” for each pod so that its traffic can be redirected through the node‑level ztunnel proxy. If the CNI plugin cannot establish this sandbox—often because it can’t reach the ztunnel socket or the ztunnel daemon isn’t running on the node—pods stall in the ContainerCreating or Pending state. As a result, those workloads never join the mesh, leaving them both unreachable and invisible to Istio’s mTLS, policy enforcement, and telemetry. This failure mode typically manifests as Kubernetes events with FailedCreatePodSandBox or CNI log entries reporting “no ztunnel connection,” and it causes immediate service downtime until the sandbox issue is resolved.

Detection

CRE-2025-0106 detects situations where the Istio CNI plugin cannot finish creating the pod’s network sandbox while the cluster is running ambient mode.

id: CRE-2025-0106
title: "Ambient CNI Sandbox Creation Failure"

rule:
  set:
    window: 60s
  event:
    source: cre.log.ambient
  match:
    - regex: "FailedCreatePodSandBox"
    - regex: "no ztunnel connection"

When you run this CRE with a supported problem detector (like preq or Prequel), events from the cre.log.ambient datasource will be watched in 60‑second windows and the rule will fire when it sees both a sandbox failure and a failed ztunnel connection, which the Istio CNI emits when it cannot reach the node‑local ztunnel. 

Resolution

Beyond the detection rule, CRE‑2025‑0106 also describes common root causes and appropriate mitigation steps. 

Common causes of an Ambient CNI sandbox creation failure include the ztunnel DaemonSet being unscheduled or unhealthy on a node and the CNI plugin’s inability to dial the ztunnel socket (/var/run/istio-cni/cni.sock). When either of these conditions occurs, pods remain stuck in the ContainerCreating or Pending state, preventing workloads from ever joining the mesh and resulting in immediate application downtime.

To resolve this issue, you should first verify the health of the ztunnel DaemonSet by running:

kubectl -n istio-system get pods -l app=ztunnel

If any ztunnel pods are missing or unhealthy, inspect their logs for connectivity errors:

kubectl -n istio-system logs <ztunnel-pod>

Once you’ve identified and addressed the root cause—such as by patching the DaemonSet to ensure it schedules on all nodes:

kubectl -n istio-system patch daemonset ztunnel --patch '{"spec":{"template":{"spec":{"nodeSelector":{}}}}}'

—you can recover by deleting and recreating the affected pods to force them through a fresh sandbox creation.

cause: |
        - ztunnel DaemonSet is unscheduled or unhealthy on the node
        - CNI plugin cannot dial the ztunnel Unix socket `/var/run/istio-cni/cni.sock`
      impact: |
        - Pods stuck in ContainerCreating or Pending
        - Workloads cannot start, leading to application downtime
      mitigation: |
        IMMEDIATE:
        - Verify ztunnel DaemonSet health: `kubectl -n istio-system get pods -l app=ztunnel`
        - Inspect ztunnel logs for connectivity errors: 
          `kubectl -n istio-system logs <ztunnel-pod>`
        RECOVERY:
        - Patch ztunnel to run on all nodes:
          `kubectl -n istio-system patch daemonset ztunnel --patch '{"spec":{"template":{"spec":{"nodeSelector":{}}}}}'`
        - Delete and recreate the affected pods
        PREVENTION:
        - Monitor Istio CNI and ztunnel agent metrics
        - Alert on repeated CNI plugin failures in control-plane logs

To prevent future occurrences, integrate monitoring tools like preq or prequel to implement alerting around both the Istio CNI plugin and the ztunnel agent. In particular, log patterns for CNI plugin failures and ztunnel connectivity errors, and configure alerts on repeated sandbox creation failures so you’re aware of issues before they cause service disruption

CRE‑2025‑0111 — Ztunnel IPv6 Bind Failure

When Istio is running in ambient mode, Ztunnel’s DNS proxy (and related control‑plane components) must bind to the IPv6 loopback address ([::1]:15053) on each node so it can intercept and resolve service names. If the node’s kernel has IPv6 support disabled—or if the IstioOperator’s IPV6_ENABLED flag isn’t explicitly set to false—the bind will fail with an error like

failed to bind to address [::1]:15053: Address family not supported

When this happens, the DNS proxy never starts, ambient DNS‑capture is disabled, and mesh traffic can’t resolve or route service names correctly, resulting in service‐to‐service connectivity failures within the mesh.

Detection

CRE‑2025‑0111 watches the cre.log.ambient datasource for Ztunnel bind errors. As soon as it sees a log entry matching the IPv6 bind failure pattern, it fires an alert.

id: CRE-2025-0111
title: "Ztunnel IPv6 Bind Failure"
rule:
  event:
    source: cre.log.ambient
  match:
    - regex: 'failed to bind to address \[::1\]:15053: Address family not supported(?: \(os error \d+\))?'

When executed by a supported problem detector (like preq or Prequel), this rule continuously scans incoming ambient‑mode logs and will alert immediately upon detecting the IPv6 bind failure.

Resolution

Beyond the detection rule, CRE‑2025‑0111 also describes common root causes and appropriate mitigation steps. 

Common causes of this Ztunnel IPv6 bind failure are either that the node’s kernel has IPv6 entirely disabled or that you haven’t explicitly disabled IPv6 binding in Ztunnel’s configuration. In both cases, the DNS proxy can’t start, which disables DNS‑capture and causes ambient traffic to be unable to resolve or misroute service names—resulting in mesh connectivity issues.

To resolve the issue, choose one of two immediate fixes: re‑enable IPv6 on the node (so the proxy can bind to [::1]:15053) or explicitly disable IPv6 binding in Ztunnel by setting IPV6_ENABLED: false in your IstioOperator. After applying your chosen fix, restart the Ztunnel DaemonSet (kubectl -n istio-system rollout restart daemonset ztunnel) to pick up the change. For long‑term stability, bake IPv6 support validation and IPV6_ENABLED checks into your cluster provisioning and CI/CD pipelines so that misconfigurations are caught before they reach production

cause: |
  - Kernel‑level IPv6 support is disabled on the node
  - `IPV6_ENABLED` not set to `false` in the IstioOperator values
impact: |
  - Ztunnel’s DNS proxy cannot start, disabling DNS‑capture
  - Ambient traffic cannot resolve or route service names, leading to connectivity failures
mitigation: |
  IMMEDIATE:
  - Re‑enable IPv6 on the node if IPv6 is required:
    ```bash
    sudo sysctl -w net.ipv6.conf.all.disable_ipv6=0
    sudo sysctl -w net.ipv6.conf.default.disable_ipv6=0
    ```
  - Or disable IPv6 binding in Ztunnel by updating your IstioOperator:
    ```yaml
    values:
      ztunnel:
        IPV6_ENABLED: false
    ```
  - Restart the Ztunnel DaemonSet to apply changes:
    ```bash
    kubectl -n istio-system rollout restart daemonset ztunnel
    ```
  LONG‑TERM:
  - Incorporate IPv6 capability checks into cluster provisioning workflows
  - Validate the `IPV6_ENABLED` flag in your CI/CD pipelines for IstioOperator manifests

CRE ‑ 2025 ‑ 0104 — Istio Ambient Traffic Fails 

Description

When Istio is running in ambient mode, Ztunnel must fetch pod workload metadata from Istiod over the XDS gRPC API before it can proxy traffic. If Ztunnel doesn’t receive the workload response—typically within about 5 seconds—it rejects the connection with an error like:

timed out waiting for workload <pod-uid> from xds

This usually indicates that Istiod is overloaded or misconfigured (for example, due to an increased PILOT_DEBOUNCE_AFTER setting), or that network policies or CNI configurations are blocking TCP port 15012 between Ztunnel and the control plane.

Detection

CRE‑2025‑0104 watches the cre.log.ambient datasource for XDS timeout errors from Ztunnel and fires when it sees a matching log entry without a subsequent “connection complete” event. The ability to tighten up rules by adding negative conditions is an important feature of CREs. 

id: CRE‑20250104
title: "Istio Ambient traffic fails with timed out waiting for workload from xds"
rule:
  set:
    window: 60s
  event:
    source: cre.log.ambient
  match:
    - regex: "timed out waiting for workload .* from xds"
  negate:
    - regex: "connection complete"

When executed by a supported detector (such as preq or Prequel), this rule evaluates ambient‑mode logs in rolling 60‑second windows and alerts immediately upon detecting an XDS timeout without a successful connection follow‑up.

Resolution

Beyond the detection rule, CRE‑2025‑0104 also describes common root causes and appropriate mitigation steps. 

Common causes of this failure mode include an overloaded or misconfigured Istiod control plane—leading to slow XDS responses—and network restrictions blocking port 15012 traffic between Ztunnel and Istiod. When Ztunnel cannot fetch workload details in time, it rejects ambient traffic with an XDS timeout error, resulting in intermittent request drops and potential circuit‑breaker trips. 

To resolve the issue, begin by examining Istiod’s resource utilization and Ztunnel logs, then ensure that port 15012/TCP is reachable across nodes. Recover by scaling or tuning Istiod and rolling your meshConfig changes back if needed. For prevention, bake XDS‑latency monitoring and alerting into your observability stack so you catch control‑plane delays before they impact workloads.

cause: |
  - Istiod is under heavy CPU or memory pressure and slow to respond
  - Network policies or the CNI plugin are blocking TCP port 15012 between Ztunnel and Istiod
  - `PILOT_DEBOUNCE_AFTER` or other Envoy debouncing settings have been increased
impact: |
  - Ztunnel connections intermittently fail with XDS timeout errors
  - Applications may experience dropped requests or circuit‑breaker activations
mitigation: |
  IMMEDIATE:
  - Check Istiod resource usage:
    ```bash
    kubectl -n istio-system top pods -l app=istiod
    ```
  - Inspect Ztunnel logs for repeated XDS timeouts:
    ```bash
    kubectl -n istio-system logs <ztunnel-pod> | grep "timed out waiting for workload"
    ```
  - Verify port 15012/TCP is open between all nodes and Istiod:
    ```bash
    kubectl -n istio-system exec <ztunnel-pod> -- nc -vz istiod-0.istio-system.svc.cluster.local 15012
    ```
  RECOVERY:
  - Scale up the Istiod deployment or increase CPU/memory requests and limits
  - Review and revert any experimental `meshConfig.PILOT_DEBOUNCE_*` changes
PREVENTION:
  - Monitor Istiod latency and error rates in your control‑plane metrics
  - Alert on repeated XDS timeout patterns in `cre.log.ambient`

CRE ‑ 2025 ‑ 0110 — Ztunnel Traffic Timeouts 

Description

When Istio is running in ambient mode, Ztunnel proxies pod‑to‑pod traffic over the HBONE (mTLS) tunnel on TCP port 15008. If network policies, firewalls, or host‑level iptables block or drop this port, Ztunnel will log timeouts such as

io error: deadline has elapsed  

connection timed out, maybe a NetworkPolicy is blocking HBONE port 15008: deadline has elapsed  

When these timeouts occur, ambient traffic hangs until the TCP deadline passes—causing silent failures or long delays and degrading service reliability as calls between workloads never complete.

Detection


CRE‑2025‑0110 watches the cre.log.ambient datasource for HBONE timeout errors in rolling three‑minute windows. When it sees both timeout patterns, it fires an alert.

id: CRE‑20250110
title: "Ztunnel Traffic timeouts in Istio Ambient Mode"
rule:
  set:
    window: 180s
  event:
    source: cre.log.ambient
  match:
    - regex: 'error +access +connection +complete +.*error="io error: deadline has elapsed"'
    - regex: 'error +access +connection +complete +.*error="connection timed out, maybe a NetworkPolicy is blocking HBONE port 15008: deadline has elapsed"'

When you run this CRE with a supported problem detector (like preq or Prequel), it will continuously scan ambient‑mode logs and alert immediately upon detecting blocked or dropped HBONE traffic.

Resolution

Beyond the detection rule, CRE‑2025‑0110 also describes common root causes and appropriate mitigation steps.

Common causes of HBONE traffic timeouts include overly restrictive NetworkPolicies or firewall rules blocking egress on port 15008, misconfigured CNI or iptables rules that prevent HBONE encapsulation, and node‑level network segmentation or security groups dropping tunnel traffic. When these issues occur, Ztunnel’s HBONE connections hang, leading to failed or delayed inter‑pod requests and reduced mesh reliability.

To resolve the issue immediately, inspect and allow egress on port 15008 in your NetworkPolicies or firewall configurations, verify that host‑level iptables aren’t dropping HBONE packets, and from a Ztunnel pod test connectivity with:

kubectl exec -n istio-system <ztunnel-pod> -- nc -vz istio-system 15008

Once connectivity is restored, ambient traffic will resume normally. For long‑term stability, enforce HBONE port rules in your CI/CD pipelines and monitor Ztunnel logs for frequent timeouts—configuring alerts on repeated occurrences so you catch emerging network segmentation issues before they impact workloads.

cause: |
  - NetworkPolicy or firewall rules blocking outbound port 15008
  - Misconfigured CNI or host-level iptables preventing HBONE tunnels
  - Node-level network segmentation or security groups dropping traffic
impact: |
  - Pod-to-pod calls via ambient mode hang until timeout
  - Application requests fail silently or with long delays
  - Service reliability degraded due to hidden networking issues
mitigation: |
  IMMEDIATE:
  - Inspect and allow egress on port 15008 in NetworkPolicies or firewalls
  - Verify host iptables aren’t dropping HBONE traffic
  - From a ztunnel pod, test connectivity:
    ```bash
    kubectl exec -n istio-system <ztunnel-pod> -- nc -vz istio-system 15008
    ```
  LONG-TERM:
  - Enforce HBONE port rules in CI/CD pipelines for IstioOperator manifests
  - Monitor ztunnel logs for frequent timeouts and alert on repeated patterns
references:
  - https://github.com/istio/istio/wiki/Troubleshooting-Istio-Ambient#scenario-traffic-timeout-with-ztunnel

CRE ‑ 2025 ‑ 0109 ‑ Ambient HTTP Status Codes by Ztunnel

Description

When Istio is running in ambient mode, Ztunnel tunnels HTTP over HBONE (HTTP CONNECT) as a TCP proxy. Even though it operates at the TCP level, Ztunnel tags each “connection complete” log line with the upstream HTTP status code (for example, 503 or 401). This CRE verifies that any non‑2xx response is correctly surfaced in Ztunnel’s logs so that operators can trace HTTP failures through the mesh. 

Detection

CRE‑2025‑0109 watches the cre.log.ambient datasource for access‑log entries where the status code indicates a client or server error. It fires immediately on seeing any log line matching a 4xx or 5xx status tag.

id: CRE‑20250109
title: "Ambient HTTP status codes by Ztunnel"
rule:
  event:
    source: cre.log.ambient
  match:
    - regex: "status=(?:4[0-9]{2}|5[0-9]{2})"
  negate: []

When run via a supported detector (like preq or Prequel), this rule continuously scans ambient‑mode logs and alerts on the very first occurrence of a non‑2xx HTTP status in Ztunnel’s access logs.

Resolution

Beyond detecting the missing or incorrect status code, CRE‑2025‑0109 outlines root causes and remediation steps to ensure HTTP error visibility in your mesh:

Common causes include:

  • Your application returning an HTTP error (e.g., 503 Service Unavailable, 401 Unauthorized).

  • Ztunnel not capturing the status code because it isn’t configured to log the status field.

When Ztunnel fails to log HTTP status codes, end‑to‑end troubleshooting becomes significantly harder—operators cannot easily correlate client‑side failures with application‑side errors.

To remediate immediately:

  1. Inspect your HTTP service logs to confirm that the upstream application is generating 4xx/5xx responses.
  2. Tail Ztunnel’s access logs and verify the status= field is present:

    kubectl -n istio-system logs -c istio-proxy <ztunnel-pod> | grep 'status='

For recovery and prevention, ensure that your Istio meshConfig proxy stats matcher includes the status field so that Ztunnel always records HTTP status codes:

cause: |
  - The application returned an HTTP error (4xx or 5xx)
  - Ztunnel isn’t logging the `status` field in its access logs
impact: |
  - Operators cannot trace HTTP failures through the mesh
  - Troubleshooting is more time‑consuming without status visibility
mitigation: |
  IMMEDIATE:
  - Inspect HTTP service logs for error responses
  - Verify Ztunnel logs contain `status=` fields:
    ```bash
    kubectl -n istio-system logs -c istio-proxy <ztunnel-pod> | grep 'status='
    ```
  RECOVERY:
  - Update your meshConfig.defaultConfig.proxyStatsMatcher.inclusionRegexps to include `status`
  - Roll out the updated IstioOperator configuration
PREVENTION:
  - Add a CI/CD validation step to enforce `status` inclusion in proxyStatsMatcher
  - Monitor Ztunnel logs for absence of `status=` patterns and alert on such events
references:
  - https://github.com/istio/istio/wiki/Troubleshooting-Istio-Ambient#scenario-traffic-fails-with-http-status- :contentReference[oaicite:2]{index=2} 

CRE ‑ 2025 ‑ 0108 ‑ Ambient Mode Readiness Probe Failures

Description

When Istio is running in ambient mode, it applies a host‑level SNAT rule so that kubelet’s readiness probe traffic appears to originate from the IP 169.254.7.127, allowing the data‑plane bypass rules to recognize and intercept those probes correctly. If this SNAT/bypass isn’t configured or is overridden by your CNI or networking stack, you’ll begin to see Kubernetes events like:

Readiness probe failed

Back-off restarting failed container

only after enabling ambient mode. In this scenario, pods never report as Ready and remain stuck in CrashLoopBackOff or Pending, causing service outages.

Detection

CRE‑2025‑0108 watches the cre.log.ambient datasource for repeated readiness probe failures in ambient mode. It fires when it sees at least four occurrences of probe failures followed by container back‑off within a rolling five‑minute window, indicating that the SNAT bypass is not working.

id: CRE‑20250108
title: "Ambient mode readiness probe failures"
rule:
  set:
    window: 300s
  event:
    source: cre.log.ambient
  match:
    - regex: "Readiness probe failed:"
      count: 4
    - regex: "Back-off restarting failed container"

When run with a detector like preq or Prequel, this rule continuously evaluates ambient‑mode logs and will alert as soon as readiness probes start failing repeatedly.

Resolution

Beyond detecting the problem, CRE‑2025‑0108 outlines the usual root causes and corrective actions.

Common causes of ambient mode readiness probe failures include:

  • The host SNAT rule for kubelet traffic (169.254.7.127) is missing, removed, or overwritten.

  • A CNI plugin (for example, Cilium with bpf.masquerade=true or Calico versions < 3.29 with BPF mode) is masquerading kubelet probe traffic unexpectedly.

  • A NetworkPolicy or security group is blocking probe ports (typically TCP 15021 or 8080).

When SNAT/bypass isn’t in place, kubelet probes never reach the application container correctly, and pods stay unready—causing downstream services to see them as unavailable.

To resolve the issue immediately:

  1. Inspect the host NAT table for the ISTIO redirect chain and SNAT rules:

iptables -t nat -L ISTIO_REDIRECT

iptables -t nat -L OUTPUT

  1. Confirm that packets from source 169.254.7.127 are being SNAT’d and bypassed by the data‑plane.

  2. Review your CNI’s ambient prerequisites (e.g., Cilium/Calico docs) and disable any masquerade settings that conflict with kubelet SNAT.

  3. Remove or adjust any NetworkPolicies or security groups that block TCP 15021/8080 on readiness probe traffic.

For long‑term prevention, ensure your cluster provisioning and CNI configurations include the necessary SNAT bypass rules for ambient mode, and add validation checks in your CI/CD pipelines to catch overrides before they reach production.

cause: |
  - Host SNAT rule for kubelet (169.254.7.127) is missing or overwritten
  - CNI plugin is masquerading kubelet probe traffic (e.g., Cilium bpf.masquerade, Calico BPF)
  - NetworkPolicy or security group blocks probe ports (15021/8080)
impact: |
  - Pods never become Ready (stuck in CrashLoopBackOff or Pending)
  - Services relying on readiness probes see pods as unavailable, causing outages
mitigation: |
  IMMEDIATE:
  - Inspect host iptables for ISTIO_REDIRECT and OUTPUT SNAT chains
  - Confirm SNAT of kubelet traffic from 169.254.7.127
  - Review and adjust CNI masquerade settings to preserve SNAT bypass
  - Remove conflicting NetworkPolicies blocking readiness ports
RECOVERY:
  - Redeploy affected pods after fixing SNAT/bypass
PREVENTION:
  - Bake ambient SNAT bypass configuration into cluster provisioning
  - Validate SNAT bypass rules in CI/CD for IstioOperator and CNI manifests
references:
  - https://github.com/istio/istio/wiki/Troubleshooting-Istio-Ambient#scenario-readiness-probes-fail-with-ztunnel :contentReference[oaicite:0]{index=0}

Summary

Common Reliability Enumerations (CREs) empower SRE and platform teams to codify deep, domain‑specific knowledge about failure modes into shareable, versioned rules that drive consistent detection and response across environments. By capturing both positive and negative conditions, CREs address both detection blind spots and noisy false positives, enabling instant, high‑fidelity insights into what’s really happening. 

The CREs referenced in this post were created by community members to specifically catch these Istio ambient issues before they escalate, turning hours of incident debugging into quick, automated detection.

If you want to proactively detect these issues, preq (100% open source) or Prequel are automatically updated with the latest community CREs.

CRE ID Detection
CRE‑2025‑0106 Ambient CNI sandbox creation failures (“FailedCreatePodSandBox” + “no ztunnel connection”)
CRE‑2025‑0111 Ztunnel IPv6 bind failures when binding to [::1]:15053 (“Address family not supported”)
CRE‑2025‑0104 XDS gRPC timeouts waiting for workload metadata (“timed out waiting for workload … from xds”)
CRE‑2025‑0110 HBONE tunnel traffic timeouts on TCP port 15008 (“io error: deadline has elapsed” / “connection timed out…HBONE port 15008”)
CRE‑2025‑0109 Non‑2xx HTTP status codes in Ztunnel access logs (status=4xx)
CRE‑2025‑0108 Ambient mode readiness probe failures due to missing SNAT bypass (Readiness probe failed: + Back-off restarting failed container)

References

  1. https://github.com/istio/istio/wiki/Troubleshooting-Istio-Ambient
  2. https://github.com/prequel-dev/cre

By clicking Sign Up you're confirming that you agree with our Terms and Conditions.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Break the cycle.

Fix Sev 1s when they're sev 5s.

See how Prequel helps engineering teams
Learn how Prequel helps companies like yours achieving their reliability goals
Get a 1-on-1 walkthrough of the platform
Ask us anything
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
“It finally feels like someone is watching my back.”
Facundo O.
Head of Infrastructure, Leading Cyber Security Company
Sign up for a free trial
Give Prequel a spin. See for yourself
how we can help you level up reliability.
Start your 30-day trial today. Use Prequel for a month on us.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
“It finally feels like someone is watching my back.”
Facundo Osimi
Head of Infrastructure, Leading Cyber Security Company