You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/en/docs/Troubleshooting/subscription.md
+31Lines changed: 31 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -47,3 +47,34 @@ api-server resource not found installing CustomResourceDefinition my-crd-name: G
47
47
This error indicates that the API apiextensions.k8s.io/v1beta1, Kind=CustomResourceDefinition is not available on-cluster. It indicates this particular GVK is not present on the api-server, which will happen on Kubernetes 1.22+ with deprecated built-in types like v1beta1 CRDs or v1beta1 RBAC.
48
48
49
49
This error can also arise when installing operator bundles with CustomResources that OLM supports, such as `VerticalPodAutoscalers` and `PrometheusRules`, but the relevant CustomResourceDefinition has not yet been installed. In this case, this error should eventually resolve itself provided the required CustomResourceDefinition gets installed on the cluster and is accepted by the api-server.
50
+
51
+
### Subscriptions failing due to unpacking errors
52
+
53
+
If A subscription references an operator bundle fails to unpack successfully, the subscription will fail with the following message:
54
+
55
+
```
56
+
bundle unpacking failed. Reason: DeadlineExceeded, and Message: Job was active longer than the specified deadline
57
+
```
58
+
59
+
This type of failure can happen due to many possible underlying reasons on the cluster, including:
60
+
1. Operator bundle image being unreachable:
61
+
1. Misconfigured network, such as an incorrectly configured proxy/firewall
62
+
2. Missing operator bundle images from the reachable image registries
63
+
3. Invalid or missing image registry credentials/secrets
64
+
4. Image registry rate limits
65
+
2. Resource limitations on the cluster
66
+
1. CPU or network limitations preventing operator bundle images from being pulled within the timeout (10 minutes)
67
+
2. Inability to schedule pods for unpacking operator bundle images
68
+
3. etcd performance issues
69
+
70
+
Once any underlying causes for the unpack failure have been addressed, deleting any failing unpack jobs and their owner configMaps will cause the subscription to retry unpacking the operator bundles once more.
71
+
72
+
You can enable automated cleanup and retry of failed unpack jobs in a namespace by setting the `operatorframework.io/bundle-unpack-min-retry-interval` annotation on the operatorGroup in that namespace. This annotation indicates the time after the last unpack failure when the unpack may be attempted again. This should not be set to an interval shorter than `5m` to avoid unnecessary load on the cluster.
This annotation does not enforce limits on the number of times an operator bundle may be unpacked on failure, preserving only 5 failing unpack attempts for inspection. Unless the underlying cause for the failure is addressed, this may cause OLM to attempt to unsuccessfully unpack the operator bundle indefinitely. Removing the annotation from the operatorGroup disables automated retries for failed unpacking jobs on that namespace.
79
+
80
+
With older versions of OLM, an installPlan may be generated for the failing subscription, in which case the subscription must be backed up; the failing installPlan, CSV, and subscription deleted; after which the subscription must be reapplied.
0 commit comments