Skip to content

OPECO-3054,OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii #600

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged

Conversation

stevekuznetsov
Copy link
Contributor

No description provided.

awgreene and others added 10 commits October 30, 2023 20:33
Signed-off-by: Alexander Greene <[email protected]>

Upstream-repository: operator-lifecycle-manager
Upstream-commit: 28977ca102802759d42fc608574fa6f29bec24f5
Generic APIServers now fail to start if the do not define the  OpenAPIV3
schemas. This commit introduces a change to define the OpenAPIV3 schemas

Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: a217efc73c7391f2737d836ce448114121290eb3
Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 9ec03f07f942dc9cef736957fa152e39157d6e13
This will specifically allow OLM to avoid the "use" verb for SCCs that
it wasn't configured to work with on OpenShift clusters.

Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 1c419a3fd20207ac8c6d41030e96d0c7b0275efd
This change tightens the alerting rules to avoid resetting the alerts
upon transient scrape failures.

It also removes the `message` annotation in favor of the `description`
annotation which is more commonly used by the Prometheus community.

Signed-off-by: Simon Pasquier <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: cda76ee57d15868284a3f412dcc19afeda4b644c
Signed-off-by: Simon Pasquier <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 2a9017153f62023c8ba68f1f9806dab5d0133ff5
A good chunk of this logic is fatally flawed - it would be much more
simple to manage a Deployment and use server-side apply to ensure that
the current server state always matched the desired state, but that
would be a large refactor and who knows how many other things are
loosely coupled here. This is the smallest change that allows for the
current serving pod to not yet have a running catalog server.

Signed-off-by: Steve Kuznetsov <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 95405d81e4c87c8113ccd7a95ba4d088b200a42a
Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-registry
Upstream-commit: f538df278daeed4b25453d15d6d282f0dc31beb7
Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-registry
Upstream-commit: b1374806c6d9028eb4cfe4343eb1a25002690237
Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: api
Upstream-commit: 071829b7c42526d9fb58023024abfb62c7a3e516
Signed-off-by: Steve Kuznetsov <[email protected]>
@openshift-ci openshift-ci bot requested review from kevinrizza and oceanc80 October 31, 2023 02:40
@stevekuznetsov
Copy link
Contributor Author

/retest

@awgreene
Copy link
Contributor

awgreene commented Oct 31, 2023

/approve
/lgtm

There may be a need to add the commit introduced in #601 to this so OLM can use the correct SCC.

@openshift-ci openshift-ci bot added the lgtm Indicates that a PR is ready to be merged. label Oct 31, 2023
Copy link
Contributor

openshift-ci bot commented Oct 31, 2023

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: awgreene, stevekuznetsov

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@openshift-ci openshift-ci bot added the approved Indicates a PR has been approved by an approver from all required OWNERS files. label Oct 31, 2023
@openshift-ci-robot
Copy link

/retest-required

Remaining retests: 0 against base HEAD 57e26f0 and 2 for PR HEAD ffddf17 in total

After reducing the RBAC granted to the OLM ServiceAccount in an earlier
commit, this commit introduces RBAC so pods that use the OLM
ServiceAccount will qualify to use the restricted-v2 SCC.

Signed-off-by: Alexander Greene <[email protected]>
Signed-off-by: Steve Kuznetsov <[email protected]>
@openshift-ci openshift-ci bot removed the lgtm Indicates that a PR is ready to be merged. label Nov 2, 2023
@awgreene
Copy link
Contributor

awgreene commented Nov 2, 2023

Infra failure on ci/prow/e2e-upgrade:

level=error
level=error msg=Error: creating EC2 Instance: InvalidNetworkInterfaceID.NotFound: The networkInterface ID 'eni-008fc950d67416d0b' does not exist
level=error msg=	status code: 400, request id: 555737c5-58c4-4a82-a15f-f67a9510a774
level=error
level=error msg=  with module.masters.aws_instance.master[1], 

/retest

Signed-off-by: Steve Kuznetsov <[email protected]>
@@ -28,6 +28,7 @@ rules:
- securitycontextconstraints
resourceNames:
- restricted-v2
- anyuid
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is needed to satisfy a feature in which we allow catalogSources in legacy mode, which cannot run under the constraints introduced by the restricted-v2 SCC, you can learn more by inspecting this test.

@stevekuznetsov
Copy link
Contributor Author

 Starting CatalogSource e2e tests configure gRPC registry pod to extract content 
Message:"failed to populate resolver cache from source catalog-7t7zb/openshift-catsrc-e2e-xqmg6: failed to list bundles: rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.224.120:50051: connect: connection refused\""

Looks bona-fide here

@stevekuznetsov
Copy link
Contributor Author

Nothing obvious in the must-gather:

catalog-operator:

2023-11-03T19:34:35.058377185Z time="2023-11-03T19:34:35Z" level=error msg="UpdateStatus - error while setting CatalogSource status" error="Operation cannot be fulfilled on catalogsources.operators.coreos.com \"catalog-lp5hw\": the object has been modified; please apply your changes to the latest version and try again" id=+zQfF source=catalog-lp5hw
2023-11-03T19:34:35.058510525Z E1103 19:34:35.058490       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: Operation cannot be fulfilled on catalogsources.operators.coreos.com "catalog-lp5hw": the object has been modified; please apply your changes to the latest version and try again
2023-11-03T19:34:56.088379536Z time="2023-11-03T19:34:56Z" level=error msg="UpdateStatus - error while setting CatalogSource status" error="Operation cannot be fulfilled on catalogsources.operators.coreos.com \"catalog-lp5hw\": the object has been modified; please apply your changes to the latest version and try again" id=s7oOB source=catalog-lp5hw
2023-11-03T19:34:56.088379536Z E1103 19:34:56.088264       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: Operation cannot be fulfilled on catalogsources.operators.coreos.com "catalog-lp5hw": the object has been modified; please apply your changes to the latest version and try again
2023-11-03T19:34:56.110049896Z time="2023-11-03T19:34:56Z" level=error msg="UpdateStatus - error while setting CatalogSource status" error="Operation cannot be fulfilled on catalogsources.operators.coreos.com \"catalog-lp5hw\": the object has been modified; please apply your changes to the latest version and try again" id=/lWgz source=catalog-lp5hw
2023-11-03T19:34:56.110049896Z E1103 19:34:56.109994       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: Operation cannot be fulfilled on catalogsources.operators.coreos.com "catalog-lp5hw": the object has been modified; please apply your changes to the latest version and try again
2023-11-03T19:35:04.438011155Z time="2023-11-03T19:35:04Z" level=error msg="UpdateStatus - error while setting CatalogSource status" error="Operation cannot be fulfilled on catalogsources.operators.coreos.com \"catalog-lp5hw\": the object has been modified; please apply your changes to the latest version and try again" id=lLBye source=catalog-lp5hw
2023-11-03T19:35:04.438141765Z E1103 19:35:04.438121       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: Operation cannot be fulfilled on catalogsources.operators.coreos.com "catalog-lp5hw": the object has been modified; please apply your changes to the latest version and try again
2023-11-03T19:35:04.889149213Z E1103 19:35:04.888921       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:04.892476612Z E1103 19:35:04.892120       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:04.898688642Z E1103 19:35:04.898629       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:04.923059502Z E1103 19:35:04.923004       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:04.967618532Z E1103 19:35:04.967491       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:05.051932332Z E1103 19:35:05.051875       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:05.089286521Z E1103 19:35:05.089213       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:05.216213951Z E1103 19:35:05.216099       1 queueinformer_operator.go:319] sync {"update" "openshift-catsrc-e2e-7fcjk/catalog-lp5hw"} failed: couldn't ensure registry server - error ensuring service account: catalog-lp5hw: serviceaccounts "catalog-lp5hw" is forbidden: unable to create new content in namespace openshift-catsrc-e2e-7fcjk because it is being terminated
2023-11-03T19:35:05.310442600Z time="2023-11-03T19:35:05Z" level=info msg="removed client for deleted catalogsource" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"

packageserver:

2023-11-03T19:34:34.851857826Z time="2023-11-03T19:34:34Z" level=info msg="connecting to source" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:34.856937457Z W1103 19:34:34.856832       1 clientconn.go:1331] [core] grpc: addrConn.createTransport failed to connect to {catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 <nil> 0 <nil>}. Err: connection error: desc = "transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused". Reconnecting...
2023-11-03T19:34:35.247542264Z time="2023-11-03T19:34:35Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:35.247601514Z time="2023-11-03T19:34:35Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:35.444364943Z time="2023-11-03T19:34:35Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:35.444530243Z time="2023-11-03T19:34:35Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:36.552236966Z time="2023-11-03T19:34:36Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:55.858830878Z W1103 19:34:55.858746       1 clientconn.go:1331] [core] grpc: addrConn.createTransport failed to connect to {catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 <nil> 0 <nil>}. Err: connection error: desc = "transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout". Reconnecting...
2023-11-03T19:34:55.859024788Z time="2023-11-03T19:34:55Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:55.859075098Z time="2023-11-03T19:34:55Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:55.859182058Z time="2023-11-03T19:34:55Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:56.079152846Z time="2023-11-03T19:34:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:56.079358286Z time="2023-11-03T19:34:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:56.100509116Z time="2023-11-03T19:34:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:56.100574566Z time="2023-11-03T19:34:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:35:04.429239375Z time="2023-11-03T19:35:04Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:35:05.310421820Z time="2023-11-03T19:35:05Z" level=warning msg="error getting bundle data" action="refresh cache" err="rpc error: code = Canceled desc = grpc: the client connection is closing" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:35:05.310829210Z time="2023-11-03T19:35:05Z" level=warning msg="error getting package stream" action="refresh cache" err="rpc error: code = Canceled desc = grpc: the client connection is closing" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:34.851907135Z time="2023-11-03T19:34:34Z" level=info msg="connecting to source" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:34.855663674Z W1103 19:34:34.855588       1 clientconn.go:1331] [core] grpc: addrConn.createTransport failed to connect to {catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 <nil> 0 <nil>}. Err: connection error: desc = "transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused". Reconnecting...
2023-11-03T19:34:35.247718081Z time="2023-11-03T19:34:35Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:35.247892101Z time="2023-11-03T19:34:35Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:35.444474815Z time="2023-11-03T19:34:35Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:35.444474815Z time="2023-11-03T19:34:35Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: connect: connection refused\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:36.552021222Z time="2023-11-03T19:34:36Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:55.856364131Z W1103 19:34:55.856292       1 clientconn.go:1331] [core] grpc: addrConn.createTransport failed to connect to {catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051 <nil> 0 <nil>}. Err: connection error: desc = "transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout". Reconnecting...
2023-11-03T19:34:55.856418301Z time="2023-11-03T19:34:55Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:55.856432181Z time="2023-11-03T19:34:55Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:55.856513321Z time="2023-11-03T19:34:55Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:56.079419992Z time="2023-11-03T19:34:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:56.079537862Z time="2023-11-03T19:34:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:34:56.100610480Z time="2023-11-03T19:34:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:34:56.100666180Z time="2023-11-03T19:34:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing dial tcp 172.30.104.104:50051: i/o timeout\"" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:35:04.429113421Z time="2023-11-03T19:35:04Z" level=info msg="updating PackageManifest based on CatalogSource changes: {catalog-lp5hw openshift-catsrc-e2e-7fcjk}" action="sync catalogsource" address="catalog-lp5hw.openshift-catsrc-e2e-7fcjk.svc:50051" name=catalog-lp5hw namespace=openshift-catsrc-e2e-7fcjk
2023-11-03T19:35:05.310717216Z time="2023-11-03T19:35:05Z" level=warning msg="error getting bundle data" action="refresh cache" err="rpc error: code = Canceled desc = grpc: the client connection is closing" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"
2023-11-03T19:35:05.310817886Z time="2023-11-03T19:35:05Z" level=warning msg="error getting package stream" action="refresh cache" err="rpc error: code = Canceled desc = grpc: the client connection is closing" source="{catalog-lp5hw openshift-catsrc-e2e-7fcjk}"

@stevekuznetsov
Copy link
Contributor Author

Actually looks like the test in question is Starting CatalogSource e2e tests image update ... which has entirely different semantics locally vs in OCP e2e downstream, so not super surprised it fails here.

There is also no logging in there whatsoever - adding it in operator-framework/operator-lifecycle-manager#3089, will rebase once that's merged upstream to pull it in.

@tmshort
Copy link
Contributor

tmshort commented Nov 7, 2023

Starting CatalogSource e2e tests image update
/go/src/github.com/openshift/operator-framework-olm/staging/operator-lifecycle-manager/test/e2e/catalog_e2e_test.go:806

Yup, that's what I saw.

@stevekuznetsov
Copy link
Contributor Author

/test e2e-gcp-olm

awgreene and others added 5 commits November 9, 2023 10:21
Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 2abfb3c18528b23f531db0b4baf64ab641f38036
This reverts commit 2abfb3c18528b23f531db0b4baf64ab641f38036.

Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: e8afbf7e5abe049f54fe5d9569e72863be9879f4
This reverts commit f7b970ea796e341941319d6e1a0966ac9af11055.

Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 17b910356999a523b327c8045ecf2ae958791802
Problem: Commit 95405d81e4c87c8113ccd7a95ba4d088b200a42ai updated the
catalog operator's logic so it does not delete the pod associated with a
catalogSource while it is in a Pending state. Unfortunately, there is a
race condition in which the pod may be admitted to the cluster without
the imagePullSecrets specified for it's serviceAccount by the admission
controller, preventing the pod from pulling its image from registries
that require authentication and causing the pod to never reach a
successful state.

Solution: Update the catalog operator to detect when a pod is missing
the imagePullSecrets granted to its serviceAccount.

Signed-off-by: Alexander Greene <[email protected]>
Upstream-repository: operator-lifecycle-manager
Upstream-commit: 0e1e089003d4fb656bc7278641c2db2fe0841598
…ft#306)

Signed-off-by: dtfranz <[email protected]>
Upstream-repository: api
Upstream-commit: 6b3567d593e430a0c9601e15ece31d031eac7ce9
Signed-off-by: Steve Kuznetsov <[email protected]>
@awgreene
Copy link
Contributor

awgreene commented Nov 9, 2023

lusteroperator/kube-apiserver is not available (<missing>) because <missing>
	clusteroperator/kube-controller-manager is not available (<missing>) because <missing>
	clusteroperator/kube-scheduler is not available (<missing>) because <missing>
	clusteroperator/kube-storage-version-migrator is not available (<missing>) because <missing>
	clusteroperator/machine-api is not available (<missing>) because <missing>
	clusteroperator/machine-approver is not available (<missing>) because <missing>
	clusteroperator/machine-config is not available (<missing>) because <missing>
	clusteroperator/marketplace is not available (<missing>) because <missing>
	clusteroperator/monitoring is not available (<missing>) because <missing>
	clusteroperator/network is not available (<missing>) because <missing>
	clusteroperator/node-tuning is not available (<missing>) because <missing>
	clusteroperator/openshift-apiserver is not available (<missing>) because <missing>
	clusteroperator/openshift-controller-manager is not available (<missing>) because <missing>
	clusteroperator/openshift-samples is not available (<missing>) because <missing>
	clusteroperator/operator-lifecycle-manager is not available (<missing>) because <missing>
	clusteroperator/operator-lifecycle-manager-catalog is not available (<missing>) because <missing>
	clusteroperator/operator-lifecycle-manager-packageserver is not available (<missing>) because <missing>
	clusteroperator/service-ca is not available (<missing>) because <missing>
	clusteroperator/storage is not available (<missing>) because <missing>

/retest

@stevekuznetsov
Copy link
Contributor Author

/retest

@awgreene
Copy link
Contributor

awgreene commented Nov 10, 2023

/retest
Trying again as the CI failures seem to have stopped.

@awgreene
Copy link
Contributor

awgreene commented Nov 10, 2023

/retest

Flake test suit will fail

@awgreene
Copy link
Contributor

/lgtm

@openshift-ci openshift-ci bot added the lgtm Indicates that a PR is ready to be merged. label Nov 10, 2023
@awgreene
Copy link
Contributor

awgreene commented Nov 10, 2023

/retitle OPRUN-3106: skuznets/bump vendor ii

@openshift-ci openshift-ci bot changed the title Skuznets/bump vendor ii OPRUN-3106: skuznets/bump vendor ii Nov 10, 2023
@openshift-ci-robot openshift-ci-robot added the jira/valid-reference Indicates that this PR references a valid Jira ticket of any type. label Nov 10, 2023
@openshift-ci-robot
Copy link

openshift-ci-robot commented Nov 10, 2023

@stevekuznetsov: This pull request references OPRUN-3106 which is a valid jira issue.

Warning: The referenced jira issue has an invalid target version for the target branch this PR targets: expected the task to target the "4.15.0" version, but no target version was set.

In response to this:

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@awgreene
Copy link
Contributor

awgreene commented Nov 10, 2023

/retitle OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii

@openshift-ci openshift-ci bot changed the title OPRUN-3106: skuznets/bump vendor ii OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii Nov 10, 2023
@openshift-ci-robot
Copy link

@stevekuznetsov: Jira Issue OCPBUGS-20347: All pull requests linked via external trackers have merged:

Jira Issue OCPBUGS-20347 has been moved to the MODIFIED state.

In response to this:

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@stevekuznetsov
Copy link
Contributor Author

@awgreene not sure the bots will be happy trying to make more than one Jira connected

@awgreene
Copy link
Contributor

@awgreene not sure the bots will be happy trying to make more than one Jira connected

@stevekuznetsov this is a supported workflow, Alex Pavel had sent an email out a few months ago:

... The Jira GitHub integration has recently been updated to support linking multiple bugs from OCPBUGS to a single PR. To link multiple bugs to a PR, simply include all bugs as a comma-separated list in the title (e.g. “OCPBUGS-123,OCPBUGS-124: Fixed the issue”). Support for multiple bugs should work for all the usual PR events that the integration usually handles, including cherrypicks and the /jira cherrypick command...

@awgreene
Copy link
Contributor

awgreene commented Nov 10, 2023

/retitle OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii

@openshift-ci openshift-ci bot changed the title OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii Nov 10, 2023
@openshift-ci-robot
Copy link

@stevekuznetsov: Jira Issue OCPBUGS-21980: All pull requests linked via external trackers have merged:

Jira Issue OCPBUGS-21980 has been moved to the MODIFIED state.

Jira Issue OCPBUGS-20347 is in an unrecognized state (ON_QA) and will not be moved to the MODIFIED state.

In response to this:

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@grokspawn
Copy link
Contributor

grokspawn commented Nov 16, 2023

/retitle OPECO-3054,OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii

@openshift-ci openshift-ci bot changed the title OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii OPECO-3054,OCPBUGS-21980,OCPBUGS-20347,OPRUN-3106: skuznets/bump vendor ii Nov 16, 2023
@openshift-ci-robot
Copy link

@stevekuznetsov: Jira Issue OCPBUGS-21980 is in an unrecognized state (Verified) and will not be moved to the MODIFIED state.

Jira Issue OCPBUGS-20347 is in an unrecognized state (Verified) and will not be moved to the MODIFIED state.

In response to this:

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved Indicates a PR has been approved by an approver from all required OWNERS files. jira/valid-reference Indicates that this PR references a valid Jira ticket of any type. lgtm Indicates that a PR is ready to be merged.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

9 participants