Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DataFlowFlexTemplateJob: long updates lose job ID #2571

Open
3 tasks done
justinsb opened this issue Aug 27, 2024 · 1 comment
Open
3 tasks done

DataFlowFlexTemplateJob: long updates lose job ID #2571

justinsb opened this issue Aug 27, 2024 · 1 comment
Assignees
Labels
bug Something isn't working
Milestone

Comments

@justinsb
Copy link
Collaborator

Checklist

Bug Description

We had a report that a DataflowFlexTemplateJob can become detached from the underlying job, when a (streaming) job update takes more than 20 minutes.

Proposed approaches:

  • Increase the timeout (sort of a hack, but works around the problem)
  • Fix the terraform-based reconciler for this scenario
  • Replace the terraform-based reconciler with a direct reconciler that does not have this issue

Additional Diagnostic Information

This makes logical sense, the update does create a new job ID, and based on the code it is likely that we don't correctly record the job ID when there is an error/timeout from the update.

Kubernetes Cluster Version

n/a

Config Connector Version

n/a

Config Connector Mode

namespaced mode (default)

Log Output

No response

Steps to reproduce the issue

n/a

YAML snippets

No response

@justinsb justinsb added the bug Something isn't working label Aug 27, 2024
@justinsb justinsb added this to the 1.122 milestone Sep 6, 2024
@justinsb justinsb self-assigned this Sep 7, 2024
@jonapgar-groupby
Copy link

jonapgar-groupby commented Oct 8, 2024

I have what I believe is the same issue with Dataflow Jobs being constantly recreated around every twenty minutes. The jobs that are being replaced are running fine (they're not being updated or queued or anything).

I'm using the latest version of the controller manager: gcr.io/gke-release/cnrm/controller:826b049

It would be very helpful if there was a way to actually view the diffs being generated during reconciliation so we could troubleshoot. I didn't see these being logged anywhere.

I noted the caveats about not using Server-side Apply for k8s. I'm using Helm to deploy these CRs and Helm doesn't support ssa. However, I couldn't see any fields actually being set on the DataflowFlexTemplateJob CR so I don't believe that is the issue.

@yuwenma yuwenma modified the milestones: 1.122, 1.126 Oct 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants
@justinsb @yuwenma @jonapgar-groupby and others