Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Communication with services sometimes failed #127

Closed
vitaliy-leschenko opened this issue Nov 24, 2020 · 10 comments
Closed

Communication with services sometimes failed #127

vitaliy-leschenko opened this issue Nov 24, 2020 · 10 comments
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@vitaliy-leschenko
Copy link
Contributor

vitaliy-leschenko commented Nov 24, 2020

Describe the bug
Windows pods can't communicate with services if target pod stay on different Windows node.
Maybe this is duplicate of another issues but symptoms are a little bit difference.

Windows pods can communicate with services from pods that run on Linux without any problem.
Linux pods can communicate with services from pods that run on Windows without any problem.

To Reproduce
Steps to reproduce the behavior:

  1. Setup new cluster with 2 linux and 2 windows nodes using PrepareNode.ps1
  2. Apply yaml
    It creates 2 daemonsets iis (on windows) and nginx (on linux) and 2 services for each daemonset.
  3. Try to connect to one of iis pods and curl:
    curl -Ik nginx.issues.svc.cluster.local <-- will be success in 100% cases
    curl -Ik iis.isses.svc.cluster.local <-- will be success only in cases when it tries to curl from pod that runs on the same node.

Expected behavior
I expect that curl -Ik iis.isses.svc.cluster.local will be success in 100% cases

Kubernetes (please complete the following information):

  • Windows Server version: 1809 (10.0.17763.1554)
  • Kubernetes Version: 1.19.0 - 1.19.4
  • CNI: flannel 0.13 vxlan

Additional context
All pods can ping each others by IP address.
All windows pods can curl from the Internet and ping each other.
Windows pods can curl from windows pods on different nodes by IP address but failed when I try to curl by Service name or Service IP address.

My PR #111 doesn't affect it because it reproduced without it.

@vitaliy-leschenko
Copy link
Contributor Author

I tried yaml from step 2 on baremetal cluster (1.17.0) created by https://github.com/kubernetes-sigs/sig-windows-tools/blob/master/kubeadm/KubeCluster.ps1 and this issue doesn't reproduced.

@vitaliy-leschenko
Copy link
Contributor Author

It is truly interesting but the issue doesn't reproduced on Windows Server 1809 (10.0.17763.1294)

@vitaliy-leschenko
Copy link
Contributor Author

It is look like a Windows issue.

@daschott
Copy link

daschott commented Dec 5, 2020

Is this still active?

@vitaliy-leschenko
Copy link
Contributor Author

@daschott Yes, it is.

@fejta-bot
Copy link

Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.

If this issue is safe to close now please do so with /close.

Send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Mar 5, 2021
@fejta-bot
Copy link

Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.

If this issue is safe to close now please do so with /close.

Send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Apr 4, 2021
@fejta-bot
Copy link

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-contributor-experience at kubernetes/community.
/close

@k8s-ci-robot
Copy link
Contributor

@fejta-bot: Closing this issue.

In response to this:

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-contributor-experience at kubernetes/community.
/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@vitaliy-leschenko
Copy link
Contributor Author

It looks like in 10.0.17763.1999 this issues was fixed. Maybe in one of previous updates but it doesn't matter. Currently we can use latest updates and don't have the issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

4 participants