-
Notifications
You must be signed in to change notification settings - Fork 28.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SPARK-3332] Revert spark-ec2 patch that identifies clusters using tags #2225
Conversation
Both patches reverted cleanly using |
QA tests have started for PR 2225. This patch merges cleanly. |
QA results for PR 2225: |
Since the original two commits, there were only two intervening commits to (see https://github.com/JoshRosen/spark/commits/revert-ec2-cluster-naming/ec2/spark_ec2.py) |
LGTM. Hmm - This is unfortunate and something that I feared given the flaky experiences I have had with tags before. I checked the intervening commits too and the reverts look good to me. Could you also manually test by launching a EC2 cluster just to be doubly sure ? |
I launched a cluster earlier using this version of the script and shut it down using the v1.0.2 one. |
FYI: It looks like branch-1.1 has an out-of-date version of |
okay cool - let's merge this then (it isn't tested by jenkins anyways) although it would be nice to update that script in branch-1.1. |
I've merged this into branch-1.1. |
This reverts #1899 and #2163, two patches that modified `spark-ec2` so that clusters are identified using tags instead of security groups. The original motivation for this patch was to allow multiple clusters to run in the same security group. Unfortunately, tagging is not atomic with launching instances on EC2, so with this approach we have the possibility of `spark-ec2` launching instances and crashing before they can be tagged, effectively orphaning those instances. The orphaned instances won't belong to any cluster, so the `spark-ec2` script will be unable to clean them up. Since this feature may still be worth supporting, there are several alternative approaches that we might consider, including detecting orphaned instances and logging warnings, or maybe using another mechanism to group instances into clusters. For the 1.1.0 release, though, I propose that we just revert this patch. Author: Josh Rosen <joshrosen@apache.org> Closes #2225 from JoshRosen/revert-ec2-cluster-naming and squashes the following commits: 0c18e86 [Josh Rosen] Revert "SPARK-2333 - spark_ec2 script should allow option for existing security group" c2ca2d4 [Josh Rosen] Revert "Spark-3213 Fixes issue with spark-ec2 not detecting slaves created with "Launch More like this""
This reverts #1899 and #2163, two patches that modified
spark-ec2
so that clusters are identified using tags instead of security groups. The original motivation for this patch was to allow multiple clusters to run in the same security group.Unfortunately, tagging is not atomic with launching instances on EC2, so with this approach we have the possibility of
spark-ec2
launching instances and crashing before they can be tagged, effectively orphaning those instances. The orphaned instances won't belong to any cluster, so thespark-ec2
script will be unable to clean them up.Since this feature may still be worth supporting, there are several alternative approaches that we might consider, including detecting orphaned instances and logging warnings, or maybe using another mechanism to group instances into clusters. For the 1.1.0 release, though, I propose that we just revert this patch.