-
Notifications
You must be signed in to change notification settings - Fork 116
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make Spark/Hadoop service installation idempotent #237
Comments
Quick thought: The same could be done for drive configuration. If (assuming we can't just look at df for /media/ephemeral0, /media/ephemeral1) a file exists with a specific name (e.g., FlintrockDrivesInstalledDontBotherDoingItAgain) then skip that as well. |
Service installation and configuration are already separated in Flintrock. We leverage this separation when adding new nodes to a cluster, for example, since when that happens all existing nodes need to have their services reconfigured but not reinstalled. I believe what you're asking for is that installation be idempotent. One easy example of Flintrock implementing a declarative-style method of managing software is To accomplish what you're looking for, we'd need to do a few things, some of which you touched on:
This is a good request, and discussing it reminds me again just how close Flintrock comes to reinventing other tools (like Ansible). 😄 Since Flintrock is strictly limited to Apache Spark and Hadoop, I'm fine with refining how we do things as long as it doesn't add a lot of complexity. It'll take a bit of work here to implement this in a non-hacky way, but I think it's possible, especially for the main services like Spark and Hadoop. |
As a user of flintrock, I would like to shave a lot of time off of spinning up new clusters.
To do so, I would like to copy an AMI from a previous flintrock install and reuse that.
Expected: without installing HDFS and Spark again, the new AMI is instantiated, the slaves files are updated, the master IP is updated in the appropriate config files and HDFS/Spark is launched.
Bonus expectation: I would love to tell Flintrock that I've already configured the drives correctly and be able to skip the ephemeral drive allocation step as well.
Actual: Today, I have to turn on the installation of each service to get them configured. No time savings for using an AMI with the software pre-installed.
The text was updated successfully, but these errors were encountered: