Sahara projects migrated to storyboard.openstack.org.
Replace the references to Launchpad, including the bugs now
available as stories.
Fix a reference to github.
Change-Id: Iadba69efc1e310b6a19463d3398bf5c6549acd73
Instead of hard-coded Spark 1.6.0, allow use of DIB_SPARK_VERSION (and
its corresponding CLI argument `-s`) to specify which version of Spark
should be included on images for Vanilla plugin.
Change-Id: Ia7c7027c9eadfc9d724733a503990ca78e487ee9
Building CDH image under version 5.5.0 is no longer support.
Remove these useless code.
Also, adding the ambari usage info in sahara-image-create
command.
Change-Id: I6fffe25ee9daf651355611be675137babb67e2a8
Update the plugin version we support(5.7) and will support(5.9) into doc.
Though "DIB_LOCAL_IMAGE" is defined in project "diskimage-builder",
add a specification in docs will be helpful.
Partially-Implements: blueprint cdh-5-9-support
Change-Id: I11ada0c05d426ea483f7c3b7ffa5b247eb77133c
we don't have vanilla 2.6.0 in supported list
in all current branches of sahara. we can
just drop that. if needed, stable/mitaka
branch should be used for building that image.
Change-Id: I81ed8209f2154f112fe7f6718029b84548793380
New version of HDP plugin uses this element for installing Ambari
- Default ambari version changed to 2.2.1.0
partially implements bp: hdp-24-support
Change-Id: I91ac4c304ffff3d665b2129845a430672aeba782
Addressing multiple problems in README.rst:
- Outdated support matrix
- Incorrect and incomplete parameter set and options
- Line wraps
- Script name inconsistencies
- Generalized and complete outdatedness
Change-Id: Iefd1bff54bafe11e519bd295795fa3b4569b731f
Closes-bug: 1596715
Spark 1.6.0 is available now for deployment.
changes:
* using spark 1.6.0 version as default version for spark plugin
* using hadoop 2.6 for spark 1.6.0 (by setting up
DIB_CDH_VERSION="5.4")
* we use only HDFS part of CDH for Spark plugin so we don't need to
create symbolic links for oozie
Depends-on: I9a46a503c7e52d756c7de8c8694dbfc51f80f2be
bp: support-spark-160
Change-Id: Id3c4fb5a0cb1349824972c1be90ea0e0be64cb4b
New version of HDP plugin uses this element for installing Ambari
partially implements bp: hdp-22-support
Change-Id: Icdbd1832fb0f1a64a0a41dd67b840f0d2a38b8b0
Update the Spark element to use the existing hadoop-cloudera element for HDFS
for Spark versions > 1.0, instead of the ad-hoc cloudera-cdh one. For Spark 1.0.2,
CDH4 via the old hadoop-cdh element is used, since a precompiled binary for CDH5
is not available.
This change also makes it possible to specify an arbitrary Spark version via the
new -s commandline switch, reducing the amount of code for supporting future
versions of Spark. The defaults for Spark are 1.3.1 and CDH 5.3, a combination
that works well in our deployments.
A small change is needed in the cloudera element: when creating a Spark image,
only the HDFS packages have to be installed.
README files have been updated to clarify that default versions are tested, while
other combinations are not. A reference to the SparkPlugin wiki page was added
to point to a table of supported versions.
Change-Id: Ifc2a0c8729981e1e1df79b556a4c2e6bd1ba893a
Implements: blueprint support-spark-1-3
Depends-On: I8fa482b6d1d6abaa6633aec309a3ba826a8b7ebb
Changes:
* Use libraries from tox env instead their downloading
* Use pbr for getting package version in sahara-version element
Change-Id: Ifa7128b2d2bc0dbf5b4bc023d5e47540f4c03408
This change removes the global system updates in favor of only updating
the required packages for diskimage creation. Additionally the script
will not automatically install the updates but will warn the user if the
packages are not installed. A switch is provided to allow the script to
install the updates.
Changes
* adding function to detect an installed package
* adding function to check for a list of installed packages
* adding command line switch for updating (-u)
* removing global update commands
* refactoring update logic to fail gracefully or accept switch
* adding -u switch to build-images gate tool
* adding note in documentation about update command line switch
Closes-Bug: 1375843
Change-Id: Iba5e676fa8ef7a852b71acd9e4c19d32f848285a
Adding Storm as one of the images that can be created by
diskimagebuilder.
Implements: blueprint insert-storm-as-image-option
Change-Id: If44ee630d090f4e032a3e0bfb71ee363f55eec56
Using IMAGE_TYPE to reflect the OS of the base image
conflicts with the way that the diskimage-builder project
uses IMAGE_TYPE (qcow2 vs tar) and was confusing.
We are now using BASE_IMAGE_OS since it properly
reflects what we are using it for and does not conflict
with the diskimage-builder project.
Change-Id: I9e84077c1bc9b8890076838e34a157cb7a3884db
Closes-Bug: #1378364
We could specify ubuntu and fedora local mirror for image build,
now we can use local mirrors for all type of images.
Yum-mirror element was divided in two elements: fedora-mirror
and centos-mirror
Change-Id: Id5016d8a7e982010b6683c038afa21e639ef4ac0
* Added the -m option
* Added detection of variable DIB_REPO_BRANCH
* Added tag 0.1.17 as default branch of DIB repo
* Added exit error if both DIB_REPO_BRANCH and -m are specified
Implements: blueprint dib-repo-branch-option
Change-Id: I44f0689f7844085c39b5560c65024b68c3424ad8
This patch modifies the common "java" element in order to allow
users to specify the directory that will contain the JDK install
and this support is for all install types now.
Change-Id: I5ba04ba5e22c7f5f5a6455669bbe8509c042a893
WARNING:
----
Before merging this commit the alias for
http://sahara-files.mirantis.com needs to be in place.
Also before merging this commit the new openstack git project must be
available at https://git.openstack.org/openstack/sahara-image-elements/
NOTE:
----
The file 'elements/hadoop-hdp/source-repository-hadoopswift' contains a
link to the HortonWorks repository that holds the Hadoop Swift rpm, this
link needs to be updated when HortonWorks makes the change.
Implements: blueprint savanna-renaming-image-elements
Change-Id: Icb9a992f8545535af3a111580ce7c9622d754c67
The script is divided into blocks corresponding version of installed plugin.
Added options for specifying image type and hadoop version. By using option
you can choose image that you want to test instead of creating them all.
Change-Id: I5be5413e978acae26454a1a98825649f7d6498cd
Add the '-p' commandline option to the diskimage-builder script to
select which Savanna plugin to target during image generation.
Vanilla and Spark plugins are supported. By default (if no option is
specified) the Vanilla plugin is selected.
Change-Id: I76af023ebd47bb026308b6034cacf01360739be1
* add parameters 'DIB_REPO_PATH' and 'SIM_REPO_PATH' for specifying pathes
* script tries to use local repository 'savanna-image-elements' as default
Change-Id: I46f3b5573e68e5dfaf5fd98b86eedc90fa108073
* As base image we use our own CentOS cloud image with preinstalled cloud-init
* Configuring that image
* Edit hadoop, mysql element for CentOS support and 'diskimage-create' script
Change-Id: If5570f31fd69c7cc809948dd2fa0249d00323f50
* Automatic building Fedora and Ubuntu cloud images with all elements from savanna-extra repository
* Using mirrors for updating images
Change-Id: I12632b5cee42b1dbfd79b7b7c3a7b26962ace625