- update_keypair now only in v2 schema
- tenant_id->project_id in cluster provision steps
- tenant_id->project_id in referenced job binaries in job templates
- proper check for job template existence, to fail early (as intended)
- hadoop_version->plugin_version for query string filter
- unbreak some data source stuff (related to tenant_id->project_id)
- fix omission of show_progress from cluster GET query string whitelist
- job_id->job_template_id for v2 jobs
- add missing release note info for strict query string checking
- release notes for all the rest
Change-Id: Idea117c406b5ab9b8d85ccf8adb175053416d6ff
Story: 2004505
Task: 28822
In order to improve return information to clients we are adding a check
to verify parameters before each call and return a more appropriate
message to the users.
Change-Id: I9923601d0903e415a3fe30bec9bdc8fc34b91ff6
Story: #2004506
Task: #28228
As part of the APIv2 work we need to change all tenant_id references
to project_id on the return payload.
Story: #2004505
Taks: #28227
Change-Id: I94bca161aa4f7bdd56d5528bae92fa81af188a43
* Drop use of stack abandon: just use a regular delete instead
* Return stack name from force delete API call
Change-Id: I33ee7323fade1b237957abb8f7c79b87eb20148f
Creation of a single cluster and creation of multiple clusters will now
share an API endpoint (in the APIv2 case). More specifically, the
original single-cluster endpoint will accept a `count` parameter in the
request and the multiple-cluster endpoint has been removed.
We can make this kind of change because APIv2 is still experimental.
Also, when creating multiple clusters, the response will now contain
all details about the clusters; previously, the response simply gave
cluster IDs.
Change-Id: I90faf4956a8ea4b4ae31a29382732771fdfddecb
Story: 2002099
Task: 19777
* Add force delete operation to engine using stack abandon
* Basic unit tests for force delete
* Necessary removal of "direct engine" code
* Change schema of APIv2 cluster delete to allow force delete
* Unit tests for new cluster delete schema
* Basic docs about force delete
* Release note
bp sahara-force-delete
Partial-Bug: #1647411
Change-Id: Ida72677c0a4110cb78edf9d62d8330cd4608ff76
Adding the option to decommission specific node from cluster.
Partially implements bp: decommission-specific-instance
Change-Id: I1a858fecc1b32f91c76aea6db14e0d5a419211d1
hadoop_version should be changed to plugin_version.
This patch only changes json requests, the complete
change should be done when APIv2 is stable
and APIv1 deprecated. Along with this should be a
data model change as well.
Partial-Implements: bp v2-api-experimental-impl
Change-Id: Id59703ce49741d024ac9dee09d28f8f515c9806a
This change renames the service api modules to be organized under the
sahara.service.api package. A v2 package has been added which holds the
equivalent v10 and v11 functions. To make the api version setup
transparent to the caller, the global OPS variable has been refactored
into the base sahara.service.api package.
Changes
* create sahara/service/api package
* rename sahara/service/api.py to sahara/service/api/v10.py
* rename sahara/service/edp/api.py to sahara/service/api/v11.py
* correct occurances of old imports in code and tests
* rename sahara/tests/unit/service/test_api.py to
sahara/tests/unit/service/api/test_v10.py
* add initial v2 equivalents for current api services
* move global OPS object into the sahara.service.api package
* add documentation for the api service layer
Partial-Implements: bp v2-api-experimental-impl
Depends-On: I16918a30a862b42edd7a982caf555be618199ac3
Change-Id: Iefbedbc76ac620ff012bcaf536c17637b6252a15
this change implements the initial v2 experimental api. it is
implemented as an optional series of class that can be configured
through the paste deploy configuration file.
* add wsgi auth validator
* add middleware router
* add v2 endpoints
* add v2 flask blueprint object
* add optional paste filter and composite
* add developer docs for v2 api
Change-Id: I74627c0879851b354b5043f8a6ff91bae8438bb1
Partial-Implements: bp v2-api-experimental-impl