From 6d2e6c5f7b7e241c904cb7da4816729244211c18 Mon Sep 17 00:00:00 2001 From: Jacob Freck Date: Mon, 29 Oct 2018 12:28:26 -0700 Subject: [PATCH] Release: v0.10.0 (#676) * fix function name * update version * changelog updates * set minor version to correct number * update changelog * changelog formatting * typos * whitespace * remove print debug statement * whitespace * whitespace --- CHANGELOG.md | 13 +++++++++++++ aztk/version.py | 4 ++-- aztk_cli/config.py | 10 +++++----- 3 files changed, 20 insertions(+), 7 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index ea6da4bb..acd6990e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Changelog +## 0.10.0 (2018-10-29) + +**Breaking Changes** +* Remove deprecated SDK API code (#671) ([fc50536](https://github.com/Azure/aztk/commit/fc50536)), closes [#671](https://github.com/Azure/aztk/issues/671) +* Remove custom scripts (#673) ([9e32b4b](https://github.com/Azure/aztk/commit/9e32b4b)), closes [#673](https://github.com/Azure/aztk/issues/673) +* Replaced states with Enums for ClusterState, JobState, ApplicationState (#677) ([e486536](https://github.com/Azure/aztk/commit/e486536)), closes [#677](https://github.com/Azure/aztk/issues/677) + +**Features** +* Spark retry docker pull (#672) ([18b74e4](https://github.com/Azure/aztk/commit/18b74e4)), closes [#672](https://github.com/Azure/aztk/issues/672) +* Spark scheduling target (#661) ([4408c4f](https://github.com/Azure/aztk/commit/4408c4f)), closes [#661](https://github.com/Azure/aztk/issues/661) +* Spark submit scheduling internal (#674) ([8c2bf0c](https://github.com/Azure/aztk/commit/8c2bf0c)), closes [#674](https://github.com/Azure/aztk/issues/674) + + ## 0.9.1 (2018-10-5) **Bug Fixes** * Fix: pin all node dependencies not in Pipfile (#667) ([0606598](https://github.com/Azure/aztk/commit/0606598)), closes [#667](https://github.com/Azure/aztk/issues/667) diff --git a/aztk/version.py b/aztk/version.py index 2223ec12..f4ff3221 100644 --- a/aztk/version.py +++ b/aztk/version.py @@ -22,8 +22,8 @@ # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. major = 0 -minor = 9 -patch = 1 +minor = 10 +patch = 0 suffix = "" diff --git a/aztk_cli/config.py b/aztk_cli/config.py index 856cd659..10fe3d4c 100644 --- a/aztk_cli/config.py +++ b/aztk_cli/config.py @@ -156,7 +156,7 @@ def merge(self, cluster_id, username, job_ui_port, job_history_ui_port, web_ui_p "Please supply a username either in the ssh.yaml configuration file or with a parameter (--username)") -def __convert_to_path(path: str): +def _convert_to_path(path: str): if path: abs_path = os.path.abspath(os.path.expanduser(path)) if not os.path.exists(abs_path): @@ -226,10 +226,10 @@ def _merge_dict(self, config): spark_configuration = config.get("spark_configuration") if spark_configuration: - self.spark_defaults_conf = __convert_to_path(spark_configuration.get("spark_defaults_conf")) - self.spark_env_sh = __convert_to_path(spark_configuration.get("spark_env_sh")) - self.core_site_xml = __convert_to_path(spark_configuration.get("core_site_xml")) - self.jars = [__convert_to_path(jar) for jar in spark_configuration.get("jars") or []] + self.spark_defaults_conf = _convert_to_path(spark_configuration.get("spark_defaults_conf")) + self.spark_env_sh = _convert_to_path(spark_configuration.get("spark_env_sh")) + self.core_site_xml = _convert_to_path(spark_configuration.get("core_site_xml")) + self.jars = [_convert_to_path(jar) for jar in spark_configuration.get("jars") or []] def _read_config_file(self, path: str = aztk.utils.constants.DEFAULT_SPARK_JOB_CONFIG): """