diff options
-rw-r--r-- | INFO.yaml | 4 | ||||
-rwxr-xr-x | check | 223 | ||||
-rwxr-xr-x | ci/build-auto.sh | 19 | ||||
-rw-r--r-- | pylintrc | 561 | ||||
-rw-r--r-- | requirements.txt | 2 | ||||
-rw-r--r-- | setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py | 602 | ||||
-rw-r--r-- | setup/VIMs/OpenStack/clouds.yaml | 99 | ||||
-rw-r--r-- | yamllintrc | 25 |
8 files changed, 1257 insertions, 278 deletions
@@ -66,11 +66,11 @@ committers: email: 'paul.vaduva@enea.com' company: 'enea.com' id: 'pvaduva' - - name: 'Martin Klozik' + - name: 'Martin Klozik' email: 'martin.klozik@tieto.com' company: 'tieto.com' id: 'mklozik' - - name: 'Richard Elias' + - name: 'Richard Elias' email: 'richard.elias@tieto.com' company: 'tieto.com' id: 'richardxelias' @@ -0,0 +1,223 @@ +#!/bin/bash + +# Copyright 2017-2018 Intel Corporation, Tieto +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +# Auto project python & yaml formatting checker +# This script is based on the file ./check from OPNFV vswitchperf +# project. + +# +# Configuration +# +PYLINT="pylint" +PYLINT_RC='pylintrc' +PYTHON_FILE_REGEX="\.py$" +YAMLLINT="yamllint" +YAMLLINT_RC='yamllintrc' +YAML_FILE_REGEX="\.yaml$" +FILE_LIST="/tmp/auto_check_list.txt" + +CHECK_PYTHON=0 +CHECK_YAML=0 + +# +# Support Functions +# +# print usage if requested +function usage() { + cat <<EOM +Usage: $0 [TARGET]... + +Performs code check for defined TARGETs. Target can be file or directory. +In case that directory is specified, then it will be searched recursively +for all python and yaml files. +If TARGET is not specified, then all python and yaml files from current AUTO +repository will be checked. + + + -h, --help Script usage + -b, --black Suppress colours. Output will be black&white. + -m, --modified Script will check python and yaml files, which have + been modified within current repository. + +Examples: + ./check + + Check all python and yaml files in current AUTO repository + + ./check INFO.yaml + + Check just one file. + + ./check -m + + Check all modified files in current AUTO repository + + ./check lib/auto/testcase lib/auto/util + + Check all python and yaml files in given directories + +EOM +} + +# get list of files to be checked +function get_file_list() { + # store file regex and shift params to get list of original ./check options + TMP_FILE_REGEX=$1 + shift + + rm $FILE_LIST &> /dev/null + if [ "x$1" == "x-m" -o "x$1" == "x--modified" ] ; then + # check of modified files requested + git status --porcelain | cut -b4- | egrep -i "${TMP_FILE_REGEX}" | sort > $FILE_LIST + elif [ "x$*" == "x" ] ; then + # list is empty, check all python files + git ls-tree --name-only -r HEAD | egrep -i "${TMP_FILE_REGEX}" | sort > $FILE_LIST + else + for item in $* ; do + if [ -d $item ] ; then + git ls-tree --name-only -r HEAD $item | egrep -i "${TMP_FILE_REGEX}" | sort >> $FILE_LIST + elif [ -f $item ] ; then + echo $item | egrep -i "${TMP_FILE_REGEX}" >> $FILE_LIST + else + echo "$item doesn't exist, thus check was aborted" + exit 1 + fi + done + fi +} + +function check_lint_binary() { + # check if lint binary is available + if ! which $1 &>/dev/null ; then + echo "$1 is not available, thus check can't be executed" + return 1 + fi + return 0 +} + + +function check_python() { + echo "Execution of pylint checks:" + + if ! check_lint_binary $PYLINT ; then + CHECK_PYTHON=1 + return + fi + + # check if there is anything to check + if [ -s $FILE_LIST ] ; then + for pyfile in `cat $FILE_LIST | sort` ; do + # get base name + pyfile_basename="'"`basename $pyfile .py`"'" + # run pylint and extract final rating + output=`$PYLINT --rcfile $PYLINT_RC $pyfile 2>/dev/null` + rating=`echo -e $output | tail -n3 | grep rated | sed -e 's/^.*rated at \(-\?[0-9.]*\).*$/\1/'` + # evaluate and display aquired rating + if [ "x$rating" == "x" ] ; then + # rating is not available for files without python statements + printf " %-70s %-6s\n" $pyfile "NA" + elif [ "$rating" == "10" ] ; then + printf " %-70s ${GREEN}%-6s${BLACK}\n" $pyfile "OK" + else + CHECK_PYTHON=1 + echo -e "$output" | awk '/^\*+ Module|^[A-Z]\:/' + printf " %-70s ${RED}%-6s${BLACK}\n" $pyfile $rating + fi + done + else + echo " Nothing to check." + fi +} + +function check_yaml() { + echo "Execution of yaml checks:" + + if ! check_lint_binary $YAMLLINT ; then + CHECK_YAML=1 + return + fi + + # check if there is anything to check + if [ -s $FILE_LIST ] ; then + for yamlfile in `cat $FILE_LIST | sort` ; do + output=`$YAMLLINT -c $YAMLLINT_RC $yamlfile 2>/dev/null` + if [ $? -eq 0 ] ; then + printf " %-70s ${GREEN}%-6s${BLACK}\n" $yamlfile "OK" + else + CHECK_YAML=1 + echo "$output" + printf " %-70s ${RED}%-6s${BLACK}\n" $yamlfile "FAILED" + fi + done + else + echo " Nothing to check." + fi +} + +# +# Main +# +# check if help is requested +if [ "x$1" == "x-h" -o "x$1" == "x--help" ] ; then + usage + exit 0 +fi + +# set colours +if [ "x$1" == "x-b" -o "x$1" == "x--black" ] ; then + shift + RED="" + GREEN="" + BLACK="" +else + RED="\e[31m" + GREEN="\e[32m" + BLACK="\e[0m" +fi + +# check if we were run within auto directory +if [ ! -x ./check 2> /dev/null ] ; then + echo "`basename $0` must be run from auto root directory" + exit 1 +fi + +# run python checks +get_file_list $PYTHON_FILE_REGEX $* +check_python + +echo + +# run yaml checks +get_file_list $YAML_FILE_REGEX $* +check_yaml + +# clean up +rm $FILE_LIST &> /dev/null + +# return success or failure based on pylint and yamllint checks +# NOTE: As of now, failure of pylint checks is not propagated into exit code. +# This will be turned on again after the rating of existing python +# files will be improved. +# if [ $CHECK_PYTHON -eq 0 -a $CHECK_YAML -eq 0 ] ; then +if [ $CHECK_YAML -eq 0 ] ; then + exit 0 +else + exit 1 +fi + +# +# The End +# diff --git a/ci/build-auto.sh b/ci/build-auto.sh index 611d83a..96588b9 100755 --- a/ci/build-auto.sh +++ b/ci/build-auto.sh @@ -30,6 +30,7 @@ # EXIT=0 EXIT_UNKNOWN_JOB_TYPE=1 +EXIT_LINT_FAILED=2 # # configuration @@ -37,6 +38,16 @@ EXIT_UNKNOWN_JOB_TYPE=1 AUTOENV_DIR="$HOME/autoenv" # +# functions +# +# execute pylint and yamllint to check code quality +function execute_auto_lint_check() { + if ! ./check -b ; then + EXIT=$EXIT_LINT_FAILED + fi +} + +# # main # echo @@ -69,7 +80,7 @@ case $1 in # Example of verify job body. Functions can call # external scripts, etc. - #execute_auto_pylint_check + execute_auto_lint_check #execute_auto_doc_check #install_opnfv MCP #install_onap @@ -87,14 +98,14 @@ case $1 in # Example of merge job body. Functions can call # external scripts, etc. - #execute_auto_pylint_check + execute_auto_lint_check #execute_auto_doc_check #install_opnfv MCP #install_onap #execute_sanity_check #execute_tests $1 - # Everything went well, so report SUCCESS to Jenkins + # propagate result to the Jenkins job exit $EXIT ;; "daily") @@ -111,7 +122,7 @@ case $1 in #execute_tests $1 #push_results_and_logs_to_artifactory - # Everything went well, so report SUCCESS to Jenkins + # propagate result to the Jenkins job exit $EXIT ;; *) diff --git a/pylintrc b/pylintrc new file mode 100644 index 0000000..c213b80 --- /dev/null +++ b/pylintrc @@ -0,0 +1,561 @@ +# Copyright 2018 Tieto +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +[MASTER] + +# A comma-separated list of package or module names from where C extensions may +# be loaded. Extensions are loading into the active Python interpreter and may +# run arbitrary code +extension-pkg-whitelist= + +# Add files or directories to the blacklist. They should be base names, not +# paths. +ignore=CVS + +# Add files or directories matching the regex patterns to the blacklist. The +# regex matches against base names, not paths. +ignore-patterns= + +# Python code to execute, usually for sys.path manipulation such as +# pygtk.require(). +#init-hook= + +# Use multiple processes to speed up Pylint. +jobs=1 + +# List of plugins (as comma separated values of python modules names) to load, +# usually to register additional checkers. +load-plugins= + +# Pickle collected data for later comparisons. +persistent=yes + +# Specify a configuration file. +#rcfile= + +# When enabled, pylint would attempt to guess common misconfiguration and emit +# user-friendly hints instead of false-positive error messages +suggestion-mode=yes + +# Allow loading of arbitrary C extensions. Extensions are imported into the +# active Python interpreter and may run arbitrary code. +unsafe-load-any-extension=no + + +[MESSAGES CONTROL] + +# Only show warnings with the listed confidence levels. Leave empty to show +# all. Valid levels: HIGH, INFERENCE, INFERENCE_FAILURE, UNDEFINED +confidence= + +# Disable the message, report, category or checker with the given id(s). You +# can either give multiple identifiers separated by comma (,) or put this +# option multiple times (only on the command line, not in the configuration +# file where it should appear only once).You can also use "--disable=all" to +# disable everything first and then reenable specific checks. For example, if +# you want to run only the similarities checker, you can use "--disable=all +# --enable=similarities". If you want to run only the classes checker, but have +# no Warning level messages displayed, use"--disable=all --enable=classes +# --disable=W" +disable=print-statement, + parameter-unpacking, + unpacking-in-except, + old-raise-syntax, + backtick, + long-suffix, + old-ne-operator, + old-octal-literal, + import-star-module-level, + non-ascii-bytes-literal, + invalid-unicode-literal, + raw-checker-failed, + bad-inline-option, + locally-disabled, + locally-enabled, + file-ignored, + suppressed-message, + useless-suppression, + deprecated-pragma, + apply-builtin, + basestring-builtin, + buffer-builtin, + cmp-builtin, + coerce-builtin, + execfile-builtin, + file-builtin, + long-builtin, + raw_input-builtin, + reduce-builtin, + standarderror-builtin, + unicode-builtin, + xrange-builtin, + coerce-method, + delslice-method, + getslice-method, + setslice-method, + no-absolute-import, + old-division, + dict-iter-method, + dict-view-method, + next-method-called, + metaclass-assignment, + indexing-exception, + raising-string, + reload-builtin, + oct-method, + hex-method, + nonzero-method, + cmp-method, + input-builtin, + round-builtin, + intern-builtin, + unichr-builtin, + map-builtin-not-iterating, + zip-builtin-not-iterating, + range-builtin-not-iterating, + filter-builtin-not-iterating, + using-cmp-argument, + eq-without-hash, + div-method, + idiv-method, + rdiv-method, + exception-message-attribute, + invalid-str-codec, + sys-max-int, + bad-python3-import, + deprecated-string-function, + deprecated-str-translate-call, + deprecated-itertools-function, + deprecated-types-field, + next-method-defined, + dict-items-not-iterating, + dict-keys-not-iterating, + dict-values-not-iterating, + deprecated-operator-function, + deprecated-urllib-function, + xreadlines-attribute, + deprecated-sys-function, + exception-escape, + comprehension-escape + +# Enable the message, report, category or checker with the given id(s). You can +# either give multiple identifier separated by comma (,) or put this option +# multiple time (only on the command line, not in the configuration file where +# it should appear only once). See also the "--disable" option for examples. +enable=c-extension-no-member + + +[REPORTS] + +# Python expression which should return a note less than 10 (10 is the highest +# note). You have access to the variables errors warning, statement which +# respectively contain the number of errors / warnings messages and the total +# number of statements analyzed. This is used by the global evaluation report +# (RP0004). +evaluation=10.0 - ((float(5 * error + warning + refactor + convention) / statement) * 10) + +# Template used to display messages. This is a python new-style format string +# used to format the message information. See doc for all details +#msg-template= + +# Set the output format. Available formats are text, parseable, colorized, json +# and msvs (visual studio).You can also give a reporter class, eg +# mypackage.mymodule.MyReporterClass. +output-format=text + +# Tells whether to display a full report or only the messages +reports=no + +# Activate the evaluation score. +score=yes + + +[REFACTORING] + +# Maximum number of nested blocks for function / method body +max-nested-blocks=5 + +# Complete name of functions that never returns. When checking for +# inconsistent-return-statements if a never returning function is called then +# it will be considered as an explicit return statement and no message will be +# printed. +never-returning-functions=optparse.Values,sys.exit + + +[FORMAT] + +# Expected format of line ending, e.g. empty (any line ending), LF or CRLF. +expected-line-ending-format= + +# Regexp for a line that is allowed to be longer than the limit. +ignore-long-lines=^\s*(# )?<?https?://\S+>?$ + +# Number of spaces of indent required inside a hanging or continued line. +indent-after-paren=4 + +# String used as indentation unit. This is usually " " (4 spaces) or "\t" (1 +# tab). +indent-string=' ' + +# Maximum number of characters on a single line. +max-line-length=160 + +# Maximum number of lines in a module +max-module-lines=1000 + +# List of optional constructs for which whitespace checking is disabled. `dict- +# separator` is used to allow tabulation in dicts, etc.: {1 : 1,\n222: 2}. +# `trailing-comma` allows a space between comma and closing bracket: (a, ). +# `empty-line` allows space-only lines. +no-space-check=trailing-comma, + dict-separator + +# Allow the body of a class to be on the same line as the declaration if body +# contains single statement. +single-line-class-stmt=no + +# Allow the body of an if to be on the same line as the test if there is no +# else. +single-line-if-stmt=no + + +[BASIC] + +# Naming style matching correct argument names +argument-naming-style=snake_case + +# Regular expression matching correct argument names. Overrides argument- +# naming-style +#argument-rgx= + +# Naming style matching correct attribute names +attr-naming-style=snake_case + +# Regular expression matching correct attribute names. Overrides attr-naming- +# style +#attr-rgx= + +# Bad variable names which should always be refused, separated by a comma +bad-names=foo, + bar, + baz, + toto, + tutu, + tata + +# Naming style matching correct class attribute names +class-attribute-naming-style=any + +# Regular expression matching correct class attribute names. Overrides class- +# attribute-naming-style +#class-attribute-rgx= + +# Naming style matching correct class names +class-naming-style=PascalCase + +# Regular expression matching correct class names. Overrides class-naming-style +#class-rgx= + +# Naming style matching correct constant names +const-naming-style=UPPER_CASE + +# Regular expression matching correct constant names. Overrides const-naming- +# style +#const-rgx= + +# Minimum line length for functions/classes that require docstrings, shorter +# ones are exempt. +docstring-min-length=-1 + +# Naming style matching correct function names +function-naming-style=snake_case + +# Regular expression matching correct function names. Overrides function- +# naming-style +#function-rgx= + +# Good variable names which should always be accepted, separated by a comma +good-names=i, + j, + k, + e, + ex, + Run, + _ + +# Include a hint for the correct naming format with invalid-name +include-naming-hint=no + +# Naming style matching correct inline iteration names +inlinevar-naming-style=any + +# Regular expression matching correct inline iteration names. Overrides +# inlinevar-naming-style +#inlinevar-rgx= + +# Naming style matching correct method names +method-naming-style=snake_case + +# Regular expression matching correct method names. Overrides method-naming- +# style +#method-rgx= + +# Naming style matching correct module names +module-naming-style=snake_case + +# Regular expression matching correct module names. Overrides module-naming- +# style +#module-rgx= + +# Colon-delimited sets of names that determine each other's naming style when +# the name regexes allow several styles. +name-group= + +# Regular expression which should only match function or class names that do +# not require a docstring. +no-docstring-rgx=^_ + +# List of decorators that produce properties, such as abc.abstractproperty. Add +# to this list to register other decorators that produce valid properties. +property-classes=abc.abstractproperty + +# Naming style matching correct variable names +variable-naming-style=snake_case + +# Regular expression matching correct variable names. Overrides variable- +# naming-style +#variable-rgx= + + +[MISCELLANEOUS] + +# List of note tags to take in consideration, separated by a comma. +notes=FIXME, + XXX, + TODO + + +[TYPECHECK] + +# List of decorators that produce context managers, such as +# contextlib.contextmanager. Add to this list to register other decorators that +# produce valid context managers. +contextmanager-decorators=contextlib.contextmanager + +# List of members which are set dynamically and missed by pylint inference +# system, and so shouldn't trigger E1101 when accessed. Python regular +# expressions are accepted. +generated-members= + +# Tells whether missing members accessed in mixin class should be ignored. A +# mixin class is detected if its name ends with "mixin" (case insensitive). +ignore-mixin-members=yes + +# This flag controls whether pylint should warn about no-member and similar +# checks whenever an opaque object is returned when inferring. The inference +# can return multiple potential results while evaluating a Python object, but +# some branches might not be evaluated, which results in partial inference. In +# that case, it might be useful to still emit no-member and other checks for +# the rest of the inferred objects. +ignore-on-opaque-inference=yes + +# List of class names for which member attributes should not be checked (useful +# for classes with dynamically set attributes). This supports the use of +# qualified names. +ignored-classes=optparse.Values,thread._local,_thread._local + +# List of module names for which member attributes should not be checked +# (useful for modules/projects where namespaces are manipulated during runtime +# and thus existing member attributes cannot be deduced by static analysis. It +# supports qualified module names, as well as Unix pattern matching. +ignored-modules= + +# Show a hint with possible names when a member name was not found. The aspect +# of finding the hint is based on edit distance. +missing-member-hint=yes + +# The minimum edit distance a name should have in order to be considered a +# similar match for a missing member name. +missing-member-hint-distance=1 + +# The total number of similar names that should be taken in consideration when +# showing a hint for a missing member. +missing-member-max-choices=1 + + +[LOGGING] + +# Logging modules to check that the string format arguments are in logging +# function parameter format +logging-modules=logging + + +[SIMILARITIES] + +# Ignore comments when computing similarities. +ignore-comments=yes + +# Ignore docstrings when computing similarities. +ignore-docstrings=yes + +# Ignore imports when computing similarities. +ignore-imports=no + +# Minimum lines number of a similarity. +min-similarity-lines=4 + + +[SPELLING] + +# Limits count of emitted suggestions for spelling mistakes +max-spelling-suggestions=4 + +# Spelling dictionary name. Available dictionaries: none. To make it working +# install python-enchant package. +spelling-dict= + +# List of comma separated words that should not be checked. +spelling-ignore-words= + +# A path to a file that contains private dictionary; one word per line. +spelling-private-dict-file= + +# Tells whether to store unknown words to indicated private dictionary in +# --spelling-private-dict-file option instead of raising a message. +spelling-store-unknown-words=no + + +[VARIABLES] + +# List of additional names supposed to be defined in builtins. Remember that +# you should avoid to define new builtins when possible. +additional-builtins= + +# Tells whether unused global variables should be treated as a violation. +allow-global-unused-variables=yes + +# List of strings which can identify a callback function by name. A callback +# name must start or end with one of those strings. +callbacks=cb_, + _cb + +# A regular expression matching the name of dummy variables (i.e. expectedly +# not used). +dummy-variables-rgx=_+$|(_[a-zA-Z0-9_]*[a-zA-Z0-9]+?$)|dummy|^ignored_|^unused_ + +# Argument names that match this expression will be ignored. Default to name +# with leading underscore +ignored-argument-names=_.*|^ignored_|^unused_ + +# Tells whether we should check for unused import in __init__ files. +init-import=no + +# List of qualified module names which can have objects that can redefine +# builtins. +redefining-builtins-modules=six.moves,past.builtins,future.builtins,io,builtins + + +[DESIGN] + +# Maximum number of arguments for function / method +max-args=5 + +# Maximum number of attributes for a class (see R0902). +max-attributes=7 + +# Maximum number of boolean expressions in a if statement +max-bool-expr=5 + +# Maximum number of branch for function / method body +max-branches=12 + +# Maximum number of locals for function / method body +max-locals=15 + +# Maximum number of parents for a class (see R0901). +max-parents=7 + +# Maximum number of public methods for a class (see R0904). +max-public-methods=20 + +# Maximum number of return / yield for function / method body +max-returns=6 + +# Maximum number of statements in function / method body +max-statements=50 + +# Minimum number of public methods for a class (see R0903). +min-public-methods=2 + + +[CLASSES] + +# List of method names used to declare (i.e. assign) instance attributes. +defining-attr-methods=__init__, + __new__, + setUp + +# List of member names, which should be excluded from the protected access +# warning. +exclude-protected=_asdict, + _fields, + _replace, + _source, + _make + +# List of valid names for the first argument in a class method. +valid-classmethod-first-arg=cls + +# List of valid names for the first argument in a metaclass class method. +valid-metaclass-classmethod-first-arg=mcs + + +[IMPORTS] + +# Allow wildcard imports from modules that define __all__. +allow-wildcard-with-all=no + +# Analyse import fallback blocks. This can be used to support both Python 2 and +# 3 compatible code, which means that the block might have code that exists +# only in one or another interpreter, leading to false positives when analysed. +analyse-fallback-blocks=no + +# Deprecated modules which should not be used, separated by a comma +deprecated-modules=optparse,tkinter.tix + +# Create a graph of external dependencies in the given file (report RP0402 must +# not be disabled) +ext-import-graph= + +# Create a graph of every (i.e. internal and external) dependencies in the +# given file (report RP0402 must not be disabled) +import-graph= + +# Create a graph of internal dependencies in the given file (report RP0402 must +# not be disabled) +int-import-graph= + +# Force import order to recognize a module as part of the standard +# compatibility libraries. +known-standard-library= + +# Force import order to recognize a module as part of a third party library. +known-third-party=enchant + + +[EXCEPTIONS] + +# Exceptions that will emit a warning when being caught. Defaults to +# "Exception" +overgeneral-exceptions=Exception diff --git a/requirements.txt b/requirements.txt index 8ef8db6..1035c76 100644 --- a/requirements.txt +++ b/requirements.txt @@ -6,3 +6,5 @@ python-glanceclient>=2.8.0 python-neutronclient>=6.3.0 python-novaclient>=9.0.0 python-heatclient>=1.6.1 +pylint==1.9.2 +yamllint==1.11.1 diff --git a/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py b/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py index bab2c59..e4b94f5 100644 --- a/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py +++ b/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py @@ -24,13 +24,15 @@ # https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095 #docstring -"""This script configures an OpenStack instance to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP. +"""This script configures an OpenStack instance to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP. It can also remove the created objects, when used in a clean-up procedure. +Use -h option to see usage (-del to delete objects, -deb to print debug information). +Requirements: python 3, OpenStack SDK (0.14 or greater), clouds.yaml file, .img files are downloaded Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095 """ ###################################################################### -# This script configures an OpenStack instance (e.g. from an OPNFV installer like FUEL/MCP) to make it -# ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP. +# This script configures an OpenStack instance (e.g. from an OPNFV installer like FUEL/MCP, Compass4nfv, ...) +# to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP. # After running this script, the created OpenStack object names/IDs can be used for example to populate # YAML&ENV files used by ONAP (installation of ONAP itself, VNF descriptor files, etc.). @@ -42,12 +44,15 @@ Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095 # (optional, probably not needed: create a new group, which can be associated to a project, and contains users) # 2) create an ONAP user within the ONAP project, so as not to use the "admin" user for ONAP # (associate user to group if applicable; credentials: name/pwd or name/APIkey, or token) -# 3) create an ONAP security group, to allow ICMP traffic (for pings) and TCP port 22 (for SSH), rather than changing default security group(s) -# (optional, probably not needed: create new region; default region RegionOne is OK) +# 3) create an ONAP security group, to allow ICMP traffic (for pings) and TCP port 22 (for SSH), +# rather than changing default security group(s) +# (optional, probably not needed: create a new region; default region RegionOne is OK) # 4) create a public network for ONAP VNFs, with subnet and CIDR block # (so components have access to the Internet, via router and gateway, on unnamed ports, dynamic IP@ allocation) -# 5) create a private and an OAM network for ONAP VNFs or other ONAP components, with their respective subnet and CIDR block -# (ONAP VNFs will be deployed in this private and/or OAM network(s), usually with named ports and static IP@ as per VNF configuration file) +# 5) create a private and an OAM network for ONAP VNFs or other ONAP components, +# with their respective subnet and CIDR block +# (ONAP VNFs will be deployed in this private and/or OAM network(s), usually with named ports +# and static IP@ as per VNF configuration file) # 6) create an OpenStack router, with interfaces to the public, private and OAM networks, # and a reference to an external network (gateway) provided by the OpenStack instance installation # 7) create VM flavors as needed: m1.medium, etc. @@ -59,11 +64,13 @@ Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095 # - python3 is installed # - OpenStack SDK is installed for python3 # - there is a clouds.yaml file (describing the OpenStack instance, especially the Auth URL and admin credentials) +# - .img files (Ubuntu Trusty Tahr, Xenial Xerus, Cirros, ... are downloaded, and stored in IMAGES_DIR # - the script connects to OpenStack as a user with admin rights # typical commands to install OpenStack SDK Python client: # apt install python3-pip # pip3 install --upgrade pip +# hash -r # pip3 list # pip3 install openstacksdk # pip3 install --upgrade openstacksdk @@ -85,7 +92,8 @@ ONAP_USER_NAME = 'ONAP_user' ONAP_USER_PASSWORD = 'auto_topsecret' ONAP_USER_DESC = 'OpenStack User created for ONAP' -ONAP_TENANT_NAME = 'ONAP_tenant' # "project" is a more generic concept than "tenant"; a tenant is type of project; quotas are per project; +ONAP_TENANT_NAME = 'ONAP_tenant' +# note: "project" is a more generic concept than "tenant"; a tenant is type of project; quotas are per project; ONAP_TENANT_DESC = 'OpenStack Project/Tenant created for ONAP' ONAP_SECU_GRP_NAME = 'ONAP_security_group' @@ -93,30 +101,35 @@ ONAP_SECU_GRP_DESC = 'Security Group created for ONAP' ONAP_PUBLIC_NET_NAME = 'ONAP_public_net' ONAP_PUBLIC_SUBNET_NAME = 'ONAP_public_subnet' -ONAP_PUBLIC_SUBNET_CIDR = '192.168.99.0/24' # some arbitrary CIDR, but typically in a private (IANA-reserved) address range +ONAP_PUBLIC_SUBNET_CIDR = '192.168.99.0/24' +# note: some arbitrary CIDR, but typically in a private (IANA-reserved) address range ONAP_PUBLIC_NET_DESC = 'Public network created for ONAP, for unnamed ports, dynamic IP@, access to the Internet (e.g., Nexus repo) via Gateway' ONAP_PRIVATE_NET_NAME = 'ONAP_private_net' ONAP_PRIVATE_SUBNET_NAME = 'ONAP_private_subnet' -ONAP_PRIVATE_SUBNET_CIDR = '10.0.0.0/16' # should match ONAP installation; Private and OAM may be the same network +ONAP_PRIVATE_SUBNET_CIDR = '10.0.0.0/16' +# note: CIDR should match ONAP installation; Private and OAM may be the same network ONAP_PRIVATE_NET_DESC = 'Private network created for ONAP, for named ports, static IP@, inter-component communication' ONAP_OAM_NET_NAME = 'ONAP_OAM_net' ONAP_OAM_SUBNET_NAME = 'ONAP_OAM_subnet' -ONAP_OAM_SUBNET_CIDR = '10.99.0.0/16' # should match ONAP installation; Private and OAM may be the same network +ONAP_OAM_SUBNET_CIDR = '10.99.0.0/16' +# note: CIDR should match ONAP installation; Private and OAM may be the same network ONAP_OAM_NET_DESC = 'OAM network created for ONAP, for named ports, static IP@, inter-component communication' ONAP_ROUTER_NAME = 'ONAP_router' ONAP_ROUTER_DESC = 'Router created for ONAP' -EXTERNAL_NETWORK_NAME = 'floating_net' # OpenStack instance external network (gateway) name to be used as router's gateway +# OpenStack instance external network (gateway) name to be used as router's gateway +EXTERNAL_NETWORK_NAME = 'floating_net' -ONAP_KEYPAIR_NAME = 'ONAP_keypair' # keypair that can be used to SSH into created servers (VNF VMs) +# keypair that can be used to SSH into created servers (VNF VMs) +ONAP_KEYPAIR_NAME = 'ONAP_keypair' # OpenStack cloud name and region name, which should be the same as in the clouds.yaml file used by this script -OPENSTACK_CLOUD_NAME = 'hpe16openstackFraser' +OPENSTACK_CLOUD_NAME = 'unh-hpe-openstack-fraser' OPENSTACK_REGION_NAME = 'RegionOne' -# OpenStack domain is: Default +# note: OpenStack domain is: Default ###################################################################### @@ -144,45 +157,58 @@ def print_debug(*args): ###################################################################### def delete_all_ONAP(): - """Delete all ONAP-specific OpenStack objects (normally not needed, but may be useful during tests).""" + """Delete all ONAP-specific OpenStack objects (normally not needed, but may be useful during tests, and for clean-up).""" print('\nOPNFV Auto, script to delete ONAP objects in an OpenStack instance') try: # connect to OpenStack instance using Connection object from OpenStack SDK print('Opening connection...') conn = openstack.connect( - identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml - cloud = OPENSTACK_CLOUD_NAME, - region_name = OPENSTACK_REGION_NAME) + identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml + cloud = OPENSTACK_CLOUD_NAME, + region_name = OPENSTACK_REGION_NAME) + # delete router; must delete router before networks (and must delete VMs before routers) print('Deleting ONAP router...') onap_router = conn.network.find_router(ONAP_ROUTER_NAME) print_debug('onap_router:',onap_router) if onap_router != None: + + # delete router interfaces before deleting router + router_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME) + if router_network != None: + if router_network.subnet_ids != None: + print_debug('router_network.subnet_ids:',router_network.subnet_ids) + for subnet_id in router_network.subnet_ids: + print(' Deleting interface to',ONAP_PUBLIC_NET_NAME,'...') + conn.network.remove_interface_from_router(onap_router, subnet_id) + + router_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME) + if router_network != None: + if router_network.subnet_ids != None: + print_debug('router_network.subnet_ids:',router_network.subnet_ids) + for subnet_id in router_network.subnet_ids: + print(' Deleting interface to',ONAP_PRIVATE_NET_NAME,'...') + conn.network.remove_interface_from_router(onap_router, subnet_id) + + router_network = conn.network.find_network(ONAP_OAM_NET_NAME) + if router_network != None: + if router_network.subnet_ids != None: + print_debug('router_network.subnet_ids:',router_network.subnet_ids) + for subnet_id in router_network.subnet_ids: + print(' Deleting interface to',ONAP_OAM_NET_NAME,'...') + conn.network.remove_interface_from_router(onap_router, subnet_id) + + # and finally delete ONAP router conn.network.delete_router(onap_router.id) + else: print('No ONAP router found...') + # TODO@@@ verify if there are ports on networks (e.g., from VMs); if yes, can't delete network - # try: # try to circumvent !=None issue with OpenStack Resource.py; nope; - # onap_router = conn.network.find_router(ONAP_ROUTER_NAME) - # print_debug('onap_router:',onap_router) - # print_debug('onap_router.name:',onap_router.name) - # print_debug('onap_router.id:',onap_router.id) - # conn.network.delete_router(onap_router.id) - # conn.network.delete_router(onap_router) - # except Exception as e: - # print('issue Deleting ONAP router...') - # print('*** Exception:',type(e), e) - # exceptionType, exceptionValue, exceptionTraceback = sys.exc_info() - # print('*** traceback.print_tb():') - # traceback.print_tb(exceptionTraceback) - # print('*** traceback.print_exception():') - # traceback.print_exception(exceptionType, exceptionValue, exceptionTraceback) - - - # delete private network (which should also delete associated subnet and ports if any) + # delete private network (which should also delete associated subnet if any) print('Deleting ONAP private network...') private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME) print_debug('private_network:',private_network) @@ -191,7 +217,7 @@ def delete_all_ONAP(): else: print('No ONAP private network found...') - # delete OAM network (which should also delete associated subnet and ports if any) + # delete OAM network (which should also delete associated subnet if any) print('Deleting ONAP OAM network...') oam_network = conn.network.find_network(ONAP_OAM_NET_NAME) print_debug('oam_network:',oam_network) @@ -200,7 +226,7 @@ def delete_all_ONAP(): else: print('No ONAP OAM network found...') - # delete public network (which should also delete associated subnet and ports if any) + # delete public network (which should also delete associated subnet if any) print('Deleting ONAP public network...') public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME) print_debug('public_network:',public_network) @@ -209,6 +235,8 @@ def delete_all_ONAP(): else: print('No ONAP public network found...') + # TODO@@@ verify if security group is in use (e.g., by a VM), otherwise can't delete it + # delete security group print('Deleting ONAP security group...') onap_security_group = conn.network.find_security_group(ONAP_SECU_GRP_NAME) @@ -245,6 +273,8 @@ def delete_all_ONAP(): else: print('No ONAP keypair found...') + # no need to delete images and flavors + except Exception as e: print('*** Exception:',type(e), e) @@ -267,65 +297,9 @@ def configure_all_ONAP(): # connect to OpenStack instance using Connection object from OpenStack SDK print('Opening connection...') conn = openstack.connect( - identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml - cloud = OPENSTACK_CLOUD_NAME, - region_name = OPENSTACK_REGION_NAME) - - ################################################################### - # TESTS: IGNORE/DELETE (BEGIN) - # gdserver_ID = '8bc274a2-8c0d-4795-9b4d-faa0a21e1d88' - # gdserver = conn.compute.get_server(gdserver_ID) - # print('\ngdserver.name=',gdserver.name) - # print('gdserver.status=',gdserver.status) - - # print("\nList Users:") - # i=1 - # for user in conn.identity.users(): - # print('User',str(i),user.name,'\n',user,'\n') - # i+=1 - - # print("\nList Projects:") - # i=1 - # for project in conn.identity.projects(): - # print('Project',str(i),project.name,'\n',project,'\n') - # i+=1 - - # print("\nList Roles:") - # i=1 - # for role in conn.identity.roles(): - # print('Role',str(i),role.name,'\n',role,'\n') - # i+=1 - - # print("\nList Networks:") - # i=1 - # for network in conn.network.networks(): - # print('Network',str(i),network.name,'\n',network,'\n') - # i+=1 - - # print("\nList Routers:") - # i=1 - # for router in conn.network.routers(): - # print('Router',str(i),router.name,'\n',router,'\n') - # i+=1 - - # print("\nList Flavors:") - # i=1 - # for flvr in conn.compute.flavors(): - # print('Flavor',str(i),flvr.name,'\n',flvr,'\n') - # i+=1 - - # print("\nList Images:") - # i=1 - # for img in conn.compute.images(): - # print('Image',str(i),img.name,'\n',img,'\n') - # i+=1 - - # router = conn.network.find_router('gd_test_router') - # print('gd router\n',router,'\n\n') - # router = conn.network.find_router('e4e59f63-8063-4774-a97a-c110c6969e4a') - # print('gd router\n',router,'\n\n') - # TESTS: IGNORE/DELETE (END) - ################################################################### + identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml + cloud = OPENSTACK_CLOUD_NAME, + region_name = OPENSTACK_REGION_NAME) print('Creating ONAP project/tenant...') @@ -396,7 +370,8 @@ def configure_all_ONAP(): port_range_min = None, port_range_max = None) except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) try: description_text = 'enable ICMP egress IPv4' @@ -411,7 +386,8 @@ def configure_all_ONAP(): port_range_min = None, port_range_max = None) except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) try: description_text = 'enable SSH (TCP port 22) ingress IPv4' @@ -426,7 +402,8 @@ def configure_all_ONAP(): port_range_min = '22', port_range_max = '22') except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) try: description_text = 'enable SSH (TCP port 22) egress IPv4' @@ -441,7 +418,8 @@ def configure_all_ONAP(): port_range_min = '22', port_range_max = '22') except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) try: description_text = 'enable IP traffic ingress IPv4' @@ -456,7 +434,8 @@ def configure_all_ONAP(): port_range_min = None, port_range_max = None) except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) try: description_text = 'enable IP traffic ingress IPv6' @@ -471,44 +450,48 @@ def configure_all_ONAP(): port_range_min = None, port_range_max = None) except Exception as e: - print(description_text, ' Exception:', type(e), e) + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) # IPv4 IP egress rule should already exist by default - # try: - # description_text = 'enable IP traffic egress IPv4' - # print(' Creating rule:',description_text,'...') - # conn.network.create_security_group_rule( - # security_group_id = onap_security_group.id, - # description = description_text, - # protocol = None, - # direction = 'egress', - # ethertype = 'IPv4', - # remote_ip_prefix = '0.0.0.0/0', - # port_range_min = None, - # port_range_max = None) - # except Exception as e: - # print(description_text, ' Exception:', type(e), e) + try: + description_text = 'enable IP traffic egress IPv4' + print(' Creating rule:',description_text,'...') + conn.network.create_security_group_rule( + security_group_id = onap_security_group.id, + description = description_text, + protocol = None, + direction = 'egress', + ethertype = 'IPv4', + remote_ip_prefix = '0.0.0.0/0', + port_range_min = None, + port_range_max = None) + except Exception as e: + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) # IPv6 IP egress rule should already exist by default - # try: - # description_text = 'enable IP traffic egress IPv6' - # print(' Creating rule:',description_text,'...') - # conn.network.create_security_group_rule( - # security_group_id = onap_security_group.id, - # description = description_text, - # protocol = None, - # direction = 'egress', - # ethertype = 'IPv6', - # remote_ip_prefix = '::/0', - # port_range_min = None, - # port_range_max = None) - # except Exception as e: - # print(description_text, ' Exception:', type(e), e) + try: + description_text = 'enable IP traffic egress IPv6' + print(' Creating rule:',description_text,'...') + conn.network.create_security_group_rule( + security_group_id = onap_security_group.id, + description = description_text, + protocol = None, + direction = 'egress', + ethertype = 'IPv6', + remote_ip_prefix = '::/0', + port_range_min = None, + port_range_max = None) + except Exception as e: + print(' rule:', description_text, ' may already exist') + print_debug(description_text, ' Exception:', type(e), e) # public network print('Creating ONAP public network...') public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME) + public_subnet = None if public_network != None: print('ONAP public network already exists') else: @@ -530,13 +513,14 @@ def configure_all_ONAP(): ip_version = 4, is_dhcp_enabled = True, dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@ + print_debug('public_subnet:',public_subnet) print_debug('public_network: after subnet',public_network) - print_debug('public_subnet:',public_subnet) # private network print('Creating ONAP private network...') private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME) + private_subnet = None if private_network != None: print('ONAP private network already exists') else: @@ -557,13 +541,14 @@ def configure_all_ONAP(): ip_version = 4, is_dhcp_enabled = True, dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@; maybe not needed for private network - print_debug('private_network: after subnet',private_network) print_debug('private_subnet:',private_subnet) + print_debug('private_network: after subnet',private_network) # OAM network print('Creating ONAP OAM network...') oam_network = conn.network.find_network(ONAP_OAM_NET_NAME) + oam_subnet = None if oam_network != None: print('ONAP OAM network already exists') else: @@ -584,8 +569,8 @@ def configure_all_ONAP(): ip_version = 4, is_dhcp_enabled = True, dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@; maybe not needed for OAM network - print_debug('oam_network: after subnet',oam_network) print_debug('oam_subnet:',oam_subnet) + print_debug('oam_network: after subnet',oam_network) # router @@ -630,131 +615,93 @@ def configure_all_ONAP(): conn.network.add_interface_to_router(onap_router, subnet_id = oam_subnet.id) print_debug('onap_router: after adding interfaces',onap_router) - # point to OpenStack external network (i.e. Gateway for router); network_id is passed in a body dictionary - # (external network such as floating_net) - # syntax: add_gateway_to_router(router, **body) - #print('Adding external network (gateway) to ONAP router...') - - # nope - # network_dict_body = {'network_id': public_network.id} - # nope - # network_dict_body = { - # 'external_fixed_ips': [{'subnet_id' : public_subnet.id}], - # 'network_id': public_network.id - # } - - # external_network = conn.network.find_network(EXTERNAL_NETWORK_NAME) - # print_debug('external_network:',external_network) - # external_subnet_ID_list = external_network.subnet_ids - # print_debug('external_subnet_ID_list:',external_subnet_ID_list) - - # # build external_fixed_ips: list of dictionaries, each with 'subnet_id' key (and may have 'ip_address' key as well) - # onap_gateway_external_subnets = [] - # for ext_subn_id in external_subnet_ID_list: # there should be only one subnet ID in the list, but go through each item, just in case - # onap_gateway_external_subnets.append({'subnet_id':ext_subn_id}) - - # #network_dict_body = {'gateway' : {'network_id' : external_network.id}} - # #network_dict_body = {'network_id' : external_network.id} - # #conn.network.add_gateway_to_router(onap_router, body=network_dict_body) - # #conn.network.add_gateway_to_router(onap_router, network_id=external_network.id) - # #conn.network.add_gateway_to_router(onap_router, **network_dict_body) - - # network_dict_body = { - # 'network_id': external_network.id, - # 'enable_snat': True, # True should be the default, so there should be no need to set it - # 'external_fixed_ips': onap_gateway_external_subnets - # } - # #conn.network.add_gateway_to_router(onap_router, **network_dict_body) - # print_debug('onap_router: after add_gateway_to_router',onap_router) - - - - - # # also create 5 flavors, from tiny to xlarge (hard-coded, no need for parameters) - # # (Flavor is a Resource) - # print('Creating flavors...') - # print('Creating m1.tiny Flavor...') - # tiny_flavor = conn.compute.find_flavor("m1.tiny") - # if tiny_flavor != None: - # print('m1.tiny Flavor already exists') - # else: - # tiny_flavor = conn.compute.create_flavor( - # name = 'm1.tiny', - # vcpus = 1, - # disk = 1, - # ram = 512, - # ephemeral = 0, - # #swap = 0, - # #rxtx_factor = 1.0, - # is_public = True) - # print_debug('tiny_flavor: ',tiny_flavor) - - # print('Creating m1.small Flavor...') - # small_flavor = conn.compute.find_flavor("m1.small") - # if small_flavor != None: - # print('m1.small Flavor already exists') - # else: - # small_flavor = conn.compute.create_flavor( - # name = 'm1.small', - # vcpus = 1, - # disk = 20, - # ram = 2048, - # ephemeral = 0, - # #swap = 0, - # #rxtx_factor = 1.0, - # is_public = True) - # print_debug('small_flavor: ',small_flavor) - - # print('Creating m1.medium Flavor...') - # medium_flavor = conn.compute.find_flavor("m1.medium") - # if medium_flavor != None: - # print('m1.medium Flavor already exists') - # else: - # medium_flavor = conn.compute.create_flavor( - # name = 'm1.medium', - # vcpus = 2, - # disk = 40, - # ram = 4096, - # ephemeral = 0, - # #swap = 0, - # #rxtx_factor = 1.0, - # is_public = True) - # print_debug('medium_flavor: ',medium_flavor) - - # print('Creating m1.large Flavor...') - # large_flavor = conn.compute.find_flavor("m1.large") - # if large_flavor != None: - # print('m1.large Flavor already exists') - # else: - # large_flavor = conn.compute.create_flavor( - # name = 'm1.large', - # vcpus = 4, - # disk = 80, - # ram = 8192, - # ephemeral = 0, - # #swap = 0, - # #rxtx_factor = 1.0, - # is_public = True) - # print_debug('large_flavor: ',large_flavor) - - # print('Creating m1.xlarge Flavor...') - # xlarge_flavor = conn.compute.find_flavor("m1.xlarge") - # if xlarge_flavor != None: - # print('m1.xlarge Flavor already exists') - # else: - # xlarge_flavor = conn.compute.create_flavor( - # name = 'm1.xlarge', - # vcpus = 8, - # disk = 160, - # ram = 16384, - # ephemeral = 0, - # #swap = 0, - # #rxtx_factor = 1.0, - # is_public = True) - # print_debug('xlarge_flavor: ',xlarge_flavor) + + # also create 5 flavors, from tiny to xlarge (hard-coded, no need for parameters) + # (Flavor is a Resource) + print('Creating flavors...') + print('Creating m1.tiny Flavor...') + tiny_flavor = conn.compute.find_flavor("m1.tiny") + if tiny_flavor != None: + print('m1.tiny Flavor already exists') + else: + tiny_flavor = conn.compute.create_flavor( + name = 'm1.tiny', + vcpus = 1, + disk = 1, + ram = 512, + ephemeral = 0, + #swap = 0, + #rxtx_factor = 1.0, + is_public = True) + print_debug('tiny_flavor: ',tiny_flavor) + + print('Creating m1.small Flavor...') + small_flavor = conn.compute.find_flavor("m1.small") + if small_flavor != None: + print('m1.small Flavor already exists') + else: + small_flavor = conn.compute.create_flavor( + name = 'm1.small', + vcpus = 1, + disk = 20, + ram = 2048, + ephemeral = 0, + #swap = 0, + #rxtx_factor = 1.0, + is_public = True) + print_debug('small_flavor: ',small_flavor) + + print('Creating m1.medium Flavor...') + medium_flavor = conn.compute.find_flavor("m1.medium") + if medium_flavor != None: + print('m1.medium Flavor already exists') + else: + medium_flavor = conn.compute.create_flavor( + name = 'm1.medium', + vcpus = 2, + disk = 40, + ram = 4096, + ephemeral = 0, + #swap = 0, + #rxtx_factor = 1.0, + is_public = True) + print_debug('medium_flavor: ',medium_flavor) + + print('Creating m1.large Flavor...') + large_flavor = conn.compute.find_flavor("m1.large") + if large_flavor != None: + print('m1.large Flavor already exists') + else: + large_flavor = conn.compute.create_flavor( + name = 'm1.large', + vcpus = 4, + disk = 80, + ram = 8192, + ephemeral = 0, + #swap = 0, + #rxtx_factor = 1.0, + is_public = True) + print_debug('large_flavor: ',large_flavor) + + print('Creating m1.xlarge Flavor...') + xlarge_flavor = conn.compute.find_flavor("m1.xlarge") + if xlarge_flavor != None: + print('m1.xlarge Flavor already exists') + else: + xlarge_flavor = conn.compute.create_flavor( + name = 'm1.xlarge', + vcpus = 8, + disk = 160, + ram = 16384, + ephemeral = 0, + #swap = 0, + #rxtx_factor = 1.0, + is_public = True) + print_debug('xlarge_flavor: ',xlarge_flavor) # create images: Ubuntu 16.04, 14.04, CirrOS, ... + # store them in images/ directory # 64-bit QCOW2 image for cirros-0.4.0-x86_64-disk.img # description: CirrOS minimal Linux distribution # http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img @@ -770,31 +717,109 @@ def configure_all_ONAP(): # http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img # user: ubuntu - # do not use compute proxy for images; there is an image proxy (v1, and v2); use shade layer, directly with Connection object; + # do not use compute proxy for images; there is an image proxy (v1, and v2); + # use shade layer, directly with Connection object: Connection.create_image() # conn.get_image() returns a Python Munch object (subclass of Dictionary) - # URL download not supported yet; download image separately, place it in the directory + # However, URL download not supported yet; download image separately, place it in the directory # https://docs.openstack.org/openstacksdk/latest/user/connection.html#openstack.connection.Connection.create_image + # image proxy: conn.image.upload_image() + # Image class: + # https://docs.openstack.org/openstacksdk/latest/user/resources/image/v2/image.html#openstack.image.v2.image.Image + # URL should be supported by image proxy + + # TODO@@@ try image v2 proxy, if it supports URLs; + # maybe load only images for current CPU (i.e. only x86 images for x86, only Arm images for Arm) + # TODO@@@ list image names/URLs in dictionary, and load then in a loop + + # Pattern: prepare an attribute dictionary, then call conn.image.upload_image() + # image_attributes_dict = {} + # image_attributes_dict['name']='cirros-0.4.0-aarch64-disk.img' + # image_attributes_dict['url']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img' + # conn.image.upload_image(disk_format='qcow2',**image_attributes_dict) + + # With a dictionary of names/URLs : + # image_ref_dict = {} + # image_ref_dict['cirros-0.4.0-x86_64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img' + # image_ref_dict['cirros-0.4.0-arm-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-arm-disk.img' + # image_ref_dict['cirros-0.4.0-aarch64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img' + # etc. + # for image_name in image_ref_dict: + # image_attributes_dict['name'] = image_name + # image_attributes_dict['url'] = image_ref_dict[image_name] + # conn.image.upload_image(disk_format='qcow2',**image_attributes_dict) + + + # Create and populate image dictionary + image_ref_dict = {} + + # Ubuntu 16.04 LTS (Xenial Xerus) images + image_ref_dict['xenial-server-cloudimg-amd64-disk1.img']='https://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-amd64-disk1.img' + image_ref_dict['xenial-server-cloudimg-arm64-disk1.img']='https://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-arm64-disk1.img' + + # Ubuntu 14.04.5 LTS (Trusty Tahr) images + image_ref_dict['trusty-server-cloudimg-amd64-disk1.img']='http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img' + image_ref_dict['trusty-server-cloudimg-arm64-disk1.img']='http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-arm64-disk1.img' + + # CirrOS images + image_ref_dict['cirros-0.4.0-x86_64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img' + image_ref_dict['cirros-0.4.0-arm-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-arm-disk.img' + image_ref_dict['cirros-0.4.0-aarch64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img' + + + # if URL-based upload using image proxy works, it will replace the section below which assumes image files + # are in a subdirectory, and uses Connection.create_image() instead of Connection.image.upload_image() + IMAGES_DIR = 'images/' + + IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution x86' + print('Creating image:',IMAGE_NAME,'...') + if conn.get_image(IMAGE_NAME) != None: + print(IMAGE_NAME,'image already exists') + else: + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-x86_64-disk.img') + + IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution ARM' + print('Creating image:',IMAGE_NAME,'...') + if conn.get_image(IMAGE_NAME) != None: + print(IMAGE_NAME,'image already exists') + else: + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-arm-disk.img') + + IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution AARCH64' + print('Creating image:',IMAGE_NAME,'...') + if conn.get_image(IMAGE_NAME) != None: + print(IMAGE_NAME,'image already exists') + else: + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-aarch64-disk.img') + + IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus x86' + print('Creating image:',IMAGE_NAME,'...') + if conn.get_image(IMAGE_NAME) != None: + print(IMAGE_NAME,'image already exists') + else: + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'xenial-server-cloudimg-amd64-disk1.img') - IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution' + IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus ARM64' print('Creating image:',IMAGE_NAME,'...') if conn.get_image(IMAGE_NAME) != None: print(IMAGE_NAME,'image already exists') else: - conn.create_image(IMAGE_NAME, filename='cirros-0.4.0-x86_64-disk.img') + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'xenial-server-cloudimg-arm64-disk1.img') - IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus' + IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr x86' print('Creating image:',IMAGE_NAME,'...') if conn.get_image(IMAGE_NAME) != None: print(IMAGE_NAME,'image already exists') else: - conn.create_image(IMAGE_NAME, filename='xenial-server-cloudimg-amd64-disk1.img') + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'trusty-server-cloudimg-amd64-disk1.img') + # End section with Connection.create_image() - IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr' + IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr ARM64' print('Creating image:',IMAGE_NAME,'...') if conn.get_image(IMAGE_NAME) != None: print(IMAGE_NAME,'image already exists') else: - conn.create_image(IMAGE_NAME, filename='trusty-server-cloudimg-amd64-disk1.img') + conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'trusty-server-cloudimg-arm64-disk1.img') + # End section with Connection.create_image() # create a keypair, if needed e.g. for VNF VMs; maybe to SSH for testing @@ -813,12 +838,46 @@ def configure_all_ONAP(): print(onap_keypair.private_key) print_debug('onap_keypair:',onap_keypair) + print('\nSUMMARY:') - print('ONAP public network ID:',public_network.id) - print('ONAP private network ID:',private_network.id) - print('ONAP private network subnet ID:',private_subnet.id) - print('ONAP private network subnet CIDR:',private_subnet.cidr) - print('\n') + # Grab live objects (don't reuse earlier references), in case the script is used on an already configured instance + # This way, the summary is still displayed even if the script execution did not create anything + # Also, this double-checks that displayed information is accurate, freshly retrieved from the OpenStack instance + + public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME) + if public_network != None: + print('ONAP public network ID:',public_network.id) + for fetched_subnet_ID in public_network.subnet_ids: + fetched_subnet = conn.network.get_subnet(fetched_subnet_ID) + if fetched_subnet != None: + print(' ONAP public network subnet ID:',fetched_subnet.id) + print(' ONAP public network subnet CIDR:',fetched_subnet.cidr) + else: + print('no ONAP public network') + + private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME) + if private_network != None: + print('ONAP private network ID:',private_network.id) + for fetched_subnet_ID in private_network.subnet_ids: + fetched_subnet = conn.network.get_subnet(fetched_subnet_ID) + if fetched_subnet != None: + print(' ONAP private network subnet ID:',fetched_subnet.id) + print(' ONAP private network subnet CIDR:',fetched_subnet.cidr) + else: + print('no ONAP private network') + + oam_network = conn.network.find_network(ONAP_OAM_NET_NAME) + if oam_network != None: + print('ONAP OAM network ID:',oam_network.id) + for fetched_subnet_ID in oam_network.subnet_ids: + fetched_subnet = conn.network.get_subnet(fetched_subnet_ID) + if fetched_subnet != None: + print(' ONAP OAM network subnet ID:',fetched_subnet.id) + print(' ONAP OAM network subnet CIDR:',fetched_subnet.cidr) + else: + print('no ONAP OAM network') + print('END SUMMARY\n') + except Exception as e: print('*** Exception:',type(e), e) @@ -849,7 +908,7 @@ def main(): help = 'delete ONAP configuration', action = 'store_true') - # parse arguments, modify global variable if need be, and use corresponding script + # parse arguments, modify global variable if need be, and use corresponding script (create objects, or delete objects) args = parser.parse_args() if args.debug: global DEBUG_VAR @@ -862,4 +921,3 @@ def main(): if __name__ == "__main__": main() - diff --git a/setup/VIMs/OpenStack/clouds.yaml b/setup/VIMs/OpenStack/clouds.yaml new file mode 100644 index 0000000..7bfd717 --- /dev/null +++ b/setup/VIMs/OpenStack/clouds.yaml @@ -0,0 +1,99 @@ +clouds: + + # Openstack instance on Arm pod, controller IP@ 172.16.10.10 + # Horizon: https://10.10.50.103/project/ + # Identity API according to Horizon dashboard: https://10.10.50.103:5000/v2.0 + # other potential auth_url: http://172.16.10.10:35357/v3 + # (OS_AUTH_URL=http://controller:35357/v3) + # 2 project names: admin, service (project = tenant) + # project ID: 122caf64b3df4818bf2ce5ba793226b2 + # EC2 URL: https://10.10.50.103:8773/services/Cloud + # EC2 access key: bcf3c69a7d1c405e9757f87f26faf19f + # 10.10.50.0/8: floating IP@ + # 10.10.10.0/8: fixed IP@ + armopenstack: + auth: + auth_url: https://10.10.50.103:5000/v2.0 + project_name: admin + username: admin + password: opnfv_secret + region_name: RegionOne + + # Openstack instance on LaaS hpe16, from OPNFV Euphrates, controller IP@ (mgt: 172.16.10.101; public: 10.16.0.101) + # keystone endpoints (openstack endpoint list --service keystone) + # admin: http://172.16.10.101:35357/v2.0 + # internal: http://172.16.10.101:5000/v2.0 + # public: http://10.16.0.101:5000/v2.0 : works on LaaS hpe16, from hpe16 + hpe16openstackEuphrates: + auth: + auth_url: http://10.16.0.101:5000/v2.0 + project_name: admin + username: admin + password: opnfv_secret + region_name: RegionOne + + # Openstack instance on generic LaaS hpe, from OPNFV Fraser, controller IP@ (mgt: 172.16.10.36; public: 10.16.0.107) + # keystone endpoints (openstack endpoint list --service keystone) + # admin: http://172.16.10.36:35357/v3 + # internal: http://172.16.10.36:5000/v3 + # public: http://10.16.0.107:5000/v3 + # Horizon: https://10.16.0.107:8078, but need SSH port forwarding through 10.10.100.26 to be reached from outside + # "If you are using Identity v3 you need to specify the user and the project domain name" + + # generic cloud name, for a UNH IOL hpe server, for OPNFV Fraser, OpenStack installed by Fuel/MCP + unh-hpe-openstack-fraser: + auth: + auth_url: http://10.16.0.107:5000/v3 + project_name: admin + username: admin + password: opnfv_secret + user_domain_name: Default + project_domain_name: Default + region_name: RegionOne + identity_api_version: 3 + +# ubuntu@ctl01:~$ openstack project show admin +# +-------------+----------------------------------+ +# | Field | Value | +# +-------------+----------------------------------+ +# | description | OpenStack Admin tenant | +# | domain_id | default | +# | enabled | True | +# | id | 04fcfe7aa83f4df79ae39ca748aa8637 | +# | is_domain | False | +# | name | admin | +# | parent_id | default | +# +-------------+----------------------------------+ + +# (openstack) domain show default +# +-------------+----------------------------------------------------------+ +# | Field | Value | +# +-------------+----------------------------------------------------------+ +# | description | Domain created automatically to support V2.0 operations. | +# | enabled | True | +# | id | default | +# | name | Default | +# +-------------+----------------------------------------------------------+ + +# (openstack) domain show heat_user_domain +# +-------------+---------------------------------------------+ +# | Field | Value | +# +-------------+---------------------------------------------+ +# | description | Contains users and projects created by heat | +# | enabled | True | +# | id | d9c29adac0fe4816922d783b257879d6 | +# | name | heat_user_domain | +# +-------------+---------------------------------------------+ + + +# export OS_AUTH_URL=http://10.16.0.107:5000/v3 +# export OS_PROJECT_ID=04fcfe7aa83f4df79ae39ca748aa8637 +# export OS_PROJECT_NAME="admin" +# export OS_USER_DOMAIN_NAME="Default" +# export OS_USERNAME="admin" +# export OS_PASSWORD="opnfv_secret" +# export OS_REGION_NAME="RegionOne" +# export OS_INTERFACE=public +# export OS_IDENTITY_API_VERSION=3 + + diff --git a/yamllintrc b/yamllintrc new file mode 100644 index 0000000..a4f3d02 --- /dev/null +++ b/yamllintrc @@ -0,0 +1,25 @@ +# Copyright 2018 Tieto +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +extends: relaxed + +rules: + empty-lines: + max-start: 1 + max-end: 1 + colons: + max-spaces-after: 1 + max-spaces-before: 1 + line-length: + max: 160 |