summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--INFO.yaml4
-rwxr-xr-xcheck223
-rwxr-xr-xci/build-auto.sh19
-rw-r--r--pylintrc561
-rw-r--r--requirements.txt2
-rw-r--r--setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py602
-rw-r--r--setup/VIMs/OpenStack/clouds.yaml99
-rw-r--r--yamllintrc25
8 files changed, 1257 insertions, 278 deletions
diff --git a/INFO.yaml b/INFO.yaml
index 00dbfda..f4e1a04 100644
--- a/INFO.yaml
+++ b/INFO.yaml
@@ -66,11 +66,11 @@ committers:
email: 'paul.vaduva@enea.com'
company: 'enea.com'
id: 'pvaduva'
- - name: 'Martin Klozik'
+ - name: 'Martin Klozik'
email: 'martin.klozik@tieto.com'
company: 'tieto.com'
id: 'mklozik'
- - name: 'Richard Elias'
+ - name: 'Richard Elias'
email: 'richard.elias@tieto.com'
company: 'tieto.com'
id: 'richardxelias'
diff --git a/check b/check
new file mode 100755
index 0000000..0428fa6
--- /dev/null
+++ b/check
@@ -0,0 +1,223 @@
+#!/bin/bash
+
+# Copyright 2017-2018 Intel Corporation, Tieto
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+# Auto project python & yaml formatting checker
+# This script is based on the file ./check from OPNFV vswitchperf
+# project.
+
+#
+# Configuration
+#
+PYLINT="pylint"
+PYLINT_RC='pylintrc'
+PYTHON_FILE_REGEX="\.py$"
+YAMLLINT="yamllint"
+YAMLLINT_RC='yamllintrc'
+YAML_FILE_REGEX="\.yaml$"
+FILE_LIST="/tmp/auto_check_list.txt"
+
+CHECK_PYTHON=0
+CHECK_YAML=0
+
+#
+# Support Functions
+#
+# print usage if requested
+function usage() {
+ cat <<EOM
+Usage: $0 [TARGET]...
+
+Performs code check for defined TARGETs. Target can be file or directory.
+In case that directory is specified, then it will be searched recursively
+for all python and yaml files.
+If TARGET is not specified, then all python and yaml files from current AUTO
+repository will be checked.
+
+
+ -h, --help Script usage
+ -b, --black Suppress colours. Output will be black&white.
+ -m, --modified Script will check python and yaml files, which have
+ been modified within current repository.
+
+Examples:
+ ./check
+
+ Check all python and yaml files in current AUTO repository
+
+ ./check INFO.yaml
+
+ Check just one file.
+
+ ./check -m
+
+ Check all modified files in current AUTO repository
+
+ ./check lib/auto/testcase lib/auto/util
+
+ Check all python and yaml files in given directories
+
+EOM
+}
+
+# get list of files to be checked
+function get_file_list() {
+ # store file regex and shift params to get list of original ./check options
+ TMP_FILE_REGEX=$1
+ shift
+
+ rm $FILE_LIST &> /dev/null
+ if [ "x$1" == "x-m" -o "x$1" == "x--modified" ] ; then
+ # check of modified files requested
+ git status --porcelain | cut -b4- | egrep -i "${TMP_FILE_REGEX}" | sort > $FILE_LIST
+ elif [ "x$*" == "x" ] ; then
+ # list is empty, check all python files
+ git ls-tree --name-only -r HEAD | egrep -i "${TMP_FILE_REGEX}" | sort > $FILE_LIST
+ else
+ for item in $* ; do
+ if [ -d $item ] ; then
+ git ls-tree --name-only -r HEAD $item | egrep -i "${TMP_FILE_REGEX}" | sort >> $FILE_LIST
+ elif [ -f $item ] ; then
+ echo $item | egrep -i "${TMP_FILE_REGEX}" >> $FILE_LIST
+ else
+ echo "$item doesn't exist, thus check was aborted"
+ exit 1
+ fi
+ done
+ fi
+}
+
+function check_lint_binary() {
+ # check if lint binary is available
+ if ! which $1 &>/dev/null ; then
+ echo "$1 is not available, thus check can't be executed"
+ return 1
+ fi
+ return 0
+}
+
+
+function check_python() {
+ echo "Execution of pylint checks:"
+
+ if ! check_lint_binary $PYLINT ; then
+ CHECK_PYTHON=1
+ return
+ fi
+
+ # check if there is anything to check
+ if [ -s $FILE_LIST ] ; then
+ for pyfile in `cat $FILE_LIST | sort` ; do
+ # get base name
+ pyfile_basename="'"`basename $pyfile .py`"'"
+ # run pylint and extract final rating
+ output=`$PYLINT --rcfile $PYLINT_RC $pyfile 2>/dev/null`
+ rating=`echo -e $output | tail -n3 | grep rated | sed -e 's/^.*rated at \(-\?[0-9.]*\).*$/\1/'`
+ # evaluate and display aquired rating
+ if [ "x$rating" == "x" ] ; then
+ # rating is not available for files without python statements
+ printf " %-70s %-6s\n" $pyfile "NA"
+ elif [ "$rating" == "10" ] ; then
+ printf " %-70s ${GREEN}%-6s${BLACK}\n" $pyfile "OK"
+ else
+ CHECK_PYTHON=1
+ echo -e "$output" | awk '/^\*+ Module|^[A-Z]\:/'
+ printf " %-70s ${RED}%-6s${BLACK}\n" $pyfile $rating
+ fi
+ done
+ else
+ echo " Nothing to check."
+ fi
+}
+
+function check_yaml() {
+ echo "Execution of yaml checks:"
+
+ if ! check_lint_binary $YAMLLINT ; then
+ CHECK_YAML=1
+ return
+ fi
+
+ # check if there is anything to check
+ if [ -s $FILE_LIST ] ; then
+ for yamlfile in `cat $FILE_LIST | sort` ; do
+ output=`$YAMLLINT -c $YAMLLINT_RC $yamlfile 2>/dev/null`
+ if [ $? -eq 0 ] ; then
+ printf " %-70s ${GREEN}%-6s${BLACK}\n" $yamlfile "OK"
+ else
+ CHECK_YAML=1
+ echo "$output"
+ printf " %-70s ${RED}%-6s${BLACK}\n" $yamlfile "FAILED"
+ fi
+ done
+ else
+ echo " Nothing to check."
+ fi
+}
+
+#
+# Main
+#
+# check if help is requested
+if [ "x$1" == "x-h" -o "x$1" == "x--help" ] ; then
+ usage
+ exit 0
+fi
+
+# set colours
+if [ "x$1" == "x-b" -o "x$1" == "x--black" ] ; then
+ shift
+ RED=""
+ GREEN=""
+ BLACK=""
+else
+ RED="\e[31m"
+ GREEN="\e[32m"
+ BLACK="\e[0m"
+fi
+
+# check if we were run within auto directory
+if [ ! -x ./check 2> /dev/null ] ; then
+ echo "`basename $0` must be run from auto root directory"
+ exit 1
+fi
+
+# run python checks
+get_file_list $PYTHON_FILE_REGEX $*
+check_python
+
+echo
+
+# run yaml checks
+get_file_list $YAML_FILE_REGEX $*
+check_yaml
+
+# clean up
+rm $FILE_LIST &> /dev/null
+
+# return success or failure based on pylint and yamllint checks
+# NOTE: As of now, failure of pylint checks is not propagated into exit code.
+# This will be turned on again after the rating of existing python
+# files will be improved.
+# if [ $CHECK_PYTHON -eq 0 -a $CHECK_YAML -eq 0 ] ; then
+if [ $CHECK_YAML -eq 0 ] ; then
+ exit 0
+else
+ exit 1
+fi
+
+#
+# The End
+#
diff --git a/ci/build-auto.sh b/ci/build-auto.sh
index 611d83a..96588b9 100755
--- a/ci/build-auto.sh
+++ b/ci/build-auto.sh
@@ -30,6 +30,7 @@
#
EXIT=0
EXIT_UNKNOWN_JOB_TYPE=1
+EXIT_LINT_FAILED=2
#
# configuration
@@ -37,6 +38,16 @@ EXIT_UNKNOWN_JOB_TYPE=1
AUTOENV_DIR="$HOME/autoenv"
#
+# functions
+#
+# execute pylint and yamllint to check code quality
+function execute_auto_lint_check() {
+ if ! ./check -b ; then
+ EXIT=$EXIT_LINT_FAILED
+ fi
+}
+
+#
# main
#
echo
@@ -69,7 +80,7 @@ case $1 in
# Example of verify job body. Functions can call
# external scripts, etc.
- #execute_auto_pylint_check
+ execute_auto_lint_check
#execute_auto_doc_check
#install_opnfv MCP
#install_onap
@@ -87,14 +98,14 @@ case $1 in
# Example of merge job body. Functions can call
# external scripts, etc.
- #execute_auto_pylint_check
+ execute_auto_lint_check
#execute_auto_doc_check
#install_opnfv MCP
#install_onap
#execute_sanity_check
#execute_tests $1
- # Everything went well, so report SUCCESS to Jenkins
+ # propagate result to the Jenkins job
exit $EXIT
;;
"daily")
@@ -111,7 +122,7 @@ case $1 in
#execute_tests $1
#push_results_and_logs_to_artifactory
- # Everything went well, so report SUCCESS to Jenkins
+ # propagate result to the Jenkins job
exit $EXIT
;;
*)
diff --git a/pylintrc b/pylintrc
new file mode 100644
index 0000000..c213b80
--- /dev/null
+++ b/pylintrc
@@ -0,0 +1,561 @@
+# Copyright 2018 Tieto
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+[MASTER]
+
+# A comma-separated list of package or module names from where C extensions may
+# be loaded. Extensions are loading into the active Python interpreter and may
+# run arbitrary code
+extension-pkg-whitelist=
+
+# Add files or directories to the blacklist. They should be base names, not
+# paths.
+ignore=CVS
+
+# Add files or directories matching the regex patterns to the blacklist. The
+# regex matches against base names, not paths.
+ignore-patterns=
+
+# Python code to execute, usually for sys.path manipulation such as
+# pygtk.require().
+#init-hook=
+
+# Use multiple processes to speed up Pylint.
+jobs=1
+
+# List of plugins (as comma separated values of python modules names) to load,
+# usually to register additional checkers.
+load-plugins=
+
+# Pickle collected data for later comparisons.
+persistent=yes
+
+# Specify a configuration file.
+#rcfile=
+
+# When enabled, pylint would attempt to guess common misconfiguration and emit
+# user-friendly hints instead of false-positive error messages
+suggestion-mode=yes
+
+# Allow loading of arbitrary C extensions. Extensions are imported into the
+# active Python interpreter and may run arbitrary code.
+unsafe-load-any-extension=no
+
+
+[MESSAGES CONTROL]
+
+# Only show warnings with the listed confidence levels. Leave empty to show
+# all. Valid levels: HIGH, INFERENCE, INFERENCE_FAILURE, UNDEFINED
+confidence=
+
+# Disable the message, report, category or checker with the given id(s). You
+# can either give multiple identifiers separated by comma (,) or put this
+# option multiple times (only on the command line, not in the configuration
+# file where it should appear only once).You can also use "--disable=all" to
+# disable everything first and then reenable specific checks. For example, if
+# you want to run only the similarities checker, you can use "--disable=all
+# --enable=similarities". If you want to run only the classes checker, but have
+# no Warning level messages displayed, use"--disable=all --enable=classes
+# --disable=W"
+disable=print-statement,
+ parameter-unpacking,
+ unpacking-in-except,
+ old-raise-syntax,
+ backtick,
+ long-suffix,
+ old-ne-operator,
+ old-octal-literal,
+ import-star-module-level,
+ non-ascii-bytes-literal,
+ invalid-unicode-literal,
+ raw-checker-failed,
+ bad-inline-option,
+ locally-disabled,
+ locally-enabled,
+ file-ignored,
+ suppressed-message,
+ useless-suppression,
+ deprecated-pragma,
+ apply-builtin,
+ basestring-builtin,
+ buffer-builtin,
+ cmp-builtin,
+ coerce-builtin,
+ execfile-builtin,
+ file-builtin,
+ long-builtin,
+ raw_input-builtin,
+ reduce-builtin,
+ standarderror-builtin,
+ unicode-builtin,
+ xrange-builtin,
+ coerce-method,
+ delslice-method,
+ getslice-method,
+ setslice-method,
+ no-absolute-import,
+ old-division,
+ dict-iter-method,
+ dict-view-method,
+ next-method-called,
+ metaclass-assignment,
+ indexing-exception,
+ raising-string,
+ reload-builtin,
+ oct-method,
+ hex-method,
+ nonzero-method,
+ cmp-method,
+ input-builtin,
+ round-builtin,
+ intern-builtin,
+ unichr-builtin,
+ map-builtin-not-iterating,
+ zip-builtin-not-iterating,
+ range-builtin-not-iterating,
+ filter-builtin-not-iterating,
+ using-cmp-argument,
+ eq-without-hash,
+ div-method,
+ idiv-method,
+ rdiv-method,
+ exception-message-attribute,
+ invalid-str-codec,
+ sys-max-int,
+ bad-python3-import,
+ deprecated-string-function,
+ deprecated-str-translate-call,
+ deprecated-itertools-function,
+ deprecated-types-field,
+ next-method-defined,
+ dict-items-not-iterating,
+ dict-keys-not-iterating,
+ dict-values-not-iterating,
+ deprecated-operator-function,
+ deprecated-urllib-function,
+ xreadlines-attribute,
+ deprecated-sys-function,
+ exception-escape,
+ comprehension-escape
+
+# Enable the message, report, category or checker with the given id(s). You can
+# either give multiple identifier separated by comma (,) or put this option
+# multiple time (only on the command line, not in the configuration file where
+# it should appear only once). See also the "--disable" option for examples.
+enable=c-extension-no-member
+
+
+[REPORTS]
+
+# Python expression which should return a note less than 10 (10 is the highest
+# note). You have access to the variables errors warning, statement which
+# respectively contain the number of errors / warnings messages and the total
+# number of statements analyzed. This is used by the global evaluation report
+# (RP0004).
+evaluation=10.0 - ((float(5 * error + warning + refactor + convention) / statement) * 10)
+
+# Template used to display messages. This is a python new-style format string
+# used to format the message information. See doc for all details
+#msg-template=
+
+# Set the output format. Available formats are text, parseable, colorized, json
+# and msvs (visual studio).You can also give a reporter class, eg
+# mypackage.mymodule.MyReporterClass.
+output-format=text
+
+# Tells whether to display a full report or only the messages
+reports=no
+
+# Activate the evaluation score.
+score=yes
+
+
+[REFACTORING]
+
+# Maximum number of nested blocks for function / method body
+max-nested-blocks=5
+
+# Complete name of functions that never returns. When checking for
+# inconsistent-return-statements if a never returning function is called then
+# it will be considered as an explicit return statement and no message will be
+# printed.
+never-returning-functions=optparse.Values,sys.exit
+
+
+[FORMAT]
+
+# Expected format of line ending, e.g. empty (any line ending), LF or CRLF.
+expected-line-ending-format=
+
+# Regexp for a line that is allowed to be longer than the limit.
+ignore-long-lines=^\s*(# )?<?https?://\S+>?$
+
+# Number of spaces of indent required inside a hanging or continued line.
+indent-after-paren=4
+
+# String used as indentation unit. This is usually " " (4 spaces) or "\t" (1
+# tab).
+indent-string=' '
+
+# Maximum number of characters on a single line.
+max-line-length=160
+
+# Maximum number of lines in a module
+max-module-lines=1000
+
+# List of optional constructs for which whitespace checking is disabled. `dict-
+# separator` is used to allow tabulation in dicts, etc.: {1 : 1,\n222: 2}.
+# `trailing-comma` allows a space between comma and closing bracket: (a, ).
+# `empty-line` allows space-only lines.
+no-space-check=trailing-comma,
+ dict-separator
+
+# Allow the body of a class to be on the same line as the declaration if body
+# contains single statement.
+single-line-class-stmt=no
+
+# Allow the body of an if to be on the same line as the test if there is no
+# else.
+single-line-if-stmt=no
+
+
+[BASIC]
+
+# Naming style matching correct argument names
+argument-naming-style=snake_case
+
+# Regular expression matching correct argument names. Overrides argument-
+# naming-style
+#argument-rgx=
+
+# Naming style matching correct attribute names
+attr-naming-style=snake_case
+
+# Regular expression matching correct attribute names. Overrides attr-naming-
+# style
+#attr-rgx=
+
+# Bad variable names which should always be refused, separated by a comma
+bad-names=foo,
+ bar,
+ baz,
+ toto,
+ tutu,
+ tata
+
+# Naming style matching correct class attribute names
+class-attribute-naming-style=any
+
+# Regular expression matching correct class attribute names. Overrides class-
+# attribute-naming-style
+#class-attribute-rgx=
+
+# Naming style matching correct class names
+class-naming-style=PascalCase
+
+# Regular expression matching correct class names. Overrides class-naming-style
+#class-rgx=
+
+# Naming style matching correct constant names
+const-naming-style=UPPER_CASE
+
+# Regular expression matching correct constant names. Overrides const-naming-
+# style
+#const-rgx=
+
+# Minimum line length for functions/classes that require docstrings, shorter
+# ones are exempt.
+docstring-min-length=-1
+
+# Naming style matching correct function names
+function-naming-style=snake_case
+
+# Regular expression matching correct function names. Overrides function-
+# naming-style
+#function-rgx=
+
+# Good variable names which should always be accepted, separated by a comma
+good-names=i,
+ j,
+ k,
+ e,
+ ex,
+ Run,
+ _
+
+# Include a hint for the correct naming format with invalid-name
+include-naming-hint=no
+
+# Naming style matching correct inline iteration names
+inlinevar-naming-style=any
+
+# Regular expression matching correct inline iteration names. Overrides
+# inlinevar-naming-style
+#inlinevar-rgx=
+
+# Naming style matching correct method names
+method-naming-style=snake_case
+
+# Regular expression matching correct method names. Overrides method-naming-
+# style
+#method-rgx=
+
+# Naming style matching correct module names
+module-naming-style=snake_case
+
+# Regular expression matching correct module names. Overrides module-naming-
+# style
+#module-rgx=
+
+# Colon-delimited sets of names that determine each other's naming style when
+# the name regexes allow several styles.
+name-group=
+
+# Regular expression which should only match function or class names that do
+# not require a docstring.
+no-docstring-rgx=^_
+
+# List of decorators that produce properties, such as abc.abstractproperty. Add
+# to this list to register other decorators that produce valid properties.
+property-classes=abc.abstractproperty
+
+# Naming style matching correct variable names
+variable-naming-style=snake_case
+
+# Regular expression matching correct variable names. Overrides variable-
+# naming-style
+#variable-rgx=
+
+
+[MISCELLANEOUS]
+
+# List of note tags to take in consideration, separated by a comma.
+notes=FIXME,
+ XXX,
+ TODO
+
+
+[TYPECHECK]
+
+# List of decorators that produce context managers, such as
+# contextlib.contextmanager. Add to this list to register other decorators that
+# produce valid context managers.
+contextmanager-decorators=contextlib.contextmanager
+
+# List of members which are set dynamically and missed by pylint inference
+# system, and so shouldn't trigger E1101 when accessed. Python regular
+# expressions are accepted.
+generated-members=
+
+# Tells whether missing members accessed in mixin class should be ignored. A
+# mixin class is detected if its name ends with "mixin" (case insensitive).
+ignore-mixin-members=yes
+
+# This flag controls whether pylint should warn about no-member and similar
+# checks whenever an opaque object is returned when inferring. The inference
+# can return multiple potential results while evaluating a Python object, but
+# some branches might not be evaluated, which results in partial inference. In
+# that case, it might be useful to still emit no-member and other checks for
+# the rest of the inferred objects.
+ignore-on-opaque-inference=yes
+
+# List of class names for which member attributes should not be checked (useful
+# for classes with dynamically set attributes). This supports the use of
+# qualified names.
+ignored-classes=optparse.Values,thread._local,_thread._local
+
+# List of module names for which member attributes should not be checked
+# (useful for modules/projects where namespaces are manipulated during runtime
+# and thus existing member attributes cannot be deduced by static analysis. It
+# supports qualified module names, as well as Unix pattern matching.
+ignored-modules=
+
+# Show a hint with possible names when a member name was not found. The aspect
+# of finding the hint is based on edit distance.
+missing-member-hint=yes
+
+# The minimum edit distance a name should have in order to be considered a
+# similar match for a missing member name.
+missing-member-hint-distance=1
+
+# The total number of similar names that should be taken in consideration when
+# showing a hint for a missing member.
+missing-member-max-choices=1
+
+
+[LOGGING]
+
+# Logging modules to check that the string format arguments are in logging
+# function parameter format
+logging-modules=logging
+
+
+[SIMILARITIES]
+
+# Ignore comments when computing similarities.
+ignore-comments=yes
+
+# Ignore docstrings when computing similarities.
+ignore-docstrings=yes
+
+# Ignore imports when computing similarities.
+ignore-imports=no
+
+# Minimum lines number of a similarity.
+min-similarity-lines=4
+
+
+[SPELLING]
+
+# Limits count of emitted suggestions for spelling mistakes
+max-spelling-suggestions=4
+
+# Spelling dictionary name. Available dictionaries: none. To make it working
+# install python-enchant package.
+spelling-dict=
+
+# List of comma separated words that should not be checked.
+spelling-ignore-words=
+
+# A path to a file that contains private dictionary; one word per line.
+spelling-private-dict-file=
+
+# Tells whether to store unknown words to indicated private dictionary in
+# --spelling-private-dict-file option instead of raising a message.
+spelling-store-unknown-words=no
+
+
+[VARIABLES]
+
+# List of additional names supposed to be defined in builtins. Remember that
+# you should avoid to define new builtins when possible.
+additional-builtins=
+
+# Tells whether unused global variables should be treated as a violation.
+allow-global-unused-variables=yes
+
+# List of strings which can identify a callback function by name. A callback
+# name must start or end with one of those strings.
+callbacks=cb_,
+ _cb
+
+# A regular expression matching the name of dummy variables (i.e. expectedly
+# not used).
+dummy-variables-rgx=_+$|(_[a-zA-Z0-9_]*[a-zA-Z0-9]+?$)|dummy|^ignored_|^unused_
+
+# Argument names that match this expression will be ignored. Default to name
+# with leading underscore
+ignored-argument-names=_.*|^ignored_|^unused_
+
+# Tells whether we should check for unused import in __init__ files.
+init-import=no
+
+# List of qualified module names which can have objects that can redefine
+# builtins.
+redefining-builtins-modules=six.moves,past.builtins,future.builtins,io,builtins
+
+
+[DESIGN]
+
+# Maximum number of arguments for function / method
+max-args=5
+
+# Maximum number of attributes for a class (see R0902).
+max-attributes=7
+
+# Maximum number of boolean expressions in a if statement
+max-bool-expr=5
+
+# Maximum number of branch for function / method body
+max-branches=12
+
+# Maximum number of locals for function / method body
+max-locals=15
+
+# Maximum number of parents for a class (see R0901).
+max-parents=7
+
+# Maximum number of public methods for a class (see R0904).
+max-public-methods=20
+
+# Maximum number of return / yield for function / method body
+max-returns=6
+
+# Maximum number of statements in function / method body
+max-statements=50
+
+# Minimum number of public methods for a class (see R0903).
+min-public-methods=2
+
+
+[CLASSES]
+
+# List of method names used to declare (i.e. assign) instance attributes.
+defining-attr-methods=__init__,
+ __new__,
+ setUp
+
+# List of member names, which should be excluded from the protected access
+# warning.
+exclude-protected=_asdict,
+ _fields,
+ _replace,
+ _source,
+ _make
+
+# List of valid names for the first argument in a class method.
+valid-classmethod-first-arg=cls
+
+# List of valid names for the first argument in a metaclass class method.
+valid-metaclass-classmethod-first-arg=mcs
+
+
+[IMPORTS]
+
+# Allow wildcard imports from modules that define __all__.
+allow-wildcard-with-all=no
+
+# Analyse import fallback blocks. This can be used to support both Python 2 and
+# 3 compatible code, which means that the block might have code that exists
+# only in one or another interpreter, leading to false positives when analysed.
+analyse-fallback-blocks=no
+
+# Deprecated modules which should not be used, separated by a comma
+deprecated-modules=optparse,tkinter.tix
+
+# Create a graph of external dependencies in the given file (report RP0402 must
+# not be disabled)
+ext-import-graph=
+
+# Create a graph of every (i.e. internal and external) dependencies in the
+# given file (report RP0402 must not be disabled)
+import-graph=
+
+# Create a graph of internal dependencies in the given file (report RP0402 must
+# not be disabled)
+int-import-graph=
+
+# Force import order to recognize a module as part of the standard
+# compatibility libraries.
+known-standard-library=
+
+# Force import order to recognize a module as part of a third party library.
+known-third-party=enchant
+
+
+[EXCEPTIONS]
+
+# Exceptions that will emit a warning when being caught. Defaults to
+# "Exception"
+overgeneral-exceptions=Exception
diff --git a/requirements.txt b/requirements.txt
index 8ef8db6..1035c76 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -6,3 +6,5 @@ python-glanceclient>=2.8.0
python-neutronclient>=6.3.0
python-novaclient>=9.0.0
python-heatclient>=1.6.1
+pylint==1.9.2
+yamllint==1.11.1
diff --git a/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py b/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py
index bab2c59..e4b94f5 100644
--- a/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py
+++ b/setup/VIMs/OpenStack/auto_script_config_openstack_for_onap.py
@@ -24,13 +24,15 @@
# https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095
#docstring
-"""This script configures an OpenStack instance to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP.
+"""This script configures an OpenStack instance to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP. It can also remove the created objects, when used in a clean-up procedure.
+Use -h option to see usage (-del to delete objects, -deb to print debug information).
+Requirements: python 3, OpenStack SDK (0.14 or greater), clouds.yaml file, .img files are downloaded
Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095
"""
######################################################################
-# This script configures an OpenStack instance (e.g. from an OPNFV installer like FUEL/MCP) to make it
-# ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP.
+# This script configures an OpenStack instance (e.g. from an OPNFV installer like FUEL/MCP, Compass4nfv, ...)
+# to make it ready to interface with an ONAP instance, for example to host VM-based VNFs deployed by ONAP.
# After running this script, the created OpenStack object names/IDs can be used for example to populate
# YAML&ENV files used by ONAP (installation of ONAP itself, VNF descriptor files, etc.).
@@ -42,12 +44,15 @@ Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095
# (optional, probably not needed: create a new group, which can be associated to a project, and contains users)
# 2) create an ONAP user within the ONAP project, so as not to use the "admin" user for ONAP
# (associate user to group if applicable; credentials: name/pwd or name/APIkey, or token)
-# 3) create an ONAP security group, to allow ICMP traffic (for pings) and TCP port 22 (for SSH), rather than changing default security group(s)
-# (optional, probably not needed: create new region; default region RegionOne is OK)
+# 3) create an ONAP security group, to allow ICMP traffic (for pings) and TCP port 22 (for SSH),
+# rather than changing default security group(s)
+# (optional, probably not needed: create a new region; default region RegionOne is OK)
# 4) create a public network for ONAP VNFs, with subnet and CIDR block
# (so components have access to the Internet, via router and gateway, on unnamed ports, dynamic IP@ allocation)
-# 5) create a private and an OAM network for ONAP VNFs or other ONAP components, with their respective subnet and CIDR block
-# (ONAP VNFs will be deployed in this private and/or OAM network(s), usually with named ports and static IP@ as per VNF configuration file)
+# 5) create a private and an OAM network for ONAP VNFs or other ONAP components,
+# with their respective subnet and CIDR block
+# (ONAP VNFs will be deployed in this private and/or OAM network(s), usually with named ports
+# and static IP@ as per VNF configuration file)
# 6) create an OpenStack router, with interfaces to the public, private and OAM networks,
# and a reference to an external network (gateway) provided by the OpenStack instance installation
# 7) create VM flavors as needed: m1.medium, etc.
@@ -59,11 +64,13 @@ Auto project: https://wiki.opnfv.org/pages/viewpage.action?pageId=12389095
# - python3 is installed
# - OpenStack SDK is installed for python3
# - there is a clouds.yaml file (describing the OpenStack instance, especially the Auth URL and admin credentials)
+# - .img files (Ubuntu Trusty Tahr, Xenial Xerus, Cirros, ... are downloaded, and stored in IMAGES_DIR
# - the script connects to OpenStack as a user with admin rights
# typical commands to install OpenStack SDK Python client:
# apt install python3-pip
# pip3 install --upgrade pip
+# hash -r
# pip3 list
# pip3 install openstacksdk
# pip3 install --upgrade openstacksdk
@@ -85,7 +92,8 @@ ONAP_USER_NAME = 'ONAP_user'
ONAP_USER_PASSWORD = 'auto_topsecret'
ONAP_USER_DESC = 'OpenStack User created for ONAP'
-ONAP_TENANT_NAME = 'ONAP_tenant' # "project" is a more generic concept than "tenant"; a tenant is type of project; quotas are per project;
+ONAP_TENANT_NAME = 'ONAP_tenant'
+# note: "project" is a more generic concept than "tenant"; a tenant is type of project; quotas are per project;
ONAP_TENANT_DESC = 'OpenStack Project/Tenant created for ONAP'
ONAP_SECU_GRP_NAME = 'ONAP_security_group'
@@ -93,30 +101,35 @@ ONAP_SECU_GRP_DESC = 'Security Group created for ONAP'
ONAP_PUBLIC_NET_NAME = 'ONAP_public_net'
ONAP_PUBLIC_SUBNET_NAME = 'ONAP_public_subnet'
-ONAP_PUBLIC_SUBNET_CIDR = '192.168.99.0/24' # some arbitrary CIDR, but typically in a private (IANA-reserved) address range
+ONAP_PUBLIC_SUBNET_CIDR = '192.168.99.0/24'
+# note: some arbitrary CIDR, but typically in a private (IANA-reserved) address range
ONAP_PUBLIC_NET_DESC = 'Public network created for ONAP, for unnamed ports, dynamic IP@, access to the Internet (e.g., Nexus repo) via Gateway'
ONAP_PRIVATE_NET_NAME = 'ONAP_private_net'
ONAP_PRIVATE_SUBNET_NAME = 'ONAP_private_subnet'
-ONAP_PRIVATE_SUBNET_CIDR = '10.0.0.0/16' # should match ONAP installation; Private and OAM may be the same network
+ONAP_PRIVATE_SUBNET_CIDR = '10.0.0.0/16'
+# note: CIDR should match ONAP installation; Private and OAM may be the same network
ONAP_PRIVATE_NET_DESC = 'Private network created for ONAP, for named ports, static IP@, inter-component communication'
ONAP_OAM_NET_NAME = 'ONAP_OAM_net'
ONAP_OAM_SUBNET_NAME = 'ONAP_OAM_subnet'
-ONAP_OAM_SUBNET_CIDR = '10.99.0.0/16' # should match ONAP installation; Private and OAM may be the same network
+ONAP_OAM_SUBNET_CIDR = '10.99.0.0/16'
+# note: CIDR should match ONAP installation; Private and OAM may be the same network
ONAP_OAM_NET_DESC = 'OAM network created for ONAP, for named ports, static IP@, inter-component communication'
ONAP_ROUTER_NAME = 'ONAP_router'
ONAP_ROUTER_DESC = 'Router created for ONAP'
-EXTERNAL_NETWORK_NAME = 'floating_net' # OpenStack instance external network (gateway) name to be used as router's gateway
+# OpenStack instance external network (gateway) name to be used as router's gateway
+EXTERNAL_NETWORK_NAME = 'floating_net'
-ONAP_KEYPAIR_NAME = 'ONAP_keypair' # keypair that can be used to SSH into created servers (VNF VMs)
+# keypair that can be used to SSH into created servers (VNF VMs)
+ONAP_KEYPAIR_NAME = 'ONAP_keypair'
# OpenStack cloud name and region name, which should be the same as in the clouds.yaml file used by this script
-OPENSTACK_CLOUD_NAME = 'hpe16openstackFraser'
+OPENSTACK_CLOUD_NAME = 'unh-hpe-openstack-fraser'
OPENSTACK_REGION_NAME = 'RegionOne'
-# OpenStack domain is: Default
+# note: OpenStack domain is: Default
######################################################################
@@ -144,45 +157,58 @@ def print_debug(*args):
######################################################################
def delete_all_ONAP():
- """Delete all ONAP-specific OpenStack objects (normally not needed, but may be useful during tests)."""
+ """Delete all ONAP-specific OpenStack objects (normally not needed, but may be useful during tests, and for clean-up)."""
print('\nOPNFV Auto, script to delete ONAP objects in an OpenStack instance')
try:
# connect to OpenStack instance using Connection object from OpenStack SDK
print('Opening connection...')
conn = openstack.connect(
- identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml
- cloud = OPENSTACK_CLOUD_NAME,
- region_name = OPENSTACK_REGION_NAME)
+ identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml
+ cloud = OPENSTACK_CLOUD_NAME,
+ region_name = OPENSTACK_REGION_NAME)
+
# delete router; must delete router before networks (and must delete VMs before routers)
print('Deleting ONAP router...')
onap_router = conn.network.find_router(ONAP_ROUTER_NAME)
print_debug('onap_router:',onap_router)
if onap_router != None:
+
+ # delete router interfaces before deleting router
+ router_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME)
+ if router_network != None:
+ if router_network.subnet_ids != None:
+ print_debug('router_network.subnet_ids:',router_network.subnet_ids)
+ for subnet_id in router_network.subnet_ids:
+ print(' Deleting interface to',ONAP_PUBLIC_NET_NAME,'...')
+ conn.network.remove_interface_from_router(onap_router, subnet_id)
+
+ router_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME)
+ if router_network != None:
+ if router_network.subnet_ids != None:
+ print_debug('router_network.subnet_ids:',router_network.subnet_ids)
+ for subnet_id in router_network.subnet_ids:
+ print(' Deleting interface to',ONAP_PRIVATE_NET_NAME,'...')
+ conn.network.remove_interface_from_router(onap_router, subnet_id)
+
+ router_network = conn.network.find_network(ONAP_OAM_NET_NAME)
+ if router_network != None:
+ if router_network.subnet_ids != None:
+ print_debug('router_network.subnet_ids:',router_network.subnet_ids)
+ for subnet_id in router_network.subnet_ids:
+ print(' Deleting interface to',ONAP_OAM_NET_NAME,'...')
+ conn.network.remove_interface_from_router(onap_router, subnet_id)
+
+ # and finally delete ONAP router
conn.network.delete_router(onap_router.id)
+
else:
print('No ONAP router found...')
+ # TODO@@@ verify if there are ports on networks (e.g., from VMs); if yes, can't delete network
- # try: # try to circumvent !=None issue with OpenStack Resource.py; nope;
- # onap_router = conn.network.find_router(ONAP_ROUTER_NAME)
- # print_debug('onap_router:',onap_router)
- # print_debug('onap_router.name:',onap_router.name)
- # print_debug('onap_router.id:',onap_router.id)
- # conn.network.delete_router(onap_router.id)
- # conn.network.delete_router(onap_router)
- # except Exception as e:
- # print('issue Deleting ONAP router...')
- # print('*** Exception:',type(e), e)
- # exceptionType, exceptionValue, exceptionTraceback = sys.exc_info()
- # print('*** traceback.print_tb():')
- # traceback.print_tb(exceptionTraceback)
- # print('*** traceback.print_exception():')
- # traceback.print_exception(exceptionType, exceptionValue, exceptionTraceback)
-
-
- # delete private network (which should also delete associated subnet and ports if any)
+ # delete private network (which should also delete associated subnet if any)
print('Deleting ONAP private network...')
private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME)
print_debug('private_network:',private_network)
@@ -191,7 +217,7 @@ def delete_all_ONAP():
else:
print('No ONAP private network found...')
- # delete OAM network (which should also delete associated subnet and ports if any)
+ # delete OAM network (which should also delete associated subnet if any)
print('Deleting ONAP OAM network...')
oam_network = conn.network.find_network(ONAP_OAM_NET_NAME)
print_debug('oam_network:',oam_network)
@@ -200,7 +226,7 @@ def delete_all_ONAP():
else:
print('No ONAP OAM network found...')
- # delete public network (which should also delete associated subnet and ports if any)
+ # delete public network (which should also delete associated subnet if any)
print('Deleting ONAP public network...')
public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME)
print_debug('public_network:',public_network)
@@ -209,6 +235,8 @@ def delete_all_ONAP():
else:
print('No ONAP public network found...')
+ # TODO@@@ verify if security group is in use (e.g., by a VM), otherwise can't delete it
+
# delete security group
print('Deleting ONAP security group...')
onap_security_group = conn.network.find_security_group(ONAP_SECU_GRP_NAME)
@@ -245,6 +273,8 @@ def delete_all_ONAP():
else:
print('No ONAP keypair found...')
+ # no need to delete images and flavors
+
except Exception as e:
print('*** Exception:',type(e), e)
@@ -267,65 +297,9 @@ def configure_all_ONAP():
# connect to OpenStack instance using Connection object from OpenStack SDK
print('Opening connection...')
conn = openstack.connect(
- identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml
- cloud = OPENSTACK_CLOUD_NAME,
- region_name = OPENSTACK_REGION_NAME)
-
- ###################################################################
- # TESTS: IGNORE/DELETE (BEGIN)
- # gdserver_ID = '8bc274a2-8c0d-4795-9b4d-faa0a21e1d88'
- # gdserver = conn.compute.get_server(gdserver_ID)
- # print('\ngdserver.name=',gdserver.name)
- # print('gdserver.status=',gdserver.status)
-
- # print("\nList Users:")
- # i=1
- # for user in conn.identity.users():
- # print('User',str(i),user.name,'\n',user,'\n')
- # i+=1
-
- # print("\nList Projects:")
- # i=1
- # for project in conn.identity.projects():
- # print('Project',str(i),project.name,'\n',project,'\n')
- # i+=1
-
- # print("\nList Roles:")
- # i=1
- # for role in conn.identity.roles():
- # print('Role',str(i),role.name,'\n',role,'\n')
- # i+=1
-
- # print("\nList Networks:")
- # i=1
- # for network in conn.network.networks():
- # print('Network',str(i),network.name,'\n',network,'\n')
- # i+=1
-
- # print("\nList Routers:")
- # i=1
- # for router in conn.network.routers():
- # print('Router',str(i),router.name,'\n',router,'\n')
- # i+=1
-
- # print("\nList Flavors:")
- # i=1
- # for flvr in conn.compute.flavors():
- # print('Flavor',str(i),flvr.name,'\n',flvr,'\n')
- # i+=1
-
- # print("\nList Images:")
- # i=1
- # for img in conn.compute.images():
- # print('Image',str(i),img.name,'\n',img,'\n')
- # i+=1
-
- # router = conn.network.find_router('gd_test_router')
- # print('gd router\n',router,'\n\n')
- # router = conn.network.find_router('e4e59f63-8063-4774-a97a-c110c6969e4a')
- # print('gd router\n',router,'\n\n')
- # TESTS: IGNORE/DELETE (END)
- ###################################################################
+ identity_api_version = 3, # must indicate Identity version (until fixed); can also be in clouds.yaml
+ cloud = OPENSTACK_CLOUD_NAME,
+ region_name = OPENSTACK_REGION_NAME)
print('Creating ONAP project/tenant...')
@@ -396,7 +370,8 @@ def configure_all_ONAP():
port_range_min = None,
port_range_max = None)
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
try:
description_text = 'enable ICMP egress IPv4'
@@ -411,7 +386,8 @@ def configure_all_ONAP():
port_range_min = None,
port_range_max = None)
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
try:
description_text = 'enable SSH (TCP port 22) ingress IPv4'
@@ -426,7 +402,8 @@ def configure_all_ONAP():
port_range_min = '22',
port_range_max = '22')
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
try:
description_text = 'enable SSH (TCP port 22) egress IPv4'
@@ -441,7 +418,8 @@ def configure_all_ONAP():
port_range_min = '22',
port_range_max = '22')
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
try:
description_text = 'enable IP traffic ingress IPv4'
@@ -456,7 +434,8 @@ def configure_all_ONAP():
port_range_min = None,
port_range_max = None)
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
try:
description_text = 'enable IP traffic ingress IPv6'
@@ -471,44 +450,48 @@ def configure_all_ONAP():
port_range_min = None,
port_range_max = None)
except Exception as e:
- print(description_text, ' Exception:', type(e), e)
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
# IPv4 IP egress rule should already exist by default
- # try:
- # description_text = 'enable IP traffic egress IPv4'
- # print(' Creating rule:',description_text,'...')
- # conn.network.create_security_group_rule(
- # security_group_id = onap_security_group.id,
- # description = description_text,
- # protocol = None,
- # direction = 'egress',
- # ethertype = 'IPv4',
- # remote_ip_prefix = '0.0.0.0/0',
- # port_range_min = None,
- # port_range_max = None)
- # except Exception as e:
- # print(description_text, ' Exception:', type(e), e)
+ try:
+ description_text = 'enable IP traffic egress IPv4'
+ print(' Creating rule:',description_text,'...')
+ conn.network.create_security_group_rule(
+ security_group_id = onap_security_group.id,
+ description = description_text,
+ protocol = None,
+ direction = 'egress',
+ ethertype = 'IPv4',
+ remote_ip_prefix = '0.0.0.0/0',
+ port_range_min = None,
+ port_range_max = None)
+ except Exception as e:
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
# IPv6 IP egress rule should already exist by default
- # try:
- # description_text = 'enable IP traffic egress IPv6'
- # print(' Creating rule:',description_text,'...')
- # conn.network.create_security_group_rule(
- # security_group_id = onap_security_group.id,
- # description = description_text,
- # protocol = None,
- # direction = 'egress',
- # ethertype = 'IPv6',
- # remote_ip_prefix = '::/0',
- # port_range_min = None,
- # port_range_max = None)
- # except Exception as e:
- # print(description_text, ' Exception:', type(e), e)
+ try:
+ description_text = 'enable IP traffic egress IPv6'
+ print(' Creating rule:',description_text,'...')
+ conn.network.create_security_group_rule(
+ security_group_id = onap_security_group.id,
+ description = description_text,
+ protocol = None,
+ direction = 'egress',
+ ethertype = 'IPv6',
+ remote_ip_prefix = '::/0',
+ port_range_min = None,
+ port_range_max = None)
+ except Exception as e:
+ print(' rule:', description_text, ' may already exist')
+ print_debug(description_text, ' Exception:', type(e), e)
# public network
print('Creating ONAP public network...')
public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME)
+ public_subnet = None
if public_network != None:
print('ONAP public network already exists')
else:
@@ -530,13 +513,14 @@ def configure_all_ONAP():
ip_version = 4,
is_dhcp_enabled = True,
dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@
+ print_debug('public_subnet:',public_subnet)
print_debug('public_network: after subnet',public_network)
- print_debug('public_subnet:',public_subnet)
# private network
print('Creating ONAP private network...')
private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME)
+ private_subnet = None
if private_network != None:
print('ONAP private network already exists')
else:
@@ -557,13 +541,14 @@ def configure_all_ONAP():
ip_version = 4,
is_dhcp_enabled = True,
dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@; maybe not needed for private network
- print_debug('private_network: after subnet',private_network)
print_debug('private_subnet:',private_subnet)
+ print_debug('private_network: after subnet',private_network)
# OAM network
print('Creating ONAP OAM network...')
oam_network = conn.network.find_network(ONAP_OAM_NET_NAME)
+ oam_subnet = None
if oam_network != None:
print('ONAP OAM network already exists')
else:
@@ -584,8 +569,8 @@ def configure_all_ONAP():
ip_version = 4,
is_dhcp_enabled = True,
dns_nameservers = [DNS_SERVER_IP]) # list of DNS IP@; maybe not needed for OAM network
- print_debug('oam_network: after subnet',oam_network)
print_debug('oam_subnet:',oam_subnet)
+ print_debug('oam_network: after subnet',oam_network)
# router
@@ -630,131 +615,93 @@ def configure_all_ONAP():
conn.network.add_interface_to_router(onap_router, subnet_id = oam_subnet.id)
print_debug('onap_router: after adding interfaces',onap_router)
- # point to OpenStack external network (i.e. Gateway for router); network_id is passed in a body dictionary
- # (external network such as floating_net)
- # syntax: add_gateway_to_router(router, **body)
- #print('Adding external network (gateway) to ONAP router...')
-
- # nope
- # network_dict_body = {'network_id': public_network.id}
- # nope
- # network_dict_body = {
- # 'external_fixed_ips': [{'subnet_id' : public_subnet.id}],
- # 'network_id': public_network.id
- # }
-
- # external_network = conn.network.find_network(EXTERNAL_NETWORK_NAME)
- # print_debug('external_network:',external_network)
- # external_subnet_ID_list = external_network.subnet_ids
- # print_debug('external_subnet_ID_list:',external_subnet_ID_list)
-
- # # build external_fixed_ips: list of dictionaries, each with 'subnet_id' key (and may have 'ip_address' key as well)
- # onap_gateway_external_subnets = []
- # for ext_subn_id in external_subnet_ID_list: # there should be only one subnet ID in the list, but go through each item, just in case
- # onap_gateway_external_subnets.append({'subnet_id':ext_subn_id})
-
- # #network_dict_body = {'gateway' : {'network_id' : external_network.id}}
- # #network_dict_body = {'network_id' : external_network.id}
- # #conn.network.add_gateway_to_router(onap_router, body=network_dict_body)
- # #conn.network.add_gateway_to_router(onap_router, network_id=external_network.id)
- # #conn.network.add_gateway_to_router(onap_router, **network_dict_body)
-
- # network_dict_body = {
- # 'network_id': external_network.id,
- # 'enable_snat': True, # True should be the default, so there should be no need to set it
- # 'external_fixed_ips': onap_gateway_external_subnets
- # }
- # #conn.network.add_gateway_to_router(onap_router, **network_dict_body)
- # print_debug('onap_router: after add_gateway_to_router',onap_router)
-
-
-
-
- # # also create 5 flavors, from tiny to xlarge (hard-coded, no need for parameters)
- # # (Flavor is a Resource)
- # print('Creating flavors...')
- # print('Creating m1.tiny Flavor...')
- # tiny_flavor = conn.compute.find_flavor("m1.tiny")
- # if tiny_flavor != None:
- # print('m1.tiny Flavor already exists')
- # else:
- # tiny_flavor = conn.compute.create_flavor(
- # name = 'm1.tiny',
- # vcpus = 1,
- # disk = 1,
- # ram = 512,
- # ephemeral = 0,
- # #swap = 0,
- # #rxtx_factor = 1.0,
- # is_public = True)
- # print_debug('tiny_flavor: ',tiny_flavor)
-
- # print('Creating m1.small Flavor...')
- # small_flavor = conn.compute.find_flavor("m1.small")
- # if small_flavor != None:
- # print('m1.small Flavor already exists')
- # else:
- # small_flavor = conn.compute.create_flavor(
- # name = 'm1.small',
- # vcpus = 1,
- # disk = 20,
- # ram = 2048,
- # ephemeral = 0,
- # #swap = 0,
- # #rxtx_factor = 1.0,
- # is_public = True)
- # print_debug('small_flavor: ',small_flavor)
-
- # print('Creating m1.medium Flavor...')
- # medium_flavor = conn.compute.find_flavor("m1.medium")
- # if medium_flavor != None:
- # print('m1.medium Flavor already exists')
- # else:
- # medium_flavor = conn.compute.create_flavor(
- # name = 'm1.medium',
- # vcpus = 2,
- # disk = 40,
- # ram = 4096,
- # ephemeral = 0,
- # #swap = 0,
- # #rxtx_factor = 1.0,
- # is_public = True)
- # print_debug('medium_flavor: ',medium_flavor)
-
- # print('Creating m1.large Flavor...')
- # large_flavor = conn.compute.find_flavor("m1.large")
- # if large_flavor != None:
- # print('m1.large Flavor already exists')
- # else:
- # large_flavor = conn.compute.create_flavor(
- # name = 'm1.large',
- # vcpus = 4,
- # disk = 80,
- # ram = 8192,
- # ephemeral = 0,
- # #swap = 0,
- # #rxtx_factor = 1.0,
- # is_public = True)
- # print_debug('large_flavor: ',large_flavor)
-
- # print('Creating m1.xlarge Flavor...')
- # xlarge_flavor = conn.compute.find_flavor("m1.xlarge")
- # if xlarge_flavor != None:
- # print('m1.xlarge Flavor already exists')
- # else:
- # xlarge_flavor = conn.compute.create_flavor(
- # name = 'm1.xlarge',
- # vcpus = 8,
- # disk = 160,
- # ram = 16384,
- # ephemeral = 0,
- # #swap = 0,
- # #rxtx_factor = 1.0,
- # is_public = True)
- # print_debug('xlarge_flavor: ',xlarge_flavor)
+
+ # also create 5 flavors, from tiny to xlarge (hard-coded, no need for parameters)
+ # (Flavor is a Resource)
+ print('Creating flavors...')
+ print('Creating m1.tiny Flavor...')
+ tiny_flavor = conn.compute.find_flavor("m1.tiny")
+ if tiny_flavor != None:
+ print('m1.tiny Flavor already exists')
+ else:
+ tiny_flavor = conn.compute.create_flavor(
+ name = 'm1.tiny',
+ vcpus = 1,
+ disk = 1,
+ ram = 512,
+ ephemeral = 0,
+ #swap = 0,
+ #rxtx_factor = 1.0,
+ is_public = True)
+ print_debug('tiny_flavor: ',tiny_flavor)
+
+ print('Creating m1.small Flavor...')
+ small_flavor = conn.compute.find_flavor("m1.small")
+ if small_flavor != None:
+ print('m1.small Flavor already exists')
+ else:
+ small_flavor = conn.compute.create_flavor(
+ name = 'm1.small',
+ vcpus = 1,
+ disk = 20,
+ ram = 2048,
+ ephemeral = 0,
+ #swap = 0,
+ #rxtx_factor = 1.0,
+ is_public = True)
+ print_debug('small_flavor: ',small_flavor)
+
+ print('Creating m1.medium Flavor...')
+ medium_flavor = conn.compute.find_flavor("m1.medium")
+ if medium_flavor != None:
+ print('m1.medium Flavor already exists')
+ else:
+ medium_flavor = conn.compute.create_flavor(
+ name = 'm1.medium',
+ vcpus = 2,
+ disk = 40,
+ ram = 4096,
+ ephemeral = 0,
+ #swap = 0,
+ #rxtx_factor = 1.0,
+ is_public = True)
+ print_debug('medium_flavor: ',medium_flavor)
+
+ print('Creating m1.large Flavor...')
+ large_flavor = conn.compute.find_flavor("m1.large")
+ if large_flavor != None:
+ print('m1.large Flavor already exists')
+ else:
+ large_flavor = conn.compute.create_flavor(
+ name = 'm1.large',
+ vcpus = 4,
+ disk = 80,
+ ram = 8192,
+ ephemeral = 0,
+ #swap = 0,
+ #rxtx_factor = 1.0,
+ is_public = True)
+ print_debug('large_flavor: ',large_flavor)
+
+ print('Creating m1.xlarge Flavor...')
+ xlarge_flavor = conn.compute.find_flavor("m1.xlarge")
+ if xlarge_flavor != None:
+ print('m1.xlarge Flavor already exists')
+ else:
+ xlarge_flavor = conn.compute.create_flavor(
+ name = 'm1.xlarge',
+ vcpus = 8,
+ disk = 160,
+ ram = 16384,
+ ephemeral = 0,
+ #swap = 0,
+ #rxtx_factor = 1.0,
+ is_public = True)
+ print_debug('xlarge_flavor: ',xlarge_flavor)
# create images: Ubuntu 16.04, 14.04, CirrOS, ...
+ # store them in images/ directory
# 64-bit QCOW2 image for cirros-0.4.0-x86_64-disk.img
# description: CirrOS minimal Linux distribution
# http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img
@@ -770,31 +717,109 @@ def configure_all_ONAP():
# http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img
# user: ubuntu
- # do not use compute proxy for images; there is an image proxy (v1, and v2); use shade layer, directly with Connection object;
+ # do not use compute proxy for images; there is an image proxy (v1, and v2);
+ # use shade layer, directly with Connection object: Connection.create_image()
# conn.get_image() returns a Python Munch object (subclass of Dictionary)
- # URL download not supported yet; download image separately, place it in the directory
+ # However, URL download not supported yet; download image separately, place it in the directory
# https://docs.openstack.org/openstacksdk/latest/user/connection.html#openstack.connection.Connection.create_image
+ # image proxy: conn.image.upload_image()
+ # Image class:
+ # https://docs.openstack.org/openstacksdk/latest/user/resources/image/v2/image.html#openstack.image.v2.image.Image
+ # URL should be supported by image proxy
+
+ # TODO@@@ try image v2 proxy, if it supports URLs;
+ # maybe load only images for current CPU (i.e. only x86 images for x86, only Arm images for Arm)
+ # TODO@@@ list image names/URLs in dictionary, and load then in a loop
+
+ # Pattern: prepare an attribute dictionary, then call conn.image.upload_image()
+ # image_attributes_dict = {}
+ # image_attributes_dict['name']='cirros-0.4.0-aarch64-disk.img'
+ # image_attributes_dict['url']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img'
+ # conn.image.upload_image(disk_format='qcow2',**image_attributes_dict)
+
+ # With a dictionary of names/URLs :
+ # image_ref_dict = {}
+ # image_ref_dict['cirros-0.4.0-x86_64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img'
+ # image_ref_dict['cirros-0.4.0-arm-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-arm-disk.img'
+ # image_ref_dict['cirros-0.4.0-aarch64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img'
+ # etc.
+ # for image_name in image_ref_dict:
+ # image_attributes_dict['name'] = image_name
+ # image_attributes_dict['url'] = image_ref_dict[image_name]
+ # conn.image.upload_image(disk_format='qcow2',**image_attributes_dict)
+
+
+ # Create and populate image dictionary
+ image_ref_dict = {}
+
+ # Ubuntu 16.04 LTS (Xenial Xerus) images
+ image_ref_dict['xenial-server-cloudimg-amd64-disk1.img']='https://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-amd64-disk1.img'
+ image_ref_dict['xenial-server-cloudimg-arm64-disk1.img']='https://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-arm64-disk1.img'
+
+ # Ubuntu 14.04.5 LTS (Trusty Tahr) images
+ image_ref_dict['trusty-server-cloudimg-amd64-disk1.img']='http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img'
+ image_ref_dict['trusty-server-cloudimg-arm64-disk1.img']='http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-arm64-disk1.img'
+
+ # CirrOS images
+ image_ref_dict['cirros-0.4.0-x86_64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img'
+ image_ref_dict['cirros-0.4.0-arm-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-arm-disk.img'
+ image_ref_dict['cirros-0.4.0-aarch64-disk.img']='http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-aarch64-disk.img'
+
+
+ # if URL-based upload using image proxy works, it will replace the section below which assumes image files
+ # are in a subdirectory, and uses Connection.create_image() instead of Connection.image.upload_image()
+ IMAGES_DIR = 'images/'
+
+ IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution x86'
+ print('Creating image:',IMAGE_NAME,'...')
+ if conn.get_image(IMAGE_NAME) != None:
+ print(IMAGE_NAME,'image already exists')
+ else:
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-x86_64-disk.img')
+
+ IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution ARM'
+ print('Creating image:',IMAGE_NAME,'...')
+ if conn.get_image(IMAGE_NAME) != None:
+ print(IMAGE_NAME,'image already exists')
+ else:
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-arm-disk.img')
+
+ IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution AARCH64'
+ print('Creating image:',IMAGE_NAME,'...')
+ if conn.get_image(IMAGE_NAME) != None:
+ print(IMAGE_NAME,'image already exists')
+ else:
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'cirros-0.4.0-aarch64-disk.img')
+
+ IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus x86'
+ print('Creating image:',IMAGE_NAME,'...')
+ if conn.get_image(IMAGE_NAME) != None:
+ print(IMAGE_NAME,'image already exists')
+ else:
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'xenial-server-cloudimg-amd64-disk1.img')
- IMAGE_NAME = 'CirrOS_0.4.0_minimal_Linux_distribution'
+ IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus ARM64'
print('Creating image:',IMAGE_NAME,'...')
if conn.get_image(IMAGE_NAME) != None:
print(IMAGE_NAME,'image already exists')
else:
- conn.create_image(IMAGE_NAME, filename='cirros-0.4.0-x86_64-disk.img')
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'xenial-server-cloudimg-arm64-disk1.img')
- IMAGE_NAME = 'Ubuntu_Server_16.04_LTS_Xenial_Xerus'
+ IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr x86'
print('Creating image:',IMAGE_NAME,'...')
if conn.get_image(IMAGE_NAME) != None:
print(IMAGE_NAME,'image already exists')
else:
- conn.create_image(IMAGE_NAME, filename='xenial-server-cloudimg-amd64-disk1.img')
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'trusty-server-cloudimg-amd64-disk1.img')
+ # End section with Connection.create_image()
- IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr'
+ IMAGE_NAME = 'Ubuntu_Server_14.04_LTS_Trusty_Tahr ARM64'
print('Creating image:',IMAGE_NAME,'...')
if conn.get_image(IMAGE_NAME) != None:
print(IMAGE_NAME,'image already exists')
else:
- conn.create_image(IMAGE_NAME, filename='trusty-server-cloudimg-amd64-disk1.img')
+ conn.create_image(IMAGE_NAME, filename=IMAGES_DIR+'trusty-server-cloudimg-arm64-disk1.img')
+ # End section with Connection.create_image()
# create a keypair, if needed e.g. for VNF VMs; maybe to SSH for testing
@@ -813,12 +838,46 @@ def configure_all_ONAP():
print(onap_keypair.private_key)
print_debug('onap_keypair:',onap_keypair)
+
print('\nSUMMARY:')
- print('ONAP public network ID:',public_network.id)
- print('ONAP private network ID:',private_network.id)
- print('ONAP private network subnet ID:',private_subnet.id)
- print('ONAP private network subnet CIDR:',private_subnet.cidr)
- print('\n')
+ # Grab live objects (don't reuse earlier references), in case the script is used on an already configured instance
+ # This way, the summary is still displayed even if the script execution did not create anything
+ # Also, this double-checks that displayed information is accurate, freshly retrieved from the OpenStack instance
+
+ public_network = conn.network.find_network(ONAP_PUBLIC_NET_NAME)
+ if public_network != None:
+ print('ONAP public network ID:',public_network.id)
+ for fetched_subnet_ID in public_network.subnet_ids:
+ fetched_subnet = conn.network.get_subnet(fetched_subnet_ID)
+ if fetched_subnet != None:
+ print(' ONAP public network subnet ID:',fetched_subnet.id)
+ print(' ONAP public network subnet CIDR:',fetched_subnet.cidr)
+ else:
+ print('no ONAP public network')
+
+ private_network = conn.network.find_network(ONAP_PRIVATE_NET_NAME)
+ if private_network != None:
+ print('ONAP private network ID:',private_network.id)
+ for fetched_subnet_ID in private_network.subnet_ids:
+ fetched_subnet = conn.network.get_subnet(fetched_subnet_ID)
+ if fetched_subnet != None:
+ print(' ONAP private network subnet ID:',fetched_subnet.id)
+ print(' ONAP private network subnet CIDR:',fetched_subnet.cidr)
+ else:
+ print('no ONAP private network')
+
+ oam_network = conn.network.find_network(ONAP_OAM_NET_NAME)
+ if oam_network != None:
+ print('ONAP OAM network ID:',oam_network.id)
+ for fetched_subnet_ID in oam_network.subnet_ids:
+ fetched_subnet = conn.network.get_subnet(fetched_subnet_ID)
+ if fetched_subnet != None:
+ print(' ONAP OAM network subnet ID:',fetched_subnet.id)
+ print(' ONAP OAM network subnet CIDR:',fetched_subnet.cidr)
+ else:
+ print('no ONAP OAM network')
+ print('END SUMMARY\n')
+
except Exception as e:
print('*** Exception:',type(e), e)
@@ -849,7 +908,7 @@ def main():
help = 'delete ONAP configuration',
action = 'store_true')
- # parse arguments, modify global variable if need be, and use corresponding script
+ # parse arguments, modify global variable if need be, and use corresponding script (create objects, or delete objects)
args = parser.parse_args()
if args.debug:
global DEBUG_VAR
@@ -862,4 +921,3 @@ def main():
if __name__ == "__main__":
main()
-
diff --git a/setup/VIMs/OpenStack/clouds.yaml b/setup/VIMs/OpenStack/clouds.yaml
new file mode 100644
index 0000000..7bfd717
--- /dev/null
+++ b/setup/VIMs/OpenStack/clouds.yaml
@@ -0,0 +1,99 @@
+clouds:
+
+ # Openstack instance on Arm pod, controller IP@ 172.16.10.10
+ # Horizon: https://10.10.50.103/project/
+ # Identity API according to Horizon dashboard: https://10.10.50.103:5000/v2.0
+ # other potential auth_url: http://172.16.10.10:35357/v3
+ # (OS_AUTH_URL=http://controller:35357/v3)
+ # 2 project names: admin, service (project = tenant)
+ # project ID: 122caf64b3df4818bf2ce5ba793226b2
+ # EC2 URL: https://10.10.50.103:8773/services/Cloud
+ # EC2 access key: bcf3c69a7d1c405e9757f87f26faf19f
+ # 10.10.50.0/8: floating IP@
+ # 10.10.10.0/8: fixed IP@
+ armopenstack:
+ auth:
+ auth_url: https://10.10.50.103:5000/v2.0
+ project_name: admin
+ username: admin
+ password: opnfv_secret
+ region_name: RegionOne
+
+ # Openstack instance on LaaS hpe16, from OPNFV Euphrates, controller IP@ (mgt: 172.16.10.101; public: 10.16.0.101)
+ # keystone endpoints (openstack endpoint list --service keystone)
+ # admin: http://172.16.10.101:35357/v2.0
+ # internal: http://172.16.10.101:5000/v2.0
+ # public: http://10.16.0.101:5000/v2.0 : works on LaaS hpe16, from hpe16
+ hpe16openstackEuphrates:
+ auth:
+ auth_url: http://10.16.0.101:5000/v2.0
+ project_name: admin
+ username: admin
+ password: opnfv_secret
+ region_name: RegionOne
+
+ # Openstack instance on generic LaaS hpe, from OPNFV Fraser, controller IP@ (mgt: 172.16.10.36; public: 10.16.0.107)
+ # keystone endpoints (openstack endpoint list --service keystone)
+ # admin: http://172.16.10.36:35357/v3
+ # internal: http://172.16.10.36:5000/v3
+ # public: http://10.16.0.107:5000/v3
+ # Horizon: https://10.16.0.107:8078, but need SSH port forwarding through 10.10.100.26 to be reached from outside
+ # "If you are using Identity v3 you need to specify the user and the project domain name"
+
+ # generic cloud name, for a UNH IOL hpe server, for OPNFV Fraser, OpenStack installed by Fuel/MCP
+ unh-hpe-openstack-fraser:
+ auth:
+ auth_url: http://10.16.0.107:5000/v3
+ project_name: admin
+ username: admin
+ password: opnfv_secret
+ user_domain_name: Default
+ project_domain_name: Default
+ region_name: RegionOne
+ identity_api_version: 3
+
+# ubuntu@ctl01:~$ openstack project show admin
+# +-------------+----------------------------------+
+# | Field | Value |
+# +-------------+----------------------------------+
+# | description | OpenStack Admin tenant |
+# | domain_id | default |
+# | enabled | True |
+# | id | 04fcfe7aa83f4df79ae39ca748aa8637 |
+# | is_domain | False |
+# | name | admin |
+# | parent_id | default |
+# +-------------+----------------------------------+
+
+# (openstack) domain show default
+# +-------------+----------------------------------------------------------+
+# | Field | Value |
+# +-------------+----------------------------------------------------------+
+# | description | Domain created automatically to support V2.0 operations. |
+# | enabled | True |
+# | id | default |
+# | name | Default |
+# +-------------+----------------------------------------------------------+
+
+# (openstack) domain show heat_user_domain
+# +-------------+---------------------------------------------+
+# | Field | Value |
+# +-------------+---------------------------------------------+
+# | description | Contains users and projects created by heat |
+# | enabled | True |
+# | id | d9c29adac0fe4816922d783b257879d6 |
+# | name | heat_user_domain |
+# +-------------+---------------------------------------------+
+
+
+# export OS_AUTH_URL=http://10.16.0.107:5000/v3
+# export OS_PROJECT_ID=04fcfe7aa83f4df79ae39ca748aa8637
+# export OS_PROJECT_NAME="admin"
+# export OS_USER_DOMAIN_NAME="Default"
+# export OS_USERNAME="admin"
+# export OS_PASSWORD="opnfv_secret"
+# export OS_REGION_NAME="RegionOne"
+# export OS_INTERFACE=public
+# export OS_IDENTITY_API_VERSION=3
+
+
diff --git a/yamllintrc b/yamllintrc
new file mode 100644
index 0000000..a4f3d02
--- /dev/null
+++ b/yamllintrc
@@ -0,0 +1,25 @@
+# Copyright 2018 Tieto
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+extends: relaxed
+
+rules:
+ empty-lines:
+ max-start: 1
+ max-end: 1
+ colons:
+ max-spaces-after: 1
+ max-spaces-before: 1
+ line-length:
+ max: 160