Age | Commit message (Collapse) | Author | Files | Lines |
|
Change-Id: I41926856b129eb7dd99c5db8765961acd241eb55
Signed-off-by: Feng Pan <fpan@redhat.com>
|
|
Change-Id: I055ec261368b341c5dd9f2ab5082c8ced15ad6e5
Signed-off-by: Feng Pan <fpan@redhat.com>
|
|
- Change scenario parsing to allow k8s scenarios in gate jobs.
- Remove scenario detection in shell scripts.
Also changes network setting file for fd.io scenarios as we
don't use separate config files any more.
Change-Id: Icbcc6cb7b6b66c7cc43f5d6c58dd97730e41ea5f
Signed-off-by: Feng Pan <fpan@redhat.com>
|
|
condition actually affects *next* phase execution and not conditional on
previous phase. This change fixes the condition so that fetch logs
always runs.
Change-Id: I50c3f9f9412682ea591c47116d406b5ed81ee2b5
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
|
|
We were using 'stream' as part of the slave label to run unit tests on.
However, we typically just use master label for everything to avoid
having to update labels on slaves. Changing the unit test to follow the
same method.
Change-Id: I6a182c3f28d541109c8a9d94ea1dae2a80f2b835
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
With this commit:
https://gerrit.opnfv.org/gerrit/#/c/47875/
We can now gather logs from after at the end of a deploy/test job and
upload them to artifacts. This patch enables that behavior on verify
and daily jobs.
Change-Id: If7af4cf1363e4bbcee78d98850af7c0e20250f3f
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
Change-Id: I7f6072121c4d88192c828e2e3cf2beb5d6c22c37
Signed-off-by: jenkins-ci <jenkins-opnfv-ci@opnfv.org>
Signed-off-by: Trevor Bramwell <tbramwell@linuxfoundation.org>
|
|
Remove passing node params and enable global queue scanning. Hopefully
this will allow 2 apex-virtual jobs to run at the same time (1 per each
slave).
Change-Id: I310dbc477e267c302d50599bab2a933ce988dba7
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
Changes Include:
- Remove defining 'node:' per job template and use slave-params
defaults
- Gate job was using daily/build slave, when it should be using virtual
slave
Change-Id: Iec2321801daef1b1fa40724a244bf2f6edf36c6e
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
apex-verify job calls apex-virtual to deploy and run functest. Apex has
2 virtual slaves attached to the apex-virtual-master label. When 2
verify jobs are triggered at the same time, apex-verify job is scheduled
on each slave correctly. However, when the mulitjob triggers
apex-virtual jobs, it only schedules both of them to a single slave.
This happens even though apex-virtual job has the same slave label and
node parameters are not passed from verify job. This patch changes the
node to be passed to apex-virtual from apex-verify. That way 2 nodes
will be scheduled on, but this is still not ideal scheduling as more
than 1 verify can run per node (but not more than one virtual job).
Change-Id: I155351c9037f70df2c5dba11bb5592423849e760
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
We were missing triggering on the ci/ path in our Apex repo. This
contains some code and a file we use to trigger verify jobs for
dependent patches.
Change-Id: I54f2826f8a16a1d0219d6ecc6ef8d257840b6399
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
Issue where apex-virtual-master label node was being passed from verify
job to deploy/functest. We allow multiple verify jobs to kick off on
virtual slaves, but then passing the host down to the deploy/functest
jobs causes the deploy/functest only run 1 job at a time on the same
node, rather than running 2 jobs at a time, one on each slave.
Change-Id: I1648eb6b84f17a2b08db4d161effe977c7952d63
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
We use apex-virtual-master label for slaves that will deploy and run
functest. Build and unit test will not use this label and will use
apex-build-master.
Change-Id: Ibf266bd37813ea8ef38fc8060f73f83462275cfd
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
JIRA: RELENG-287
Change-Id: Ie50fdddc47b47764d3e3064904f19015d5d39341
Signed-off-by: Trevor Bramwell <tbramwell@linuxfoundation.org>
|