diff --git a/Jenkinsfile b/Jenkinsfile index cc5d660..1ba0f65 100644 --- a/Jenkinsfile +++ b/Jenkinsfile @@ -185,7 +185,11 @@ pipeline { } finally { library("tailor-meta@${params.tailor_meta}") cleanDocker() - deleteDir() + try { + deleteDir() + } catch (e) { + println e + } } }}] } @@ -234,7 +238,11 @@ pipeline { artifacts: "$debian_dir/rules*, $debian_dir/control*, $debian_dir/Dockerfile*", allowEmptyArchive: true) library("tailor-meta@${params.tailor_meta}") cleanDocker() - deleteDir() + try { + deleteDir() + } catch (e) { + println e + } } }}] } @@ -255,7 +263,7 @@ pipeline { docker.withRegistry(params.docker_registry, docker_credentials) { bundle_image.pull() } } bundle_image.inside("-v $HOME/tailor/ccache:/ccache -e CCACHE_DIR=/ccache") { - // Invoke the Jenkins Job Cacher Plugin via the cache method. + // Invoke the Jenkins Job Cacher Plugin via the cache method. // Set the max cache size to 4GB, as S3 only allows a 5GB max upload at once cache(maxCacheSize: 4000, caches: [ arbitraryFileCache(path: '${HOME}/tailor/ccache', cacheName: recipe_label) @@ -275,7 +283,11 @@ pipeline { // archiveArtifacts(artifacts: "*.deb", allowEmptyArchive: true) library("tailor-meta@${params.tailor_meta}") cleanDocker() - deleteDir() + try { + deleteDir() + } catch (e) { + println e + } } }}] } @@ -313,7 +325,11 @@ pipeline { } finally { library("tailor-meta@${params.tailor_meta}") cleanDocker() - deleteDir() + try { + deleteDir() + } catch (e) { + println e + } } }}] }