Remove legacy Jenkins quality pipeline. (#1510)
- it has been used only at Google for a quality pipeline - the worker.py is Python 2 and closely tied to Google Cloud, so I don't believe there is much to be reused by someone else nowadays and not worth the effort to port to Python 3. Signed-off-by: Michael Grupp <grupp@magazino.eu>master
parent
f30d1f7e1e
commit
051a018c47
|
@ -1,122 +0,0 @@
|
||||||
# Copyright 2016 The Cartographer Authors
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
# you may not use this file except in compliance with the License.
|
|
||||||
# You may obtain a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
# See the License for the specific language governing permissions and
|
|
||||||
# limitations under the License.
|
|
||||||
|
|
||||||
FROM ros:kinetic
|
|
||||||
|
|
||||||
ARG CARTOGRAPHER_VERSION=master
|
|
||||||
|
|
||||||
# Xenial's base image doesn't ship with sudo.
|
|
||||||
RUN apt-get update && apt-get install -y sudo time && rm -rf /var/lib/apt/lists/*
|
|
||||||
# First, we invalidate the entire cache if cartographer-project/cartographer has
|
|
||||||
# changed. This file's content changes whenever master changes. See:
|
|
||||||
# http://stackoverflow.com/questions/36996046/how-to-prevent-dockerfile-caching-git-clone
|
|
||||||
ADD https://api.github.com/repos/cartographer-project/cartographer/git/refs/heads/master \
|
|
||||||
cartographer_ros/cartographer_version.json
|
|
||||||
|
|
||||||
# wstool needs the updated rosinstall file to clone the correct repos.
|
|
||||||
COPY cartographer_ros.rosinstall cartographer_ros/
|
|
||||||
COPY scripts/prepare_jenkins_catkin_workspace.sh cartographer_ros/scripts/
|
|
||||||
|
|
||||||
# Invalidates the Docker cache to ensure this command is always executed.
|
|
||||||
ARG CACHEBUST=1
|
|
||||||
RUN CARTOGRAPHER_VERSION=$CARTOGRAPHER_VERSION \
|
|
||||||
cartographer_ros/scripts/prepare_jenkins_catkin_workspace.sh
|
|
||||||
|
|
||||||
# rosdep needs the updated package.xml files to install the correct debs.
|
|
||||||
COPY cartographer_ros/package.xml catkin_ws/src/cartographer_ros/cartographer_ros/
|
|
||||||
COPY cartographer_ros_msgs/package.xml catkin_ws/src/cartographer_ros/cartographer_ros_msgs/
|
|
||||||
COPY cartographer_rviz/package.xml catkin_ws/src/cartographer_ros/cartographer_rviz/
|
|
||||||
COPY scripts/install_debs.sh cartographer_ros/scripts/
|
|
||||||
RUN cartographer_ros/scripts/install_debs.sh && rm -rf /var/lib/apt/lists/*
|
|
||||||
|
|
||||||
# Install proto3.
|
|
||||||
RUN /catkin_ws/src/cartographer/scripts/install_proto3.sh
|
|
||||||
|
|
||||||
# Build, install, and test all packages individually to allow caching. The
|
|
||||||
# ordering of these steps must match the topological package ordering as
|
|
||||||
# determined by Catkin.
|
|
||||||
COPY scripts/install.sh cartographer_ros/scripts/
|
|
||||||
COPY scripts/catkin_test_results.sh cartographer_ros/scripts/
|
|
||||||
|
|
||||||
COPY cartographer_ros_msgs catkin_ws/src/cartographer_ros/cartographer_ros_msgs/
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer_ros_msgs && \
|
|
||||||
cartographer_ros/scripts/install.sh --pkg cartographer_ros_msgs \
|
|
||||||
--catkin-make-args run_tests && \
|
|
||||||
cartographer_ros/scripts/catkin_test_results.sh build_isolated/cartographer_ros_msgs
|
|
||||||
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg ceres-solver
|
|
||||||
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer && \
|
|
||||||
cartographer_ros/scripts/install.sh --pkg cartographer --make-args test
|
|
||||||
|
|
||||||
COPY cartographer_ros catkin_ws/src/cartographer_ros/cartographer_ros/
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer_ros && \
|
|
||||||
cartographer_ros/scripts/install.sh --pkg cartographer_ros \
|
|
||||||
--catkin-make-args run_tests && \
|
|
||||||
cartographer_ros/scripts/catkin_test_results.sh build_isolated/cartographer_ros
|
|
||||||
|
|
||||||
COPY cartographer_rviz catkin_ws/src/cartographer_ros/cartographer_rviz/
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer_rviz && \
|
|
||||||
cartographer_ros/scripts/install.sh --pkg cartographer_rviz \
|
|
||||||
--catkin-make-args run_tests && \
|
|
||||||
cartographer_ros/scripts/catkin_test_results.sh build_isolated/cartographer_rviz
|
|
||||||
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer_toru
|
|
||||||
RUN cartographer_ros/scripts/install.sh --pkg cartographer_fetch
|
|
||||||
|
|
||||||
COPY scripts/ros_entrypoint.sh /
|
|
||||||
# A BTRFS bug may prevent us from cleaning up these directories.
|
|
||||||
# https://btrfs.wiki.kernel.org/index.php/Problem_FAQ#I_cannot_delete_an_empty_directory
|
|
||||||
RUN rm -rf cartographer_ros catkin_ws || true
|
|
||||||
|
|
||||||
RUN sudo apt-get update
|
|
||||||
RUN sudo apt-get -y install openjdk-8-jdk python-pip
|
|
||||||
|
|
||||||
ENV HOME /home/jenkins
|
|
||||||
RUN addgroup --system --gid 10000 jenkins
|
|
||||||
RUN adduser --system --ingroup jenkins --home $HOME --uid 10000 jenkins
|
|
||||||
|
|
||||||
LABEL Description="This is a base image, which provides the Jenkins agent executable (slave.jar)" Vendor="Jenkins project" Version="3.17"
|
|
||||||
|
|
||||||
ARG VERSION=3.17
|
|
||||||
ARG AGENT_WORKDIR=/home/jenkins/agent
|
|
||||||
|
|
||||||
RUN curl --create-dirs -sSLo /usr/share/jenkins/slave.jar https://repo.jenkins-ci.org/public/org/jenkins-ci/main/remoting/${VERSION}/remoting-${VERSION}.jar \
|
|
||||||
&& chmod 755 /usr/share/jenkins \
|
|
||||||
&& chmod 644 /usr/share/jenkins/slave.jar
|
|
||||||
# USER jenkins
|
|
||||||
ENV AGENT_WORKDIR=${AGENT_WORKDIR}
|
|
||||||
RUN mkdir /home/jenkins/.jenkins && mkdir -p ${AGENT_WORKDIR}
|
|
||||||
|
|
||||||
VOLUME /home/jenkins/.jenkins
|
|
||||||
VOLUME ${AGENT_WORKDIR}
|
|
||||||
WORKDIR /home/jenkins
|
|
||||||
|
|
||||||
COPY jenkins/jenkins-slave /usr/local/bin/jenkins-slave
|
|
||||||
|
|
||||||
ENV CLOUDSDK_CORE_DISABLE_PROMPTS 1
|
|
||||||
ENV PATH /opt/google-cloud-sdk/bin:$PATH
|
|
||||||
|
|
||||||
USER root
|
|
||||||
|
|
||||||
# Install Google Cloud Components
|
|
||||||
RUN curl https://sdk.cloud.google.com | bash && mv google-cloud-sdk /opt
|
|
||||||
RUN gcloud components install kubectl
|
|
||||||
|
|
||||||
RUN pip install --upgrade google-cloud-datastore
|
|
||||||
RUN pip install --upgrade google-cloud-bigquery
|
|
||||||
COPY jenkins/worker.py /worker.py
|
|
||||||
|
|
||||||
# USER root
|
|
||||||
ENTRYPOINT ["jenkins-slave"]
|
|
|
@ -1,110 +0,0 @@
|
||||||
podTemplate(label: 'node-0', containers: [
|
|
||||||
containerTemplate(
|
|
||||||
name: 'jnlp',
|
|
||||||
image: 'eggsy84/gcp-jenkins-slave-k8s-seed:latest',
|
|
||||||
ttyEnabled: false,
|
|
||||||
command: '',
|
|
||||||
privileged: true,
|
|
||||||
alwaysPullImage: false,
|
|
||||||
workingDir: '/home/jenkins',
|
|
||||||
args: '${computer.jnlpmac} ${computer.name}'
|
|
||||||
)
|
|
||||||
],
|
|
||||||
volumes: [
|
|
||||||
secretVolume(mountPath: '/opt/config', secretName: 'gcloud-svc-account'),
|
|
||||||
hostPathVolume(mountPath: '/var/run/docker.sock', hostPath: '/var/run/docker.sock'),
|
|
||||||
persistentVolumeClaim(claimName: 'data-claim-compile', mountPath: '/data'),
|
|
||||||
]
|
|
||||||
) {
|
|
||||||
node('node-0') {
|
|
||||||
stage('Compile') {
|
|
||||||
sh 'gcloud auth activate-service-account --key-file=/opt/config/gcloud-svc-account.json'
|
|
||||||
sh 'cd /data && rm -Rf *'
|
|
||||||
sh 'cd /data && git clone https://github.com/cartographer-project/cartographer_ros'
|
|
||||||
sh 'cd /data/cartographer_ros && docker build -f jenkins/Dockerfile.kinetic -t kinetic-jenkins-slave --build-arg CACHEBUST=$(date +%s) .'
|
|
||||||
}
|
|
||||||
stage('Push') {
|
|
||||||
sh 'docker tag kinetic-jenkins-slave eu.gcr.io/cartographer-141408/kinetic-jenkins-slave'
|
|
||||||
sh 'gcloud docker -- push eu.gcr.io/cartographer-141408/kinetic-jenkins-slave'
|
|
||||||
sh 'cd /data && rm -Rf *'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
podTemplate(label: 'node-1', containers: [
|
|
||||||
containerTemplate(
|
|
||||||
name: 'jnlp',
|
|
||||||
image: 'eu.gcr.io/cartographer-141408/kinetic-jenkins-slave:latest',
|
|
||||||
ttyEnabled: false,
|
|
||||||
command: '',
|
|
||||||
privileged: true,
|
|
||||||
alwaysPullImage: true,
|
|
||||||
workingDir: '/home/jenkins',
|
|
||||||
args: '${computer.jnlpmac} ${computer.name}'
|
|
||||||
)
|
|
||||||
],
|
|
||||||
volumes: [
|
|
||||||
secretVolume(mountPath: '/opt/config', secretName: 'gcloud-svc-account'),
|
|
||||||
hostPathVolume(mountPath: '/var/run/docker.sock', hostPath: '/var/run/docker.sock'),
|
|
||||||
persistentVolumeClaim(claimName: 'data-claim-compile', mountPath: '/data'),
|
|
||||||
]
|
|
||||||
) {
|
|
||||||
node('node-1') {
|
|
||||||
stage('Run Fetch Pipeline') {
|
|
||||||
sh 'gcloud auth activate-service-account --key-file=/opt/config/gcloud-svc-account.json'
|
|
||||||
sh 'GOOGLE_APPLICATION_CREDENTIALS="/opt/config/gcloud-svc-account.json" GOOGLE_CLOUD_DISABLE_GRPC=True python /worker.py --worker_id 0 --num_workers 1 --pipeline_id fetch'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
podTemplate(label: 'node-2', containers: [
|
|
||||||
containerTemplate(
|
|
||||||
name: 'jnlp',
|
|
||||||
image: 'eu.gcr.io/cartographer-141408/kinetic-jenkins-slave:latest',
|
|
||||||
ttyEnabled: false,
|
|
||||||
command: '',
|
|
||||||
privileged: true,
|
|
||||||
alwaysPullImage: true,
|
|
||||||
workingDir: '/home/jenkins',
|
|
||||||
args: '${computer.jnlpmac} ${computer.name}'
|
|
||||||
)
|
|
||||||
],
|
|
||||||
volumes: [
|
|
||||||
secretVolume(mountPath: '/opt/config', secretName: 'gcloud-svc-account'),
|
|
||||||
hostPathVolume(mountPath: '/var/run/docker.sock', hostPath: '/var/run/docker.sock'),
|
|
||||||
persistentVolumeClaim(claimName: 'data-claim-compile', mountPath: '/data'),
|
|
||||||
]
|
|
||||||
) {
|
|
||||||
node('node-2') {
|
|
||||||
stage('Run Backpack Pipeline') {
|
|
||||||
sh 'gcloud auth activate-service-account --key-file=/opt/config/gcloud-svc-account.json'
|
|
||||||
sh 'GOOGLE_APPLICATION_CREDENTIALS="/opt/config/gcloud-svc-account.json" GOOGLE_CLOUD_DISABLE_GRPC=True python /worker.py --worker_id 0 --num_workers 1 --pipeline_id backpack'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
podTemplate(label: 'node-3', containers: [
|
|
||||||
containerTemplate(
|
|
||||||
name: 'jnlp',
|
|
||||||
image: 'eu.gcr.io/cartographer-141408/kinetic-jenkins-slave:latest',
|
|
||||||
ttyEnabled: false,
|
|
||||||
command: '',
|
|
||||||
privileged: true,
|
|
||||||
alwaysPullImage: true,
|
|
||||||
workingDir: '/home/jenkins',
|
|
||||||
args: '${computer.jnlpmac} ${computer.name}'
|
|
||||||
)
|
|
||||||
],
|
|
||||||
volumes: [
|
|
||||||
secretVolume(mountPath: '/opt/config', secretName: 'gcloud-svc-account'),
|
|
||||||
hostPathVolume(mountPath: '/var/run/docker.sock', hostPath: '/var/run/docker.sock'),
|
|
||||||
persistentVolumeClaim(claimName: 'data-claim-compile', mountPath: '/data'),
|
|
||||||
]
|
|
||||||
) {
|
|
||||||
node('node-3') {
|
|
||||||
stage('Run Toru Pipeline') {
|
|
||||||
sh 'gcloud auth activate-service-account --key-file=/opt/config/gcloud-svc-account.json'
|
|
||||||
sh 'GOOGLE_APPLICATION_CREDENTIALS="/opt/config/gcloud-svc-account.json" GOOGLE_CLOUD_DISABLE_GRPC=True python /worker.py --worker_id 0 --num_workers 1 --pipeline_id toru'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,84 +0,0 @@
|
||||||
#!/usr/bin/env sh
|
|
||||||
|
|
||||||
# The MIT License
|
|
||||||
#
|
|
||||||
# Copyright (c) 2015, CloudBees, Inc.
|
|
||||||
#
|
|
||||||
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
||||||
# of this software and associated documentation files (the "Software"), to deal
|
|
||||||
# in the Software without restriction, including without limitation the rights
|
|
||||||
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
||||||
# copies of the Software, and to permit persons to whom the Software is
|
|
||||||
# furnished to do so, subject to the following conditions:
|
|
||||||
#
|
|
||||||
# The above copyright notice and this permission notice shall be included in
|
|
||||||
# all copies or substantial portions of the Software.
|
|
||||||
#
|
|
||||||
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
||||||
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
||||||
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
||||||
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
||||||
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
||||||
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
||||||
# THE SOFTWARE.
|
|
||||||
|
|
||||||
# Usage jenkins-slave.sh [options] -url http://jenkins [SECRET] [AGENT_NAME]
|
|
||||||
# Optional environment variables :
|
|
||||||
# * JENKINS_TUNNEL : HOST:PORT for a tunnel to route TCP traffic to jenkins host, when jenkins can't be directly accessed over network
|
|
||||||
# * JENKINS_URL : alternate jenkins URL
|
|
||||||
# * JENKINS_SECRET : agent secret, if not set as an argument
|
|
||||||
# * JENKINS_AGENT_NAME : agent name, if not set as an argument
|
|
||||||
|
|
||||||
if [ $# -eq 1 ]; then
|
|
||||||
|
|
||||||
# if `docker run` only has one arguments, we assume user is running alternate command like `bash` to inspect the image
|
|
||||||
exec "$@"
|
|
||||||
|
|
||||||
else
|
|
||||||
|
|
||||||
# if -tunnel is not provided try env vars
|
|
||||||
case "$@" in
|
|
||||||
*"-tunnel "*) ;;
|
|
||||||
*)
|
|
||||||
if [ ! -z "$JENKINS_TUNNEL" ]; then
|
|
||||||
TUNNEL="-tunnel $JENKINS_TUNNEL"
|
|
||||||
fi ;;
|
|
||||||
esac
|
|
||||||
|
|
||||||
if [ -n "$JENKINS_URL" ]; then
|
|
||||||
URL="-url $JENKINS_URL"
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ -n "$JENKINS_NAME" ]; then
|
|
||||||
JENKINS_AGENT_NAME="$JENKINS_NAME"
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ -z "$JNLP_PROTOCOL_OPTS" ]; then
|
|
||||||
echo "Warning: JnlpProtocol3 is disabled by default, use JNLP_PROTOCOL_OPTS to alter the behavior"
|
|
||||||
JNLP_PROTOCOL_OPTS="-Dorg.jenkinsci.remoting.engine.JnlpProtocol3.disabled=true"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# If both required options are defined, do not pass the parameters
|
|
||||||
OPT_JENKINS_SECRET=""
|
|
||||||
if [ -n "$JENKINS_SECRET" ]; then
|
|
||||||
case "$@" in
|
|
||||||
*"${JENKINS_SECRET}"*) echo "Warning: SECRET is defined twice in command-line arguments and the environment variable" ;;
|
|
||||||
*)
|
|
||||||
OPT_JENKINS_SECRET="${JENKINS_SECRET}" ;;
|
|
||||||
esac
|
|
||||||
fi
|
|
||||||
|
|
||||||
OPT_JENKINS_AGENT_NAME=""
|
|
||||||
if [ -n "$JENKINS_AGENT_NAME" ]; then
|
|
||||||
case "$@" in
|
|
||||||
*"${JENKINS_AGENT_NAME}"*) echo "Warning: AGENT_NAME is defined twice in command-line arguments and the environment variable" ;;
|
|
||||||
*)
|
|
||||||
OPT_JENKINS_AGENT_NAME="${JENKINS_AGENT_NAME}" ;;
|
|
||||||
esac
|
|
||||||
fi
|
|
||||||
|
|
||||||
#TODO: Handle the case when the command-line and Environment variable contain different values.
|
|
||||||
#It is fine it blows up for now since it should lead to an error anyway.
|
|
||||||
|
|
||||||
exec java $JAVA_OPTS $JNLP_PROTOCOL_OPTS -cp /usr/share/jenkins/slave.jar hudson.remoting.jnlp.Main -headless $TUNNEL $URL $OPT_JENKINS_SECRET $OPT_JENKINS_AGENT_NAME "$@"
|
|
||||||
fi
|
|
|
@ -1,350 +0,0 @@
|
||||||
"""This is the script executed by workers of the quality control pipline."""
|
|
||||||
|
|
||||||
import argparse
|
|
||||||
import datetime
|
|
||||||
from os.path import basename
|
|
||||||
from pprint import pprint
|
|
||||||
import re
|
|
||||||
import subprocess
|
|
||||||
|
|
||||||
from google.cloud import bigquery
|
|
||||||
from google.cloud import datastore
|
|
||||||
|
|
||||||
|
|
||||||
class Pattern(object):
|
|
||||||
"""Defines a pattern for regular expression matching."""
|
|
||||||
|
|
||||||
def __init__(self, pattern):
|
|
||||||
self.regex = re.compile(pattern, re.MULTILINE)
|
|
||||||
|
|
||||||
def extract(self, text):
|
|
||||||
"""Returns a dictionary of named capture groups to extracted output.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
text: input to parse
|
|
||||||
|
|
||||||
Returns an empty dict if no match was found.
|
|
||||||
"""
|
|
||||||
match = self.regex.search(text)
|
|
||||||
if match is None:
|
|
||||||
return {}
|
|
||||||
return match.groupdict()
|
|
||||||
|
|
||||||
def extract_last_occurence(self, text):
|
|
||||||
"""Returns tuple of extracted outputs.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
text: input to parse
|
|
||||||
|
|
||||||
Returns the information extracted from the last match. Returns
|
|
||||||
None if no match was found.
|
|
||||||
"""
|
|
||||||
matches = self.regex.findall(text)
|
|
||||||
if not matches:
|
|
||||||
return None
|
|
||||||
return matches[-1]
|
|
||||||
|
|
||||||
|
|
||||||
# BigQuery table schema
|
|
||||||
SCHEMA = [
|
|
||||||
bigquery.SchemaField('date', 'DATE'),
|
|
||||||
bigquery.SchemaField('commit_sha1', 'STRING'),
|
|
||||||
bigquery.SchemaField('job_id', 'INTEGER'),
|
|
||||||
bigquery.SchemaField('rosbag', 'STRING'),
|
|
||||||
bigquery.SchemaField('user_time_secs', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('system_time_secs', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('wall_time_secs', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('max_set_size_kbytes', 'INTEGER'),
|
|
||||||
bigquery.SchemaField('constraints_count', 'INTEGER'),
|
|
||||||
bigquery.SchemaField('constraints_score_minimum', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('constraints_score_maximum', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('constraints_score_mean', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_abs_trans_err', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_abs_trans_err_dev', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_sqr_trans_err', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_sqr_trans_err_dev', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_abs_rot_err', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_abs_rot_err_dev', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_sqr_rot_err', 'FLOAT'),
|
|
||||||
bigquery.SchemaField('ground_truth_sqr_rot_err_dev', 'FLOAT')
|
|
||||||
]
|
|
||||||
|
|
||||||
# Pattern matchers for the various fields of the '/usr/bin/time -v' output
|
|
||||||
USER_TIME_PATTERN = Pattern(
|
|
||||||
r'^\s*User time \(seconds\): (?P<user_time>\d+.\d+|\d+)')
|
|
||||||
SYSTEM_TIME_PATTERN = Pattern(
|
|
||||||
r'^\s*System time \(seconds\): (?P<system_time>\d+.\d+|\d+)')
|
|
||||||
WALL_TIME_PATTERN = Pattern(
|
|
||||||
r'^\s*Elapsed \(wall clock\) time \(h:mm:ss or m:ss\): '
|
|
||||||
r'((?P<hours>\d{1,2}):|)(?P<minutes>\d{1,2}):(?P<seconds>\d{2}\.\d{2})')
|
|
||||||
MAX_RES_SET_SIZE_PATTERN = Pattern(
|
|
||||||
r'^\s*Maximum resident set size \(kbytes\): (?P<max_set_size>\d+)')
|
|
||||||
CONSTRAINT_STATS_PATTERN = Pattern(
|
|
||||||
r'Score histogram:[\n\r]+'
|
|
||||||
r'Count:\s+(?P<constraints_count>\d+)\s+'
|
|
||||||
r'Min:\s+(?P<constraints_score_min>\d+\.\d+)\s+'
|
|
||||||
r'Max:\s+(?P<constraints_score_max>\d+\.\d+)\s+'
|
|
||||||
r'Mean:\s+(?P<constraints_score_mean>\d+\.\d+)')
|
|
||||||
GROUND_TRUTH_STATS_PATTERN = Pattern(
|
|
||||||
r'Result:[\n\r]+'
|
|
||||||
r'Abs translational error (?P<abs_trans_err>\d+\.\d+) '
|
|
||||||
r'\+/- (?P<abs_trans_err_dev>\d+\.\d+) m[\n\r]+'
|
|
||||||
r'Sqr translational error (?P<sqr_trans_err>\d+\.\d+) '
|
|
||||||
r'\+/- (?P<sqr_trans_err_dev>\d+\.\d+) m\^2[\n\r]+'
|
|
||||||
r'Abs rotational error (?P<abs_rot_err>\d+\.\d+) '
|
|
||||||
r'\+/- (?P<abs_rot_err_dev>\d+\.\d+) deg[\n\r]+'
|
|
||||||
r'Sqr rotational error (?P<sqr_rot_err>\d+\.\d+) '
|
|
||||||
r'\+/- (?P<sqr_rot_err_dev>\d+\.\d+) deg\^2')
|
|
||||||
|
|
||||||
# Pattern matcher for extracting the HEAD commit SHA-1 hash.
|
|
||||||
GIT_SHA1_PATTERN = Pattern(r'^(?P<sha1>[0-9a-f]{40})\s+HEAD')
|
|
||||||
|
|
||||||
|
|
||||||
def get_head_git_sha1():
|
|
||||||
"""Returns the SHA-1 hash of the commit tagged HEAD."""
|
|
||||||
output = subprocess.check_output([
|
|
||||||
'git', 'ls-remote',
|
|
||||||
'https://github.com/cartographer-project/cartographer.git'
|
|
||||||
])
|
|
||||||
parsed = GIT_SHA1_PATTERN.extract(output)
|
|
||||||
return parsed['sha1']
|
|
||||||
|
|
||||||
|
|
||||||
def extract_stats(inp):
|
|
||||||
"""Returns a dictionary of stats."""
|
|
||||||
result = {}
|
|
||||||
|
|
||||||
parsed = USER_TIME_PATTERN.extract(inp)
|
|
||||||
result['user_time_secs'] = float(parsed['user_time'])
|
|
||||||
|
|
||||||
parsed = SYSTEM_TIME_PATTERN.extract(inp)
|
|
||||||
result['system_time_secs'] = float(parsed['system_time'])
|
|
||||||
|
|
||||||
parsed = WALL_TIME_PATTERN.extract(inp)
|
|
||||||
result['wall_time_secs'] = float(parsed['hours'] or 0.) * 3600 + float(
|
|
||||||
parsed['minutes']) * 60 + float(parsed['seconds'])
|
|
||||||
|
|
||||||
parsed = MAX_RES_SET_SIZE_PATTERN.extract(inp)
|
|
||||||
result['max_set_size_kbytes'] = int(parsed['max_set_size'])
|
|
||||||
|
|
||||||
parsed = CONSTRAINT_STATS_PATTERN.extract_last_occurence(inp)
|
|
||||||
print parsed
|
|
||||||
result['constraints_count'] = int(parsed[0])
|
|
||||||
result['constraints_score_min'] = float(parsed[1])
|
|
||||||
result['constraints_score_max'] = float(parsed[2])
|
|
||||||
result['constraints_score_mean'] = float(parsed[3])
|
|
||||||
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def extract_ground_truth_stats(input_text):
|
|
||||||
"""Returns a dictionary of ground truth stats."""
|
|
||||||
result = {}
|
|
||||||
parsed = GROUND_TRUTH_STATS_PATTERN.extract(input_text)
|
|
||||||
for name in ('abs_trans_err', 'sqr_trans_err', 'abs_rot_err', 'sqr_rot_err'):
|
|
||||||
result['ground_truth_{}'.format(name)] = float(parsed[name])
|
|
||||||
result['ground_truth_{}_dev'.format(name)] = float(
|
|
||||||
parsed['{}_dev'.format(name)])
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def retrieve_entity(datastore_client, kind, identifier):
|
|
||||||
"""Convenience function for Datastore entity retrieval."""
|
|
||||||
key = datastore_client.key(kind, identifier)
|
|
||||||
return datastore_client.get(key)
|
|
||||||
|
|
||||||
|
|
||||||
def create_job_selector(worker_id, num_workers):
|
|
||||||
"""Constructs a round-robin job selector."""
|
|
||||||
return lambda job_id: job_id % num_workers == worker_id
|
|
||||||
|
|
||||||
|
|
||||||
def run_cmd(cmd):
|
|
||||||
"""Runs command both printing its stdout output and returning it as string."""
|
|
||||||
print cmd
|
|
||||||
p = subprocess.Popen(
|
|
||||||
cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
|
||||||
run_cmd.output = []
|
|
||||||
|
|
||||||
def process(line):
|
|
||||||
run_cmd.output.append(line)
|
|
||||||
print line.rstrip()
|
|
||||||
|
|
||||||
while p.poll() is None:
|
|
||||||
process(p.stdout.readline())
|
|
||||||
process(p.stdout.read())
|
|
||||||
return '\n'.join(run_cmd.output)
|
|
||||||
|
|
||||||
|
|
||||||
def run_ros_cmd(ros_distro, ros_cmd):
|
|
||||||
"""Runs command similar to run_cmd but sets ROS environment variables."""
|
|
||||||
cmd = ('/bin/bash -c \"source /opt/ros/{}/setup.bash && source '
|
|
||||||
'/opt/cartographer_ros/setup.bash && {}\"').format(
|
|
||||||
ros_distro, ros_cmd)
|
|
||||||
return run_cmd(cmd)
|
|
||||||
|
|
||||||
|
|
||||||
class Job(object):
|
|
||||||
"""Represents a single job to be executed.
|
|
||||||
|
|
||||||
A job consists of a combination of rosbag and configuration and launch files.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, datastore_client, job_id):
|
|
||||||
self.id = job_id
|
|
||||||
entity = retrieve_entity(datastore_client, 'Job', job_id)
|
|
||||||
self.launch_file = entity['launch_file']
|
|
||||||
self.assets_writer_launch_file = entity['assets_writer_launch_file']
|
|
||||||
self.assets_writer_config_file = entity['assets_writer_config_file']
|
|
||||||
self.rosbag = entity['rosbag']
|
|
||||||
self.ros_package = entity['ros_package']
|
|
||||||
|
|
||||||
def __repr__(self):
|
|
||||||
return 'Job: id : {} launch_file: {} rosbag: {}'.format(
|
|
||||||
self.id, self.launch_file, self.rosbag)
|
|
||||||
|
|
||||||
def run(self, ros_distro, run_id):
|
|
||||||
"""Runs the job with ROS distro 'ros_distro'."""
|
|
||||||
print 'running job {}'.format(self.id)
|
|
||||||
|
|
||||||
# Garbage collect any left-overs from previous runs.
|
|
||||||
run_cmd('rm -rf /data/*')
|
|
||||||
|
|
||||||
# Copy the rosbag to scratch space
|
|
||||||
scratch_dir = '/data/{}'.format(self.id)
|
|
||||||
rosbag_filename = basename(self.rosbag)
|
|
||||||
run_cmd('mkdir {}'.format(scratch_dir))
|
|
||||||
run_cmd('gsutil cp gs://{} {}/{}'.format(self.rosbag, scratch_dir,
|
|
||||||
rosbag_filename))
|
|
||||||
|
|
||||||
# Creates pbstream
|
|
||||||
output = run_ros_cmd(ros_distro,
|
|
||||||
'/usr/bin/time -v roslaunch {} {} '
|
|
||||||
'bag_filenames:={}/{} no_rviz:=true'.format(
|
|
||||||
self.ros_package, self.launch_file, scratch_dir,
|
|
||||||
rosbag_filename))
|
|
||||||
info = extract_stats(output)
|
|
||||||
|
|
||||||
# Creates assets.
|
|
||||||
run_ros_cmd(
|
|
||||||
ros_distro, '/usr/bin/time -v roslaunch {} {} '
|
|
||||||
'bag_filenames:={}/{} pose_graph_filename:='
|
|
||||||
'{}/{}.pbstream config_file:={}'.format(
|
|
||||||
self.ros_package, self.assets_writer_launch_file, scratch_dir,
|
|
||||||
rosbag_filename, scratch_dir, rosbag_filename,
|
|
||||||
self.assets_writer_config_file))
|
|
||||||
|
|
||||||
# Copies assets to bucket.
|
|
||||||
run_cmd('gsutil cp {}/{}.pbstream '
|
|
||||||
'gs://cartographer-ci-artifacts/{}/{}/{}.pbstream'.format(
|
|
||||||
scratch_dir, rosbag_filename, run_id, self.id, rosbag_filename))
|
|
||||||
run_cmd('gsutil cp {}/{}_* gs://cartographer-ci-artifacts/{}/{}/'.format(
|
|
||||||
scratch_dir, rosbag_filename, run_id, self.id))
|
|
||||||
|
|
||||||
# Download ground truth relations file.
|
|
||||||
run_cmd('gsutil cp gs://cartographer-ci-ground-truth/{}/relations.pb '
|
|
||||||
'{}/relations.pb'.format(self.id, scratch_dir))
|
|
||||||
|
|
||||||
# Calculate metrics.
|
|
||||||
output = run_ros_cmd(ros_distro, 'cartographer_compute_relations_metrics '
|
|
||||||
'-relations_filename {}/relations.pb '
|
|
||||||
'-pose_graph_filename {}/{}.pbstream'.format(
|
|
||||||
scratch_dir, scratch_dir, rosbag_filename))
|
|
||||||
|
|
||||||
# Add ground truth stats.
|
|
||||||
info.update(extract_ground_truth_stats(output))
|
|
||||||
|
|
||||||
info['rosbag'] = rosbag_filename
|
|
||||||
return info
|
|
||||||
|
|
||||||
|
|
||||||
class Worker(object):
|
|
||||||
"""Represents a single worker that executes a sequence of Jobs."""
|
|
||||||
|
|
||||||
def __init__(self, datastore_client, pipeline_id, run_id):
|
|
||||||
entity = retrieve_entity(datastore_client, 'PipelineConfig', pipeline_id)
|
|
||||||
self.pipeline_id = pipeline_id
|
|
||||||
self.jobs = [Job(datastore_client, job_id) for job_id in entity['jobs']]
|
|
||||||
self.scratch_dir = entity['scratch_dir']
|
|
||||||
self.ros_distro = entity['ros_distro']
|
|
||||||
self.run_id = run_id
|
|
||||||
|
|
||||||
def __repr__(self):
|
|
||||||
result = 'Worker: pipeline_id: {}\n'.format(self.pipeline_id)
|
|
||||||
for job in self.jobs:
|
|
||||||
result += '{}\n'.format(str(job))
|
|
||||||
return result
|
|
||||||
|
|
||||||
def run_jobs(self, selector):
|
|
||||||
outputs = {}
|
|
||||||
for idx, job in enumerate(self.jobs):
|
|
||||||
if selector(idx):
|
|
||||||
output = job.run(self.ros_distro, self.run_id)
|
|
||||||
outputs[job.id] = output
|
|
||||||
else:
|
|
||||||
print 'job {}: skip'.format(job.id)
|
|
||||||
return outputs
|
|
||||||
|
|
||||||
|
|
||||||
def publish_stats_to_big_query(stats_dict, now, head_sha1):
|
|
||||||
"""Publishes metrics to BigQuery."""
|
|
||||||
bigquery_client = bigquery.Client()
|
|
||||||
dataset = bigquery_client.dataset('Cartographer')
|
|
||||||
table = dataset.table('metrics')
|
|
||||||
rows_to_insert = []
|
|
||||||
for job_identifier, job_info in stats_dict.iteritems():
|
|
||||||
print job_info
|
|
||||||
data = ('{}-{}-{}'.format(
|
|
||||||
now.year, now.month,
|
|
||||||
now.day), head_sha1, job_identifier, job_info['rosbag'],
|
|
||||||
job_info['user_time_secs'], job_info['system_time_secs'],
|
|
||||||
job_info['wall_time_secs'], job_info['max_set_size_kbytes'],
|
|
||||||
job_info['constraints_count'], job_info['constraints_score_min'],
|
|
||||||
job_info['constraints_score_max'],
|
|
||||||
job_info['constraints_score_mean'],
|
|
||||||
job_info['ground_truth_abs_trans_err'],
|
|
||||||
job_info['ground_truth_abs_trans_err_dev'],
|
|
||||||
job_info['ground_truth_sqr_trans_err'],
|
|
||||||
job_info['ground_truth_sqr_trans_err_dev'],
|
|
||||||
job_info['ground_truth_abs_rot_err'],
|
|
||||||
job_info['ground_truth_abs_rot_err_dev'],
|
|
||||||
job_info['ground_truth_sqr_rot_err'],
|
|
||||||
job_info['ground_truth_sqr_rot_err_dev'])
|
|
||||||
|
|
||||||
rows_to_insert.append(data)
|
|
||||||
|
|
||||||
errors = bigquery_client.create_rows(
|
|
||||||
table, rows_to_insert, selected_fields=SCHEMA)
|
|
||||||
if not errors:
|
|
||||||
print 'Pushed {} row(s) into Cartographer:metrics'.format(
|
|
||||||
len(rows_to_insert))
|
|
||||||
else:
|
|
||||||
print 'Errors:'
|
|
||||||
pprint(errors)
|
|
||||||
|
|
||||||
|
|
||||||
def parse_arguments():
|
|
||||||
"""Parses the command line arguments."""
|
|
||||||
parser = argparse.ArgumentParser(
|
|
||||||
description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter)
|
|
||||||
parser.add_argument('--worker_id', type=int)
|
|
||||||
parser.add_argument('--num_workers', type=int)
|
|
||||||
parser.add_argument('--pipeline_id', type=str)
|
|
||||||
return parser.parse_args()
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
|
||||||
args = parse_arguments()
|
|
||||||
ds_client = datastore.Client()
|
|
||||||
job_selector = create_job_selector(int(args.worker_id), int(args.num_workers))
|
|
||||||
head_sha1 = get_head_git_sha1()
|
|
||||||
now = datetime.datetime.now()
|
|
||||||
pipeline_run_id = '{}-{}-{}_{}'.format(now.year, now.month, now.day,
|
|
||||||
head_sha1)
|
|
||||||
worker = Worker(ds_client, args.pipeline_id, pipeline_run_id)
|
|
||||||
stats_dict = worker.run_jobs(job_selector)
|
|
||||||
publish_stats_to_big_query(stats_dict, now, head_sha1)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
|
||||||
main()
|
|
Loading…
Reference in New Issue