Compare commits

...

8 Commits

15 changed files with 238 additions and 511 deletions

1
.gitignore vendored
View File

@ -10,7 +10,6 @@
*.pyc
*.swp
*~
Makefile
build
build-*
.deps

12
Makefile Normal file
View File

@ -0,0 +1,12 @@
docker:
docker build -t fail-base -f fail-base.dockerfile . --build-arg CACHE_DATE="$(shell date)"
docker build -t fail-generic-tracing -f fail-generic-tracing.dockerfile .
docker build -t fail-demo -f fail-demo.dockerfile . --build-arg CACHE_DATE="$(shell date)"
run: .compose ssh
.compose:
docker compose up -d --force-recreate --renew-anon-volumes -y
ssh:
ssh -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -p 5022 fail@127.0.0.1

View File

@ -7,13 +7,13 @@ services:
MYSQL_USER: fail
MYSQL_PASSWORD: fail
MYSQL_DATABASE: fail
# ports:
# - "3306:3306"
ports:
- "3306:3306"
networks:
- fail-network
fail-demo:
image: danceos/fail-demo
image: fail-demo
container_name: fail-demo
ports:
- "5000:5000" # Result Browser

View File

@ -119,10 +119,14 @@ RUN wget http://www.aspectc.org/releases/"$acversion"/ac-bin-linux-x86-64bit-"$a
ENV PATH=/home/fail/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
# Clone FAIL*
# Break docker layer cache
ARG CACHE_DATE=1970-01-01
# RUN git clone https://github.com/danceos/fail.git
RUN git clone https://gitea.vps.chriphost.de/christoph/fail
WORKDIR fail
USER root
# Accept SSH connections
EXPOSE 22
CMD ["/usr/sbin/sshd", "-D"]

View File

@ -3,7 +3,7 @@
# generic-tracing experiment was already built and the binaries are in
# place (~fail/bin/*)
FROM danceos/fail-generic-tracing
FROM fail-generic-tracing
LABEL org.opencontainers.image.authors="Christian Dietrich <stettberger@dokucode.de>"
@ -14,10 +14,7 @@ RUN chown fail /home/fail/.my.cnf
USER fail
WORKDIR /home/fail
RUN echo 'export PATH=$HOME/bin:$PATH' >> ~/.profile \
&& echo 'cd $HOME/fail-targets' >> ~/.profile
# RUN git clone https://github.com/danceos/fail-targets.git
RUN git clone https://gitea.vps.chriphost.de/christoph/fail-targets
&& echo 'cd $HOME/fail-wasm/examples' >> ~/.profile
WORKDIR fail
RUN mkdir build; cd build
@ -135,6 +132,7 @@ RUN apt-get update \
unzip \
&& apt-get clean
# Fix old shebangs
RUN ln -sf /usr/bin/python2 /usr/bin/python
# Install python packages
@ -143,5 +141,17 @@ RUN wget https://bootstrap.pypa.io/pip/2.7/get-pip.py \
&& wget https://raw.githubusercontent.com/paulfitz/mysql-connector-c/master/include/my_config.h -O /usr/include/mysql/my_config.h \
&& pip2 install flask pyyaml MySQL-python
# For the resultbrowser, we expose port 5000 to the outside world.
USER fail
# Clone FAIL* targets
# Break docker layer cache
ARG CACHE_DATE=1970-01-01
WORKDIR /home/fail
# RUN git clone https://github.com/danceos/fail-targets.git
RUN git clone https://gitea.vps.chriphost.de/christoph/fail-targets
RUN git clone https://gitea.vps.chriphost.de/christoph/fail-wasm
USER root
# Resultbrowser
EXPOSE 5000

View File

@ -1,6 +1,6 @@
# Inherit from docker container that has the fail source code prepared,
# including all tools which are needed to build FAIL*
FROM danceos/fail-base
FROM fail-base
LABEL org.opencontainers.image.authors="Christian Dietrich <stettberger@dokucode.de>"
@ -23,13 +23,13 @@ RUN cmake \
-DBUILD_CONVERT_TRACE=OFF \
-DBUILD_DATA_AGGREGATOR=OFF \
-DBUILD_DUMP_HOPS=OFF \
-DBUILD_DUMP_TRACE=OFF \
-DBUILD_DUMP_TRACE=ON \
-DBUILD_FAULTSPACEPLOT=OFF \
-DBUILD_GEM5=OFF \
-DBUILD_IMPORT_TRACE=OFF \
-DBUILD_IMPORT_TRACE=ON \
-DBUILD_LLVM_DISASSEMBLER=ON \
-DBUILD_PANDA=OFF \
-DBUILD_PRUNE_TRACE=OFF \
-DBUILD_PRUNE_TRACE=ON \
-DBUILD_QEMU=OFF \
-DBUILD_T32=OFF \
-DBUILD_X86=ON \

59
flake.lock generated
View File

@ -1,59 +0,0 @@
{
"nodes": {
"flake-utils": {
"inputs": {
"systems": "systems"
},
"locked": {
"lastModified": 1731533236,
"narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=",
"owner": "numtide",
"repo": "flake-utils",
"rev": "11707dc2f618dd54ca8739b309ec4fc024de578b",
"type": "github"
},
"original": {
"owner": "numtide",
"repo": "flake-utils",
"type": "github"
}
},
"nixpkgs": {
"locked": {
"lastModified": 1768032153,
"narHash": "sha256-6kD1MdY9fsE6FgSwdnx29hdH2UcBKs3/+JJleMShuJg=",
"owner": "NixOS",
"repo": "nixpkgs",
"rev": "3146c6aa9995e7351a398e17470e15305e6e18ff",
"type": "github"
},
"original": {
"id": "nixpkgs",
"type": "indirect"
}
},
"root": {
"inputs": {
"flake-utils": "flake-utils",
"nixpkgs": "nixpkgs"
}
},
"systems": {
"locked": {
"lastModified": 1681028828,
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
"owner": "nix-systems",
"repo": "default",
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
"type": "github"
},
"original": {
"owner": "nix-systems",
"repo": "default",
"type": "github"
}
}
},
"root": "root",
"version": 7
}

334
flake.nix
View File

@ -1,334 +0,0 @@
rec {
description = "FAIL* - Fault Injection Leveraged";
inputs = {
nixpkgs.url = "nixpkgs"; # Use nixpkgs from system registry
flake-utils.url = "github:numtide/flake-utils";
};
outputs = {
self,
nixpkgs,
flake-utils,
}:
# Create a shell (and possibly package) for each possible system, not only x86_64-linux
flake-utils.lib.eachDefaultSystem (system: let
pkgs = import nixpkgs {
inherit system;
config.allowUnfree = true;
overlays = [];
};
inherit (pkgs) lib stdenv;
# ===========================================================================================
# Define custom dependencies
# ===========================================================================================
# 64 bit C/C++ compilers that don't collide (use the same libc)
bintools = pkgs.wrapBintoolsWith {
bintools = pkgs.bintools.bintools; # Unwrapped bintools
libc = pkgs.glibc;
};
gcc = lib.hiPrio (pkgs.wrapCCWith {
cc = pkgs.gcc.cc; # Unwrapped gcc
libc = pkgs.glibc;
bintools = bintools;
});
clang = pkgs.wrapCCWith {
cc = pkgs.clang.cc; # Unwrapped clang
libc = pkgs.glibc;
bintools = bintools;
};
# Multilib C/C++ compilers that don't collide (use the same libc)
# bintools_multilib = pkgs.wrapBintoolsWith {
# bintools = pkgs.bintools.bintools; # Unwrapped bintools
# libc = pkgs.glibc_multi;
# };
# gcc_multilib = lib.hiPrio (pkgs.wrapCCWith {
# cc = pkgs.gcc.cc; # Unwrapped gcc
# libc = pkgs.glibc_multi;
# bintools = bintools_multilib;
# });
# clang_multilib = pkgs.wrapCCWith {
# cc = pkgs.clang.cc; # Unwrapped clang
# libc = pkgs.glibc_multi;
# bintools = bintools_multilib;
# };
aspectcxx = stdenv.mkDerivation rec {
pname = "aspectcxx";
version = "1.2";
src = pkgs.fetchurl {
url = "http://www.aspectc.org/releases/${version}/ac-bin-linux-x86-64bit-${version}.tar.gz";
sha256 = "sha256-8wOrPYC99E3aV5/Js2EBI4V/OoD9y10fYZt8ikPtvt4="; # 1.2
# sha256 = "sha256-gmqoZFkPvs60b6yuMHNtYvuGJXGnbwAYEiyXxp/fmPI="; # 2.0
# sha256 = "sha256-+rEflemXNwn4XZZwSOqCsr6/KFGV8wxW6PeXzHHUK0o="; # 2.5
};
nativeBuildInputs = [
pkgs.autoPatchelfHook
];
unpackPhase = ''
tar xvzf $src
'';
installPhase = ''
mkdir -p $out/bin
cp aspectc++/ac++ $out/bin/
cp aspectc++/ag++ $out/bin/
'';
};
libpcl = stdenv.mkDerivation rec {
pname = "libpcl";
version = "1.12-2";
src = pkgs.fetchurl {
url = "https://launchpadlibrarian.net/521269537/libpcl1_1.12-2_amd64.deb";
sha256 = "sha256-GL3mjPAccAtRMAJPnDMCHiDf6xNvGi4oUWylOIqBjP0=";
};
nativeBuildInputs = with pkgs; [
dpkg
autoPatchelfHook
];
unpackPhase = ''
dpkg-deb -x $src .
'';
installPhase = ''
mkdir -p $out
cp -r usr/* $out/
'';
};
libpcl-dev = stdenv.mkDerivation rec {
pname = "libpcl-dev";
version = "1.12-2";
src = pkgs.fetchurl {
url = "https://launchpadlibrarian.net/521269536/libpcl1-dev_${version}_amd64.deb";
sha256 = "sha256-Z1wP0K8hfV1f9ypee9XIx6H0JOTidhtXDBe82mlRaOg=";
};
nativeBuildInputs = with pkgs; [
dpkg
autoPatchelfHook
];
unpackPhase = ''
dpkg-deb -x $src .
'';
installPhase = ''
mkdir -p $out
cp -r usr/* $out/
# Hacky bullshit
cp ${libpcl}/lib/x86_64-linux-gnu/libpcl.so.1.0.11 $out/lib/x86_64-linux-gnu/libpcl.so.1.0.11
rm $out/share/doc/libpcl1-dev/changelog.Debian.gz
'';
};
# ===========================================================================================
# Specify dependencies
# https://nixos.org/manual/nixpkgs/stable/#ssec-stdenv-dependencies-overview
# Just for a "nix develop" shell, buildInputs can be used for everything.
# ===========================================================================================
# Add dependencies to nativeBuildInputs if they are executed during the build:
# - Those which are needed on $PATH during the build, for example cmake and pkg-config
# - Setup hooks, for example makeWrapper
# - Interpreters needed by patchShebangs for build scripts (with the --build flag), which can be the case for e.g. perl
nativeBuildInputs = with pkgs; [
# Languages:
bintools
gcc
# clang
aspectcxx
# C/C++:
gnumake
cmake
pkg-config
doxygen
];
# Add dependencies to buildInputs if they will end up copied or linked into the final output or otherwise used at runtime:
# - Libraries used by compilers, for example zlib
# - Interpreters needed by patchShebangs for scripts which are installed, which can be the case for e.g. perl
buildInputs = with pkgs; [
# C/C++:
libpcl-dev
libiberty
libelf
libdwarf
boost
llvmPackages_18.llvm
mariadb-connector-c
fontconfig
zlib
capstone
protobuf
binutils
# No clue what I need from those
xorg.libX11
xorg.libXext
xorg.libXrender
xorg.libXrandr
xorg.libXinerama
xorg.libXcursor
xorg.libXi
xorg.libXfixes
];
# ===========================================================================================
# Define buildable + installable packages
# ===========================================================================================
package = stdenv.mkDerivation {
inherit nativeBuildInputs buildInputs;
pname = "fail";
version = "1.0.1";
src = ./.;
enableParallelBuilding = true;
cmakeFlags = [
# Our CMake is too new :( Try it anyway, can still downgrade later...
"-DCMAKE_POLICY_VERSION_MINIMUM=3.5"
# AspectC++
# "-DCMAKE_AGPP_FLAGS=--c_compiler${clang}/bin/clang++"
"-DCMAKE_AGPP_FLAGS=-D__NO_MATH_INLINES"
# "-DCMAKE_AGPP_FLAGS=--c_compiler${clang}/bin/clang++ -D__NO_MATH_INLINES -D__STRICT_ANSI__"
# Tell CMake where the libs are
"-DLibIberty_INCLUDE_DIRS=${pkgs.libiberty}/include"
"-DLibIberty_LIBRARIES=${pkgs.libiberty}/lib/libiberty.a"
"-DLIBELF_INCLUDE_DIRS=${pkgs.libelf}/include"
"-DLIBELF_LIBRARIES=${pkgs.libelf}/lib/libelf.a"
"-DLIBDWARF_INCLUDE_DIRS=${pkgs.libdwarf}/include"
"-DLIBDWARF_LIBRARIES=${pkgs.libdwarf}/lib/libdwarf.a"
"-DCAPSTONE_INCLUDE_DIR=${pkgs.capstone}/include"
"-DCAPSTONE_LIBRARY=${pkgs.capstone}/lib/libcapstone.a"
"-DLIBPCL_LIBRARIES=${libpcl-dev}/lib/libpcl.a"
];
installPhase = ''
mkdir -p $out/bin
mv ./fail $out/bin/
'';
};
in rec {
# Provide package for "nix build"
defaultPackage = package;
defaultApp = flake-utils.lib.mkApp {
drv = defaultPackage;
};
# Provide environment for "nix develop"
devShell = pkgs.mkShell {
inherit nativeBuildInputs buildInputs;
name = description;
# =========================================================================================
# Define environment variables
# =========================================================================================
# Custom dynamic libraries:
# LD_LIBRARY_PATH = builtins.concatStringsSep ":" [
# # Rust Bevy GUI app:
# # "${pkgs.xorg.libX11}/lib"
# # "${pkgs.xorg.libXcursor}/lib"
# # "${pkgs.xorg.libXrandr}/lib"
# # "${pkgs.xorg.libXi}/lib"
# # "${pkgs.libGL}/lib"
#
# # JavaFX app:
# # "${pkgs.libGL}/lib"
# # "${pkgs.gtk3}/lib"
# # "${pkgs.glib.out}/lib"
# # "${pkgs.xorg.libXtst}/lib"
# ];
# Dynamic libraries from buildinputs:
LD_LIBRARY_PATH = nixpkgs.lib.makeLibraryPath buildInputs;
# Set matplotlib backend
# MPLBACKEND = "TkAgg";
# =========================================================================================
# Define shell environment
# =========================================================================================
# Setup the shell when entering the "nix develop" environment (bash script).
shellHook = let
mkCmakeScript = type: let
typeLower = lib.toLower type;
in
pkgs.writers.writeFish "cmake-${typeLower}.fish" ''
cd $FLAKE_PROJECT_ROOT
echo "Removing build directory ./cmake-build-${typeLower}/"
rm -rf ./cmake-build-${typeLower}
echo "Creating build directory"
mkdir cmake-build-${typeLower}
cd cmake-build-${typeLower}
echo "Running cmake"
cmake -G "Unix Makefiles" -DCMAKE_BUILD_TYPE="${type}" -DCMAKE_EXPORT_COMPILE_COMMANDS="On" ..
echo "Linking compile_commands.json"
cd ..
ln -sf ./cmake-build-${typeLower}/compile_commands.json ./compile_commands.json
'';
cmakeDebug = mkCmakeScript "Debug";
cmakeRelease = mkCmakeScript "Release";
mkBuildScript = type: let
typeLower = lib.toLower type;
in
pkgs.writers.writeFish "cmake-build.fish" ''
cd $FLAKE_PROJECT_ROOT/cmake-build-${typeLower}
echo "Running cmake"
cmake --build .
'';
buildDebug = mkBuildScript "Debug";
buildRelease = mkBuildScript "Release";
# Use this to specify commands that should be ran after entering fish shell
initProjectShell = pkgs.writers.writeFish "init-shell.fish" ''
echo "Entering \"${description}\" environment..."
# Determine the project root, used e.g. in cmake scripts
set -g -x FLAKE_PROJECT_ROOT (git rev-parse --show-toplevel)
# C/C++:
# abbr -a cmake-debug "${cmakeDebug}"
# abbr -a cmake-release "${cmakeRelease}"
# abbr -a build-debug "${buildDebug}"
# abbr -a build-release "${buildRelease}"
'';
in
builtins.concatStringsSep "\n" [
# Launch into pure fish shell
''
exec "$(type -p fish)" -C "source ${initProjectShell} && abbr -a menu '${pkgs.bat}/bin/bat "${initProjectShell}"'"
''
];
};
});
}

View File

@ -1,5 +1,5 @@
[client]
host=127.0.0.1
host=fail-db
user=fail
password=fail
database=fail

View File

@ -1,12 +0,0 @@
build-all:
docker build -t danceos/fail-base fail-base
docker build -t danceos/fail-generic-tracing fail-generic-tracing
docker build -t danceos/fail-demo fail-demo
run-all: .compose ssh
.compose:
docker compose up -d --force-recreate --renew-anon-volumes -y
ssh:
ssh -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -p 5022 fail@127.0.0.1

View File

@ -1,9 +1,12 @@
from pprint import pprint
from . import details
from . import model
from app.detaildealer import detaildealer
def scrub(table_name):
return ''.join( chr for chr in table_name if chr.isalnum() or chr == '_' )
return "".join(chr for chr in table_name if chr.isalnum() or chr == "_")
class Resulttype:
def __init__(self, name, count):
@ -16,28 +19,27 @@ class Resulttype:
def getCount(self):
return self.count
class Variant:
def __init__(self, id, name, table, benchmark, detail):
self.id = id
self.dbname = name
self.parenttable = table # TableDetails
self.details = detail # VariantDetails
self.benchmark = benchmark # BenchmarkDetails
self.parenttable = table # TableDetails
self.details = detail # VariantDetails
self.benchmark = benchmark # BenchmarkDetails
self.results = {}
self.totalresults = 0
def getMapper(self):
mapper = self.benchmark.getMapper()
if not mapper: #try benchmark mapper
if not mapper: # try benchmark mapper
mapper = self.details.getMapper()
if not mapper: # of not there, try parent tables mapper
if not mapper: # of not there, try parent tables mapper
mapper = self.parenttable.getMapper()
if not mapper: # no mapper found at all, try default mapper
mapper = model.detaildealer.getDefaultMapper()
if not mapper: # no mapper found at all, try default mapper
mapper = detaildealer.getDefaultMapper()
return mapper
def addResulttype(self, name, count):
mapper = self.getMapper()
label = mapper.getLabel(name)
@ -70,17 +72,28 @@ class Variant:
return self.totalresults
def __str__(self):
ret = "Variant: " + self.getDetails().getTitle() + " - " + self.getBenchmarkDetails().getTitle() +" (id: " + str( self.id )+ ")" + " "
ret += "Total Results: " + str( self.totalresults ) + "\n"
ret = (
"Variant: "
+ self.getDetails().getTitle()
+ " - "
+ self.getBenchmarkDetails().getTitle()
+ " (id: "
+ str(self.id)
+ ")"
+ " "
)
ret += "Total Results: " + str(self.totalresults) + "\n"
for v in self.results:
ret += "\t" + v.name + ": " + str( v.count ) + "\n"
ret += "\t" + v.name + ": " + str(v.count) + "\n"
return ret
__repr__ = __str__
'''A ResultTable contains n Variants'''
class ResultTable:
"""A ResultTable contains n Variants"""
class ResultTable:
def __init__(self, name, cfg):
self.name = scrub(name)
self.details = cfg.getTable(name)
@ -89,7 +102,7 @@ class ResultTable:
def addVariant(self, var):
if var.getId() in self.variants:
return
self.variants[var.getId()] = var # Add if not existing yet
self.variants[var.getId()] = var # Add if not existing yet
def getVariant(self, id):
if id in self.variants:
@ -97,7 +110,7 @@ class ResultTable:
return None
def getVariantById(self, varid):
for k,v in self.variants.items():
for k, v in self.variants.items():
if int(v.getId()) == int(varid):
return v
return None
@ -110,12 +123,16 @@ class ResultTable:
def __str__(self):
ret = "Result: " + self.getDetails().getTitle() + "\n"
for k,v in self.variants.items():
for k, v in self.variants.items():
ret += "\t" + str(v) + "\n"
return ret
__repr__ = __str__
'''Overview has n ResultTables'''
"""Overview has n ResultTables"""
class Overview:
def __init__(self):
self.tables = {}
@ -130,7 +147,7 @@ class Overview:
return self.tables.get(dbname, None)
def getVariantById(self, variant_id):
for key,table in self.tables.items():
for key, table in self.tables.items():
variant = table.getVariantById(variant_id)
if variant:
return variant
@ -139,5 +156,3 @@ class Overview:
def length(self):
return len(self.tables)

View File

@ -0,0 +1,4 @@
# Instantiate global detail dealer, will be initialized in reloadOverview
from app import details
detaildealer = details.DetailDealer()

View File

@ -1,31 +1,60 @@
#!/usr/bin/env python
import MySQLdb
import MySQLdb.cursors
import yaml
import sys
import os.path
import yaml
import sys
import MySQLdb
import MySQLdb.cursors
from app.detaildealer import detaildealer
from pprint import pprint
from . import data
from . import details
from . import data, details
"""Get command line options"""
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-c", "--conf", type="string", help="MySQL config file", dest="config", default= os.path.join(os.path.expanduser("~"),".my.cnf"))
parser.add_option("-s", "--host", type="string", help="Webserver hostname", dest="host", default="localhost")
parser.add_option("-d", "--details", type="string", help="Detailed information (YAML configuration file)", dest="details", default=None)
parser.add_option("-p", "--port", type="string", help="Webserver port", dest="port", default="5000")
parser.add_option(
"-c",
"--conf",
type="string",
help="MySQL config file",
dest="config",
default=os.path.join(os.path.expanduser("~"), ".my.cnf"),
)
parser.add_option(
"-s",
"--host",
type="string",
help="Webserver hostname",
dest="host",
default="localhost",
)
parser.add_option(
"-d",
"--details",
type="string",
help="Detailed information (YAML configuration file)",
dest="details",
default=None,
)
parser.add_option(
"-p", "--port", type="string", help="Webserver port", dest="port", default="5000"
)
opts, args = parser.parse_args()
"""Check if configuration files exist"""
def checkConfigFile(msg, fname):
if not os.path.isfile(fname):
sys.exit("Error: '" + fname + "' not found")
else:
print(msg, "->", fname)
# Check sql config
sqlconfig = opts.config
checkConfigFile("MySQL config", sqlconfig)
@ -35,50 +64,74 @@ if opts.details:
checkConfigFile("Details", opts.details)
# Instantiate global detail dealer, will be initialized in reloadOverview
detaildealer = details.DetailDealer()
# detaildealer = details.DetailDealer()
"""Remove all characters from string except alphanuermics and _"""
def scrub(table_name):
return ''.join( chr for chr in table_name if chr.isalnum() or chr == '_' )
return "".join(chr for chr in table_name if chr.isalnum() or chr == "_")
"""Global mysql handles"""
db = None
cur = None
def loadSession(dbconf):
global db
if db:
db.close()
db = MySQLdb.connect(read_default_file=dbconf, cursorclass=MySQLdb.cursors.DictCursor)
db = MySQLdb.connect(
read_default_file=dbconf, cursorclass=MySQLdb.cursors.DictCursor
)
return db.cursor()
def closeSession():
if cur: cur.close()
if cur:
cur.close()
global db
db.close()
db = None
'''Populate variant results for overview data'''
"""Populate variant results for overview data"""
def getVariants(cur, table):
restbl = table.getDetails().getDBName()
cur.execute("""SELECT sum((t.time2 - t.time1 + 1) * width) AS total, resulttype,variant, v.id as variant_id, benchmark, details FROM variant v JOIN trace t ON v.id = t.variant_id JOIN fspgroup g ON g.variant_id = t.variant_id AND g.instr2 = t.instr2 AND g.data_address = t.data_address JOIN %s r ON r.pilot_id = g.pilot_id JOIN fsppilot p ON r.pilot_id = p.id GROUP BY v.id, resulttype, details""" % (restbl)) # % is used here, as a tablename must not be quoted
cur.execute(
"""SELECT sum((t.time2 - t.time1 + 1) * width) AS total, resulttype,variant, v.id as variant_id, benchmark, details FROM variant v JOIN trace t ON v.id = t.variant_id JOIN fspgroup g ON g.variant_id = t.variant_id AND g.instr2 = t.instr2 AND g.data_address = t.data_address JOIN %s r ON r.pilot_id = g.pilot_id JOIN fsppilot p ON r.pilot_id = p.id GROUP BY v.id, resulttype, details"""
% (restbl)
) # % is used here, as a tablename must not be quoted
res = cur.fetchall()
rdic = {}
# Build dict with variant id as key
for r in res:
# if variant entry already exists:
variant = table.getVariant(int(r['variant_id']))
if not variant: # if variant did not exist yet, create it:
variant_details = detaildealer.getVariant(restbl, r['variant'])
benchmark_details = detaildealer.getBenchmark(restbl, r['variant'], r['benchmark'])
variant = table.getVariant(int(r["variant_id"]))
if not variant: # if variant did not exist yet, create it:
variant_details = detaildealer.getVariant(restbl, r["variant"])
benchmark_details = detaildealer.getBenchmark(
restbl, r["variant"], r["benchmark"]
)
table_details = detaildealer.getTable(restbl)
variant = data.Variant(int(r['variant_id']), r['variant'], table_details, benchmark_details, variant_details)
variant.addResulttype(r['resulttype'], r['total'])
variant = data.Variant(
int(r["variant_id"]),
r["variant"],
table_details,
benchmark_details,
variant_details,
)
variant.addResulttype(r["resulttype"], r["total"])
table.addVariant(variant)
'''Get overview data for index page'''
"""Get overview data for index page"""
def reloadOverview():
overview = data.Overview()
detaildealer.reload(opts.details)
@ -89,33 +142,42 @@ def reloadOverview():
for rdic in result_tables:
# r is the tablename, -> result_FOOBAR
for key, tablename in rdic.items():
table = data.ResultTable(tablename,detaildealer)
table = data.ResultTable(tablename, detaildealer)
getVariants(cur, table)
overview.add(table)
# Check if objdump table exists
cur.execute("SHOW TABLES like 'objdump'")
objdump_exists = (len(cur.fetchall()) == 1)
objdump_exists = len(cur.fetchall()) == 1
closeSession()
return overview, objdump_exists
"""Load overview data at server startup"""
print("Loading overview data from database. This may take a while ...")
overview_data, objdump_exists = reloadOverview()
print("done.")
## Get overview data for views.index()
def getOverview():
return overview_data
def objdumpExists():
return objdump_exists
"""Get Results for one variant id"""
def getVariantResult(table, variantid):
cur = loadSession(sqlconfig)
restbl = scrub(table)
stmt = "SELECT resulttype, count(*) as total from %s r join fsppilot on r.pilot_id=fsppilot.id join variant on fsppilot.variant_id=variant.id" % (restbl)
stmt = (
"SELECT resulttype, count(*) as total from %s r join fsppilot on r.pilot_id=fsppilot.id join variant on fsppilot.variant_id=variant.id"
% (restbl)
)
where = " WHERE variant.id = %s group by resulttype ORDER BY resulttype "
stmt = stmt + where
cur.execute(stmt, variantid)
@ -123,10 +185,13 @@ def getVariantResult(table, variantid):
closeSession()
return res
'''Show objdump together with according injection result types.'''
"""Show objdump together with according injection result types."""
def getCode(result_table, variant_id, resultlabel=None):
result_table = scrub(result_table)
filt = ''
filt = ""
if not variant_id or not result_table:
return None
variant = overview_data.getVariantById(variant_id)
@ -137,15 +202,18 @@ def getCode(result_table, variant_id, resultlabel=None):
filt = " and ( "
for dbn in dbnames[:-1]:
filt += "resulttype = '" + dbn + "' OR "
filt += "resulttype = '" + dbnames[-1] +"' ) "
filt += "resulttype = '" + dbnames[-1] + "' ) "
else:
filt = " and resulttype = '" + resultlabel + "' "
# I especially like this one:
select = "SELECT instr_address, opcode, disassemble, comment, sum(t.time2 - t.time1 + 1) as totals, GROUP_CONCAT(DISTINCT resulttype SEPARATOR ', ') as results FROM variant v "
join = " JOIN trace t ON v.id = t.variant_id JOIN fspgroup g ON g.variant_id = t.variant_id AND g.instr2 = t.instr2 AND g.data_address = t.data_address JOIN %s r ON r.pilot_id = g.pilot_id JOIN fsppilot p ON r.pilot_id = p.id JOIN objdump ON objdump.variant_id = v.id AND objdump.instr_address = injection_instr_absolute " %(scrub(result_table))
where = "WHERE v.id = %s "
group = "GROUP BY injection_instr_absolute ORDER BY totals DESC "
join = (
" JOIN trace t ON v.id = t.variant_id JOIN fspgroup g ON g.variant_id = t.variant_id AND g.instr2 = t.instr2 AND g.data_address = t.data_address JOIN %s r ON r.pilot_id = g.pilot_id JOIN fsppilot p ON r.pilot_id = p.id JOIN objdump ON objdump.variant_id = v.id AND objdump.instr_address = injection_instr_absolute "
% (scrub(result_table))
)
where = "WHERE v.id = %s "
group = "GROUP BY injection_instr_absolute ORDER BY totals DESC "
cur = loadSession(sqlconfig)
stmt = select + join + where + filt + group
@ -153,27 +221,35 @@ def getCode(result_table, variant_id, resultlabel=None):
dump = cur.fetchall()
closeSession()
resulttypes = variant.getResultLabels()
resulttypes = variant.getResultLabels()
return dump, resulttypes
def getCodeExcerpt(variant_id, instr_addr):
code = {}
limit = 8
cur = loadSession(sqlconfig)
cur.execute( """(SELECT instr_address, opcode, disassemble, comment FROM objdump \
cur.execute(
"""(SELECT instr_address, opcode, disassemble, comment FROM objdump \
WHERE instr_address < %s AND variant_id = %s \
ORDER BY instr_address DESC LIMIT %s) \
ORDER BY instr_address ASC""" , (instr_addr, variant_id, limit))
ORDER BY instr_address ASC""",
(instr_addr, variant_id, limit),
)
below = cur.fetchall()
code['below'] = below
cur.execute("""SELECT instr_address, opcode, disassemble, comment FROM objdump \
code["below"] = below
cur.execute(
"""SELECT instr_address, opcode, disassemble, comment FROM objdump \
WHERE instr_address >= %s AND variant_id = %s \
ORDER BY instr_address ASC LIMIT %s""", (instr_addr, variant_id, limit+1))
ORDER BY instr_address ASC LIMIT %s""",
(instr_addr, variant_id, limit + 1),
)
upper = cur.fetchall()
code['upper'] = upper
code["upper"] = upper
closeSession()
return code
def getResultsbyInstruction(result_table, variant_id, instr_addr, resultlabel=None):
restypefilter = None
if resultlabel:
@ -185,12 +261,15 @@ def getResultsbyInstruction(result_table, variant_id, instr_addr, resultlabel=No
restypefilter = " and ( "
for dbn in dbnames[:-1]:
restypefilter += "resulttype = '" + dbn + "' OR "
restypefilter += "resulttype = '" + dbnames[-1] +"' ) "
restypefilter += "resulttype = '" + dbnames[-1] + "' ) "
select = "SELECT bitoffset as 'Bit Offset', hex(injection_instr_absolute) as 'Instruction Address', hex(original_value) as 'Original Value', hex(data_address) as 'Data Address', resulttype as 'Result Type', details as 'Details' from %s " % scrub(result_table)
join = "JOIN fsppilot ON pilot_id = fsppilot.id "
where = "WHERE variant_id = %s and injection_instr_absolute = %s "
order = "ORDER BY data_address, bitoffset"
select = (
"SELECT bitoffset as 'Bit Offset', hex(injection_instr_absolute) as 'Instruction Address', hex(original_value) as 'Original Value', hex(data_address) as 'Data Address', resulttype as 'Result Type', details as 'Details' from %s "
% scrub(result_table)
)
join = "JOIN fsppilot ON pilot_id = fsppilot.id "
where = "WHERE variant_id = %s and injection_instr_absolute = %s "
order = "ORDER BY data_address, bitoffset"
cur = loadSession(sqlconfig)
if not restypefilter:
@ -204,8 +283,7 @@ def getResultsbyInstruction(result_table, variant_id, instr_addr, resultlabel=No
closeSession()
return res
def showDBstatus():
res = "TODO"
return res

View File

@ -1,42 +1,53 @@
from flask import render_template,request
from flask import render_template, request
from app import app
# import model
# import data
from . import model
from . import data
@app.route('/')
@app.route('/index')
@app.route("/")
@app.route("/index")
def index():
reload_overview = request.args.get('reload', False)
reload_overview = request.args.get("reload", False)
if reload_overview:
print("Reloading overview...")
model.reloadOverview()
return render_template("index.html", overview=model.getOverview(), objdump_there = model.objdumpExists())
return render_template(
"index.html", overview=model.getOverview(), objdump_there=model.objdumpExists()
)
@app.route('/code')
@app.route("/code")
def code():
variant_id = request.args.get('variant_id', None)
resulttype = request.args.get('resulttype', None)
table = request.args.get('table', None)
res,restypes = model.getCode(table, variant_id, resulttype)
variant_id = request.args.get("variant_id", None)
resulttype = request.args.get("resulttype", None)
table = request.args.get("table", None)
res, restypes = model.getCode(table, variant_id, resulttype)
var_dets = model.getOverview().getVariantById(variant_id)
return render_template("code.html", results=res, resulttypes=restypes, variant_details=var_dets )
return render_template(
"code.html", results=res, resulttypes=restypes, variant_details=var_dets
)
@app.route('/instr_details')
@app.route("/instr_details")
def instr_details():
table = request.args.get('table', None)
variant_id = request.args.get('variant_id', None)
instr_addr = request.args.get('instr_address', None)
resulttype = request.args.get('resulttype', None)
table = request.args.get("table", None)
variant_id = request.args.get("variant_id", None)
instr_addr = request.args.get("instr_address", None)
resulttype = request.args.get("resulttype", None)
codeexcerpt = model.getCodeExcerpt(variant_id, instr_addr)
var_dets = model.getOverview().getVariantById(variant_id)
results = model.getResultsbyInstruction(table, variant_id, instr_addr, resulttype)
return render_template("instr_details.html", code=codeexcerpt, result=results, variant_details=var_dets)
return render_template(
"instr_details.html", code=codeexcerpt, result=results, variant_details=var_dets
)
@app.route('/about')
@app.route("/about")
def about():
stat = model.showDBstatus()
return render_template("about.html", status=stat)

View File

@ -1,5 +1,4 @@
#!/usr/bin/env python
from app import app
from app import model
from app import app, model
app.run(debug=False, port=int(model.opts.port), host=model.opts.host)