I am Robin Schmitt

Data Scientist with a passion for the transparent
and ethical application of artificial intelligence.

My work experience and education

This section provides shortcuts to details about my work experience, education and a blog-style outline of my Master's thesis.

Work experience

Academic career

Master's thesis

My work experience

At the moment, I am working as a data scientist at Mercedes-Benz,
more precisely on the #HeyMercedes in-car voice assistant.
My responsibilities include:

Data engineering
End-to-end responsibility for data collection pipeline
Technologies: PySpark, Databricks, SQL

User research
Conducting user research utilising customer data to gain strategic insights, providing BI tools
Technologies: Python, PowerBI

Machine learning
Applying machine learning algorithms on sequential interaction data to improve our voice assistant
Technologies: Python, Jupyter Notebooks, pm4py, sklearn, spacy, pymc3

A full tabular cv of all my positions within Mercedes-Benz is found here:

To - From Occupation
now - 2021 Data scientist for the Mercedes-Benz voice assistant:
End-to-end responsibility for data collection pipeline
Conducting user research utilising customer data to gain strategic insights
Applying machine learning algorithms on sequential interaction data to improve our voice assistant
2021 - 2018 Development engineer for hard- and software:
System architecture and strategy for in-vehicle infotainment
Multi-factor optimisation of future infotainment architectures based on historical data utilising predictive models
2018 - 2015 Project engineer in quality management:
Quality and maturity management in vehicle development projects
Integrating module maturity reports into actionable project reports

My academic career

In 2022, I graduated as Master of Science in the field of Data Science at Albstadt-Sigmaringen University.
Since then, I am continuously taking courses to further improve my skills in relevant areas. Beginning in August 2023, I joined the King's College London as visiting researcher to work on metabolomics, single cell RNA sequencing data and risk score prediction. The first publication to which I contributed has been published in the European Journal of Heart Failure: Serum metabolomics improves risk stratification for incident heart failure.
Initially, I studied Mechanical Engineering (Bachelor of Engineering) at the Cooperative State University Baden-Württemberg.
This unique combination of engineering and data science enables me to understand the technical background of the data I am working with and to break down complex tasks into smaller, more manageable ones. I am capable to quickly adapt to new work environments and apply my knowledge to produce reliable and fast results.

A full tabular cv of my eduation and certifications is found here:

To - From Occupation
now - 2023 Visiting Researcher at King's College London
School of Cardiovascular and Metabolic Medicine and Sciences
2023 XCS330 Multi-task and Meta-learning (Stanford Course)
Certificate
2023 - 2022 XCS234 Reinforcement learning (Stanford Course)
Certificate
2022 - 2017 Master of Science - Data Science (Albstadt Sigmaringen University)
Final grade: 1.2 (Best grade: 1.0)
2015 - 2012 Bachelor of Engineering - Mechanical Engineering (Cooperative State University Baden-Württemberg)
Final grade: 1.6 (Best grade: 1.0)

Utilising explainable AI algorithms to perform adversarial attacks on neural networks

Above you can see the original image (left) of a ship of the CALTECH256 dataset, a manipulated adversarial image of that ship (middle) and the change in classifier confidence for the correct class 8 (y8) and a wrong class 1 (y1), which would be a truck. After about 60 iterations, the classifier misclassifies the image for a truck instead of a ship.

On explainable AI

First, let's dive into explainable AI - this is a field that aims to provide insights into the decision making of artificial intelligence systems. Research and regulatory interest in explainable AI has risen steeply in recent years. One prominent example of regulators investigating this is the European Union - in GDPR article 5, transparency is mentioned as one of the key principles under which data has to be handled by companies. Such explanations can have different target groups:

  • System developers: Those who develop intelligent systems and want to perform debugging.
  • Regulatory bodies: Government bodies responsible with certification and oversight of the transparent and ethic use of AI.
  • End users: People who oftentimes are not aware that they are even subject to an AI-based product.

A developer may use the insights from explainable AI to judge whether the system can be released to the public. Regulatory bodies may require insights into the decision-making in order to certify products (e.g., Mercedes-Benz level 3 autonomous driving). Studies have shown that end users trust recommendations more when an explanation for the recommendation is given (Explaining Collaborative Filtering Recommendations).
The actual form an explanation takes can vary depending on the purpose of the explanation. Explanations can be:

  • Learned representations: Maximizing the activation of single neurons to obtain learned features / representations.
  • Individual predictions: Attribute to every input dimension its contribution to the output.
  • Model behaviour: Identify the general strategy of a model in achieving its task.
  • Representative examples: Identify the training data that impacts / informs the classification result most.

In my work, I focused on individual predictions, because my goal was to use the information to manipulate inputs to create adversarial inputs. To this end, I created a custom implementation of the Layer-wise Relevance Propagation (LRP) explainable AI algorithm for tensorflow models. LRP describes a mathematical framework under which explanations for individual decisions can be obtained. Generally speaking, relevance messages are back-propagated throughout a neural network to obtain the contribution of each input dimension to the output.

The network structure and weights that are present during inference (left) are used when calculating the back-propagation (right) of relevance messages.

There are three principles in LRP:

  • Relevance conservation: Throughout the back-propagation, the sum of all relevances stays constant.
  • Relevance guidance: The distribution of relevance during back-propagation is guided by the neuron connections and weights.
  • Specific propagation rules: To optimise the explanations, layer-specific LRP-rules can be used.

Ultimately, an explanation can be computed for a model, with a specific input and a specific target label. The resulting explanation can then be represented in the input dimensions, giving each input dimension a positive or negative relevance score for the target label.

The explanation for which pixel of the input image (left, from the MNIST dataset) seven contributes positively or negatively (right) to the classifier decision. Here, ca denotes the actual class, ct the target class of the explanation, a and b denote hyperparameters for the explanation.

On adversarial attacks

Studies showed that neural networks are vulnerable to perturbed inputs, so called adversarial examples, that are only slightly off the data distribution of training data. One famous example of such adversarial attacks is the Fast Gradient Sign Method (FGSM) , which has infamously shown that one small step accross all input dimensions in the direction of the gradient of the input with regard to (w.r.t) the output can fool image classification models.

Here, an image of the CALTECH256 dataset was manipulated using an iterative FGSM algorithm. The original image (left) and adversarial image (right) are barely distinguishable for humans. In the middle, the pixelwise manipulation is depicted using a heatmap, the actual change would not be recognizable.

Just like in xAI, adversarial attacks can take different forms. One way to distinguish attacks is to differ between white-box and black-box attacks. White-box attacks assume that the machine learning model as well as training data distribution is fully available. Thus, white-box attacks are generally easier and more effective. Black-box attacks on the other hand are performed with just limited access to the model - for example via an API. For my work, I focused on white-box attacks (such as FGSM).

Utilising xAI to perform adversarial attacks

Finally, as both building blocks - xAI and adversarial attacks - have been introduced, I want to share the connection of these topics. Intuitively, when attacking a classifier, we want to know which information is most valuable and obstruct / manipulate it. FGSM for example computes the gradient of the input w.r.t. the output. These gradients are also in some cases used for explanation purposes. Some xAI methods are even based on local perturbations (e.g., LIME) and use precisely the change in classification to determine the most important input dimensions.
The best-performing novel adversarial attack that I proposed in my thesis is the LRP-grad attack.

This algorithm takes the model parameters, input, the model, original and target labels as well as an interation count as input. Each loop, the most relevant pixels are identified and manipulated in direction of the gradient of the input w.r.t the output.

This attack can be also be seen in action in the following gif:

The left side depicts the relevance matrix per iteration, the right side shows the adversarial image over all iterations. After about 60 iterations, the classifier in this case was fooled.

Ultimately, I was able to show that the proposed LRP-grad attack outperforms iterative FGSM due to reaching a higher attack success rate (ASR) at lower Lp-norms. The following graph shows this comparison:

Here, a comparison of the L1, L2 and Linf norms over the ASR is performed. Four attack methods are compared - IFGSM, LRP-flip, LRP-mean and LRP-grad.

Of course, this overview only touches the surface of my Master's thesis - so feel free to contact me if you have any inputs, ideas or other thoughts. To sum it up, my Master's thesis lead to the following contributions:

  • Custom implementation of layerwise relevance propagation and adversarial attacks.
  • Proposal of two novel adversarial attack methods.
  • Quantitative benchmark of IFGSM, LRP-flip, LRP-grad and LRP-mean.

Contact

I am happy for any input that you might have, so feel free to contact me.