Dataless Black-Box Model Comparison


Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription Access

Abstract

In a time where the training of new machine learning models is extremely time-consuming and resource-intensive and the sale of these models or the access to them is more popular than ever, it is important to think about ways to ensure the protection of these models against theft. In this paper, we present a method for estimating the similarity or distance between two black-box models. Our approach does not depend on the knowledge about specific training data and therefore may be used to identify copies of or stolen machine learning models. It can also be applied to detect instances of license violations regarding the use of datasets. We validate our proposed method empirically on the CIFAR-10 and MNIST datasets using convolutional neural networks, generative adversarial networks and support vector machines. We show that it can clearly distinguish between models trained on different datasets. Theoretical foundations of our work are also given.

About the authors

C. Theiss

Computer Vision Group

Author for correspondence.
Email: christoph.theiss@uni-jena.de
Germany, Jena

C. A. Brust

Computer Vision Group

Email: christoph.theiss@uni-jena.de
Germany, Jena

J. Denzler

Computer Vision Group

Email: christoph.theiss@uni-jena.de
Germany, Jena

Supplementary files

Supplementary Files
Action
1. JATS XML

Copyright (c) 2018 Pleiades Publishing, Ltd.