Deadline for manuscript submissions: 31 October 2018
Machine learning is clearly a research area that will continue creating real-world impact, as computing power becomes increasingly more readily available. Security and privacy considerations, however, are vital, in particular since machine learning algorithms are often perceived as magical black boxes, in which the inner workings are not easily made transparent. Important topics that warrant new research are, among others:
The right to be forgotten. How much of the ‘original’ personal data is embedded in trained neural networks? Can we delete this data without retraining? How can we measure the anonymity/pseudonymity of training data embedded in a trained network?
How easy is it to attack training sets and trained networks? If ML is used for real-world applications such as autonomous driving, successful attacks may have huge impact.
We look forward to receiving research papers that address, not only the aforementioned examples, but also any excellent research that investigates privacy and security aspects in ML in depth.
Edgar Weippl, Francesco Buccafurri, Guest Editors
- Threat models
- Attacks against machine learning
- Malware detection
- Black-box attacks against machine learning
- Adversarial training and defensive distillation
- Privacy-preserving machine learning
- Application of machine learning to security and privacy