How would you design a data poisoning attack on the Quick, Draw! dataset by inserting invisible or redundant vector strokes that a human would not detect, but that would systematically induce the model to confuse one class with another?
Saturday, 01 November 2025
by JOSE ALFONSIN PENA
Designing a data poisoning attack on the Quick, Draw! dataset, specifically by inserting invisible or redundant vector strokes, requires a multifaceted understanding of how vector-based sketch data is represented, how convolutional and recurrent neural networks process such data, and how imperceptible modifications can manipulate a model’s decision boundaries without alerting human annotators or users. Understanding

