Logo
Pattern

Discover published sets by community

Explore tens of thousands of sets crafted by our community.

Machine Learning Security and Privacy

10

Flashcards

0/10

Still learning
StarStarStarStar

Fairness and Bias

StarStarStarStar

Fairness and bias issues arise when a model's decisions disproportionately affect certain groups. Mitigation includes auditing datasets for bias, using fairness metrics during model evaluation, and applying fairness-aware algorithms.

StarStarStarStar

Replay Attacks

StarStarStarStar

Replay attacks use previously captured data to fool a system. Preventing replay attacks involves using anomaly detection, timestamping inputs, and ensuring contextual relevance of the data.

StarStarStarStar

Data Poisoning

StarStarStarStar

Data poisoning involves injecting malicious data into a dataset, leading to compromised model performance. Mitigation strategies include robust data validation, anomaly detection, and regular model retraining with trusted data sources.

StarStarStarStar

Information Leakage

StarStarStarStar

Information leakage occurs when a model unintentionally exposes sensitive information. This can be mitigated by reducing model complexity, applying data anonymization, and ensuring proper access controls.

StarStarStarStar

Membership Inference Attacks

StarStarStarStar

Membership inference attacks determine if a specific data point was used in the model's training set. To prevent this, use techniques like data generalization, noise addition, and regularizing models to reduce overfitting.

StarStarStarStar

Evasion Attacks (Adversarial Examples)

StarStarStarStar

Evasion attacks, or adversarial examples, involve subtly altering inputs to mislead models. Defenses include training on adversarial examples, using model ensembles, and deploying adversarial detection systems.

StarStarStarStar

Backdoor Attacks

StarStarStarStar

Backdoor attacks embed hidden behavior in a model during training, which can be activated by trigger inputs. Countermeasures include thorough validation of training data, model inspection, and anomaly detection to identify triggers.

StarStarStarStar

Model Stealing (Extraction)

StarStarStarStar

Model stealing is the unauthorized replication of a model by probing the system with inputs and observing outputs. Mitigation includes API rate limiting, monitoring for suspicious activity, and employing watermarking techniques.

StarStarStarStar

Model Inversion Attacks

StarStarStarStar

Model inversion attacks aim to recover sensitive information from a model's output. Protecting against this involves limiting model access, applying differential privacy techniques, and reducing model overfitting.

StarStarStarStar

Trojan Attacks

StarStarStarStar

Trojan attacks involve altering a model to respond to certain inputs with a misclassification. Defense strategies include inspecting and cleaning the training dataset, implementing strong model validation, and avoiding use of untrusted pre-trained models.

Know
0
Still learning
Click to flip
Know
0
Logo

© Hypatia.Tech. 2024 All rights reserved.