AI Secure (AI-secure)

AI Secure

AI-secure

Geek Repo

UIUC Secure Learning Lab

Location:University of Illinois at Urbana-Champaign

Home Page:https://aisecure.github.io/

Github PK Tool:Github PK Tool

AI Secure's repositories

DecodingTrust

A Comprehensive Assessment of Trustworthiness in GPT Models

Language:PythonLicense:CC-BY-SA-4.0Stargazers:207Issues:6Issues:18

Certified-Robustness-SoK-Oldver

This repo keeps track of popular provable training and verification approaches towards robust neural networks, including leaderboards on popular datasets and paper categorization.

VeriGauge

A united toolbox for running major robustness verification approaches for DNNs. [S&P 2023]

InfoBERT

[ICLR 2021] "InfoBERT: Improving Robustness of Language Models from An Information Theoretic Perspective" by Boxin Wang, Shuohang Wang, Yu Cheng, Zhe Gan, Ruoxi Jia, Bo Li, Jingjing Liu

FLBenchmark-toolkit

Federated Learning Framework Benchmark (UniFed)

Language:PythonLicense:Apache-2.0Stargazers:45Issues:3Issues:5

Robustness-Against-Backdoor-Attacks

RAB: Provable Robustness Against Backdoor Attacks

DataLens

[CCS 2021] "DataLens: Scalable Privacy Preserving Training via Gradient Compression and Aggregation" by Boxin Wang*, Fan Wu*, Yunhui Long*, Luka Rimanic, Ce Zhang, Bo Li

G-PATE

[NeurIPS 2021] "G-PATE: Scalable Differentially Private Data Generator via Private Aggregation of Teacher Discriminators" by Yunhui Long*, Boxin Wang*, Zhuolin Yang, Bhavya Kailkhura, Aston Zhang, Carl A. Gunter, Bo Li

semantic-randomized-smoothing

[CCS 2021] TSS: Transformation-specific smoothing for robustness certification

SemAttack

[NAACL 2022] "SemAttack: Natural Textual Attacks via Different Semantic Spaces" by Boxin Wang, Chejian Xu, Xiangyu Liu, Yu Cheng, Bo Li

aug-pe

Differentially Private Synthetic Data via Foundation Model APIs 2: Text

Language:PythonLicense:Apache-2.0Stargazers:10Issues:0Issues:0

adversarial-glue

[NeurIPS 2021] "Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models" by Boxin Wang*, Chejian Xu*, Shuohang Wang, Zhe Gan, Yu Cheng, Jianfeng Gao, Ahmed Hassan Awadallah, Bo Li.

Language:PythonStargazers:8Issues:1Issues:0

COPA

[ICLR 2022] COPA: Certifying Robust Policies for Offline Reinforcement Learning against Poisoning Attacks

CROP

[ICLR 2022] CROP: Certifying Robust Policies for Reinforcement Learning through Functional Smoothing

CoPur

CoPur: Certifiably Robust Collaborative Inference via Feature Purification (NeurIPS 2022)

Language:PythonStargazers:6Issues:1Issues:0

DPFL-Robustness

[CCS 2023] Unraveling the Connections between Privacy and Certified Robustness in Federated Learning Against Poisoning Attacks

Language:PythonStargazers:6Issues:3Issues:0

Certified-Fairness

Code for Certifying Some Distributional Fairness with Subpopulation Decomposition [NeurIPS 2022]

Language:PythonStargazers:4Issues:3Issues:0

SecretGen

A general model inversion attack against large pre-trained models.

Language:PythonLicense:MITStargazers:4Issues:1Issues:1

TextGuard

TextGuard: Provable Defense against Backdoor Attacks on Text Classification

Language:PythonStargazers:4Issues:3Issues:0

FedGame

Official implementation for paper "FedGame: A Game-Theoretic Defense against Backdoor Attacks in Federated Learning" (NeurIPS 2023).

Language:PythonStargazers:1Issues:2Issues:0
Language:PythonStargazers:0Issues:3Issues:0
Language:Jupyter NotebookStargazers:0Issues:2Issues:0
Language:PythonStargazers:0Issues:0Issues:0
Language:HTMLStargazers:0Issues:1Issues:0

helm

Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

hf-blog

Public repo for HF blog posts

Language:Jupyter NotebookStargazers:0Issues:0Issues:0

VFL-ADMM

Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM (SaTML 2024)

License:Apache-2.0Stargazers:0Issues:2Issues:0