{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"DecodingTrust","owner":"AI-secure","isFork":false,"description":"A Comprehensive Assessment of Trustworthiness in GPT Models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":9,"starsCount":220,"forksCount":48,"license":"Creative Commons Attribution Share Alike 4.0 International","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-08T19:15:25.998Z"}},{"type":"Public","name":"VFL-ADMM","owner":"AI-secure","isFork":false,"description":"Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM (SaTML 2024)","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T21:37:35.870Z"}},{"type":"Public","name":"aug-pe","owner":"AI-secure","isFork":false,"description":"[ICML 2024] Differentially Private Synthetic Data via Foundation Model APIs 2: Text","allTopics":["language-model","differential-privacy","ai-privacy","large-language-models","prompt-engineering"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":16,"forksCount":2,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,5,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-14T20:34:20.757Z"}},{"type":"Public","name":"DPFL-Robustness","owner":"AI-secure","isFork":false,"description":"[CCS 2023] Unraveling the Connections between Privacy and Certified Robustness in Federated Learning Against Poisoning Attacks ","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-15T06:49:21.030Z"}},{"type":"Public","name":"hf-blog","owner":"AI-secure","isFork":true,"description":"Public repo for HF blog posts","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":641,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-26T19:11:47.752Z"}},{"type":"Public","name":"helm","owner":"AI-secure","isFork":true,"description":"Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":230,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-07T02:09:34.017Z"}},{"type":"Public","name":"DecodingTrust-Data-Legacy","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[3,9,0,0,0,1,0,0,76,2,0,1,5,1,1,2,1,2,4,6,3,0,2,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-25T22:21:19.521Z"}},{"type":"Public","name":"TextGuard","owner":"AI-secure","isFork":false,"description":"TextGuard: Provable Defense against Backdoor Attacks on Text Classification","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-07T07:17:16.937Z"}},{"type":"Public","name":"InfoBERT","owner":"AI-secure","isFork":false,"description":"[ICLR 2021] \"InfoBERT: Improving Robustness of Language Models from An Information Theoretic Perspective\" by Boxin Wang, Shuohang Wang, Yu Cheng, Zhe Gan, Ruoxi Jia, Bo Li, Jingjing Liu","allTopics":["information-theory","language-models","bert","adversarial-attacks","roberta","adversarial-defense","adversarial-robustness"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":82,"forksCount":7,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-25T18:55:13.693Z"}},{"type":"Public","name":"FedGame","owner":"AI-secure","isFork":false,"description":"Official implementation for paper \"FedGame: A Game-Theoretic Defense against Backdoor Attacks in Federated Learning\" (NeurIPS 2023).","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-12T00:02:28.701Z"}},{"type":"Public","name":"Robustness-Against-Backdoor-Attacks","owner":"AI-secure","isFork":false,"description":"RAB: Provable Robustness Against Backdoor Attacks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":2,"starsCount":39,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T21:46:41.474Z"}},{"type":"Public","name":"semantic-randomized-smoothing","owner":"AI-secure","isFork":false,"description":"[CCS 2021] TSS: Transformation-specific smoothing for robustness certification","allTopics":["security","robustness-verification","deep-learning"],"primaryLanguage":{"name":"Roff","color":"#ecdebe"},"pullRequestCount":5,"issueCount":0,"starsCount":20,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T21:45:55.203Z"}},{"type":"Public","name":"FLBenchmark-toolkit","owner":"AI-secure","isFork":false,"description":"Federated Learning Framework Benchmark (UniFed)","allTopics":["benchmark","federated-learning","machine-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":46,"forksCount":5,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-14T05:32:37.733Z"}},{"type":"Public","name":"SecretGen","owner":"AI-secure","isFork":false,"description":"A general model inversion attack against large pre-trained models.","allTopics":["machine-learning","privacy"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-22T20:37:17.413Z"}},{"type":"Public","name":"adversarial-glue","owner":"AI-secure","isFork":false,"description":"[NeurIPS 2021] \"Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models\" by Boxin Wang*, Chejian Xu*, Shuohang Wang, Zhe Gan, Yu Cheng, Jianfeng Gao, Ahmed Hassan Awadallah, Bo Li.","allTopics":["nlp","machine-learning","nlp-library","adversarial-examples","adversarial-attacks"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-03T08:06:14.880Z"}},{"type":"Public","name":"VeriGauge","owner":"AI-secure","isFork":false,"description":"A united toolbox for running major robustness verification approaches for DNNs. [S&P 2023]","allTopics":["robustness","verfication","deep-learning"],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":4,"issueCount":3,"starsCount":87,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-24T22:24:33.933Z"}},{"type":"Public","name":"Certified-Fairness","owner":"AI-secure","isFork":false,"description":"Code for Certifying Some Distributional Fairness with Subpopulation Decomposition [NeurIPS 2022]","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-03T06:12:19.470Z"}},{"type":"Public","name":"CoPur","owner":"AI-secure","isFork":false,"description":"CoPur: Certifiably Robust Collaborative Inference via Feature Purification (NeurIPS 2022)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-07T04:37:17.676Z"}},{"type":"Public","name":"transferability-versus-robustness","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-06T17:08:23.768Z"}},{"type":"Public","name":"DMLW2022","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-03T21:24:40.807Z"}},{"type":"Public","name":"Certified-Robustness-SoK-Oldver","owner":"AI-secure","isFork":false,"description":"This repo keeps track of popular provable training and verification approaches towards robust neural networks, including leaderboards on popular datasets and paper categorization.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":99,"forksCount":10,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-18T21:04:43.723Z"}},{"type":"Public","name":"Layerwise-Orthogonal-Training","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-11T20:30:47.862Z"}},{"type":"Public","name":"CROP","owner":"AI-secure","isFork":false,"description":"[ICLR 2022] CROP: Certifying Robust Policies for Reinforcement Learning through Functional Smoothing","allTopics":["reinforcement-learning","certification","robustness"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":8,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-16T00:09:44.508Z"}},{"type":"Public","name":"SemAttack","owner":"AI-secure","isFork":false,"description":"[NAACL 2022] \"SemAttack: Natural Textual Attacks via Different Semantic Spaces\" by Boxin Wang, Chejian Xu, Xiangyu Liu, Yu Cheng, Bo Li","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-11T10:17:59.532Z"}},{"type":"Public","name":"Transferability-Reduced-Smooth-Ensemble","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":22,"forksCount":8,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-15T18:01:56.293Z"}},{"type":"Public","name":"COPA","owner":"AI-secure","isFork":false,"description":"[ICLR 2022] COPA: Certifying Robust Policies for Offline Reinforcement Learning against Poisoning Attacks","allTopics":["certification","robustness","reinforcement-learning"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-05T23:31:03.919Z"}},{"type":"Public","name":"COPA_Highway","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-19T06:44:00.620Z"}},{"type":"Public","name":"COPA_Atari","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-18T00:39:46.554Z"}},{"type":"Public","name":"DataLens","owner":"AI-secure","isFork":false,"description":"[CCS 2021] \"DataLens: Scalable Privacy Preserving Training via Gradient Compression and Aggregation\" by Boxin Wang*, Fan Wu*, Yunhui Long*, Luka Rimanic, Ce Zhang, Bo Li","allTopics":["generative-model","differential-privacy","gradient-compression"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":37,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-28T20:00:32.484Z"}},{"type":"Public","name":"G-PATE","owner":"AI-secure","isFork":false,"description":"[NeurIPS 2021] \"G-PATE: Scalable Differentially Private Data Generator via Private Aggregation of Teacher Discriminators\" by Yunhui Long*, Boxin Wang*, Zhuolin Yang, Bhavya Kailkhura, Aston Zhang, Carl A. Gunter, Bo Li","allTopics":["generative-model","differential-privacy"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":30,"forksCount":9,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-26T03:23:18.322Z"}}],"repositoryCount":51,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}