Skip to content
View Muzammal-Naseer's full-sized avatar
🐻
🐻
Block or Report

Block or report Muzammal-Naseer

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Muzammal-Naseer/README.md

Hi there πŸ‘‹

  • πŸ”­ My research interests are robust visual perception by understanding and explaining AI behavior through adversarial machine learning, temporal perception, representation learning (self-supervision, self-distillation, self-critique), and configuring the role of language models (LLMs) in building visual AI systems.
  • 🌱 You are welcome to explore my research work along with the provided code below. Seven of the papers are accepted as Oral/Spotlight at ICLR, NeurIPS, AAAI, CVPR, BMVC, and ACCV.
  • πŸ“« How to reach me: muz.pak@gmail.com
  • ⚑ Fun fact: I am really into fitness and thinking of joining the GYM for quite some time now πŸ˜„

🌱 Repositories

Topic Application Paper Repo Venue
Vision-Language Learning Composed Video Retrieval Composed Video Retrieval via Enriched Context and Discriminative Embeddings composed-video-retrieval CVPR'24
Self-supervision Multi-Spectral Satellite Imagery Rethinking Transformers Pre-training for Multi-Spectral Satellite Imagery satmae_pp CVPR'24
Vision-Language Learning Video grounding Video-GroundingDINO: Towards Open-Vocabulary Spatio-Temporal Video Grounding Video-GroundingDINO CVPR'24
Vision-Language Learning Language Driven VLM for Remote Sensing Geochat: Grounded large vision-language model for remote sensing GeoChat CVPR'24
Vision-Language Learning Leaverging LLM to generate complex scenes (Zero-Shot) LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts llmblueprint ICLR'24
Self-supervision Self-structural Alignment of Foundational Models (Zero-Shot) Towards Realistic Zero-Shot Classification via Self Structural Semantic Alignment S3A AAAI'24-Oral
Vision-Language Learning Test-Time Alignment of Foundational Models (Zero-Shot) Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization PromptAlign NeurIPS'23
Vision-Language Learning Regulating Foundational Models Self-regulating Prompts: Foundational Model Adaptation without Forgetting PromptSRC ICCV'23
Network Engineering Video Recognition Video-FocalNets: Spatio-Temporal Focal Modulation for Video Action Recognition Video-FocalNets ICCV'23
Vision-Language Learning Face Anti-spoofing FLIP: Cross-domain Face Anti-spoofing with Language Guidance FLIP ICCV'23
3D Medical Segmentation Adversarial Training Frequency Domain Adversarial Training for Robust Volumetric Medical Segmentation VAFA MICCAI'23
Vision-Language Learning Facial Privacy CLIP2Protect: Protecting Facial Privacy Using Text-Guided Makeup via Adversarial Latent Search Clip2Protect CVPR'23
Vision-Language Learning Video Recognition (Zero-shot) Vita-CLIP: Video and text adaptive CLIP via Multimodal Prompting Vita-CLIP CVPR'23
Prompt learning Image Recognition (Category Discovery) PromptCAL for Generalized Novel Category Discovery PromptCAL CVPR'23
Prompt learning Adversarial Attack Boosting Adversarial Transferability using Dynamic Cues DCViT-AT ICLR'23
Self-supervision Video Recognition Self-Supervised Video Transformer SVT CVPR'22-Oral
Contrastive learning Adversarial Defense Stylized Adversarial Training SAT IEEE-TPAMI'22
Self-supervision Adversarial Attack Adversarial Pixel Restoration as a Pretext Task for Transferable Perturbations ARP BMVC'22-Oral
Self-supervision Image Recognition How to Train Vision Transformer on Small-scale Datasets? VSSD BMVC'22
Self-distillation Image Recognition (Domain Generalization) Self-Distilled Vision Transformer for Domain Generalization SDViT ACCV'22-Oral
Attention Analysis Understanding Vision Transformer Intriguing Properties of Vision Transformers IPViT NeurIPS'21-Spotlight
Self-ensemble Adversarial Attack On Improving Adversarial Transferability of Vision Transformers ATViT ICLR'21-Spotlight
Distribution matching Adversarial Attack On Generating Transferable Targeted Perturbations TTP ICCV'21
Contrastive learning Image Recognition Orthogonal Projection Loss OPL ICCV'21
Self-supervision Adversarial Defense A Self-supervised Approach for Adversarial Robustness NRP CVPR'20-Oral
Relativistic optimization Adversarial Attack Cross-Domain Transferability of Adversarial Perturbations CDA NeurIPS'19
Gradient Smoothing Adversarial Defense Local Gradients Smoothing: Defense Against Localized Adversarial Attacks LGS WACV'19

Pinned

  1. hananshafi/vits-for-small-scale-datasets hananshafi/vits-for-small-scale-datasets Public

    [BMVC 2022] Official repository for "How to Train Vision Transformer on Small-scale Datasets?"

    Python 130 12

  2. IPViT IPViT Public

    Official repository for "Intriguing Properties of Vision Transformers" (NeurIPS 2021--Spotlight)

    Python 173 19

  3. NRP NRP Public

    Official repository for "A Self-supervised Approach for Adversarial Robustness" (CVPR 2020--Oral)

    Python 89 17

  4. ATViT ATViT Public

    Official repository for "On Improving Adversarial Transferability of Vision Transformers" (ICLR 2022--Spotlight)

    Python 67 11

  5. TTP TTP Public

    Official repository for "On Generating Transferable Targeted Perturbations" (ICCV 2021)

    Python 61 9

  6. CDA CDA Public

    Official repository for "Cross-Domain Transferability of Adversarial Perturbations" (NeurIPS 2019)

    Python 57 11