{"payload":{"header_redesign_enabled":false,"results":[{"id":"493887055","archived":false,"color":"#DA5B0B","followers":231,"has_funding_file":false,"hl_name":"implus/UM-MAE","hl_trunc_description":"Official Codes for \"Uniform Masking: Enabling MAE Pre-training for Pyramid-based Vision Transformers with Locality\"","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":493887055,"name":"UM-MAE","owner_id":4126718,"owner_login":"implus","updated_at":"2022-12-03T03:45:49.235Z","has_issues":true}},"sponsorable":false,"topics":["coco","mae","ade20k","self-supervised-learning","masked-autoencoder","imagenet-classification","pyramid-vision-transformer","swin-transformer","masked-image-modeling","hierarchical-vision-transformer"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":78,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aimplus%252FUM-MAE%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/implus/UM-MAE/star":{"post":"YdHvgw-CxC9cqn7I9t3DvYrAo_7Jopg-qZBadZ5dMJqxLU0-73mXoR4TaWFfBoHEMMvW48CyjjWmnN7hFd1nsw"},"/implus/UM-MAE/unstar":{"post":"edo0jNCCQX55LXXJGCIHH27qUCAv4EirNeyU4-8MLpX7F3Co2QgXRRy6ExaR92HKeMGQ5YgwJAmdTSKiOOLM3Q"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"QeCfxIIbOmmmk34HMZaxZ2Rgjzo7FaJooAl223tPR0BbZx89IXutaJupmmyFa_T956ctr5TjFd-UEXNz6B7Bng"}}},"title":"Repository search results"}