{"payload":{"header_redesign_enabled":false,"results":[{"id":"278810244","archived":false,"color":"#3572A5","followers":3958,"has_funding_file":false,"hl_name":"open-mmlab/mmaction2","hl_trunc_description":"OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":278810244,"name":"mmaction2","owner_id":10245193,"owner_login":"open-mmlab","updated_at":"2024-05-16T12:11:57.704Z","has_issues":true}},"sponsorable":false,"topics":["benchmark","deep-learning","pytorch","ava","x3d","action-recognition","video-understanding","video-classification","tsm","non-local","i3d","tsn","slowfast","temporal-action-localization","spatial-temporal-action-detection","openmmlab","posec3d","uniformerv2"],"type":"Public","help_wanted_issues_count":5,"good_first_issue_issues_count":2,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":68,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aopen-mmlab%252Fmmaction2%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/open-mmlab/mmaction2/star":{"post":"r65A3E8CFZ5bpRjcpSc_yoCLp2Z-NtE6kxdKJ8T3V70Xm_8DZlJxEOrTbuYEKiuUzOnQKx-ZeJjrhx3DaipkAw"},"/open-mmlab/mmaction2/unstar":{"post":"D_i1j_k2-C3FzjdRjRjeARf6JZzU5CuSA6xguXMjlDbTZpQWOjrYHhpRWIFKfCPZ7bLVbsTo4SlLmitXl4PJPw"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"5KheCV3MRBzC1S3-uR2stIl6JakPJu2fUzNU_SyPeDNvtQSiDNc_Op5mMTFLmVxs0_IPn1m5k3L8fgOY8gbRpg"}}},"title":"Repository search results"}