{"payload":{"header_redesign_enabled":false,"results":[{"id":"198125003","archived":false,"color":"#3572A5","followers":23,"has_funding_file":false,"hl_name":"catalina17/VideoNavQA","hl_trunc_description":"An alternative EQA paradigm and informative benchmark + models (BMVC 2019, ViGIL 2019 spotlight)","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":198125003,"name":"VideoNavQA","owner_id":11073663,"owner_login":"catalina17","updated_at":"2022-06-22T00:01:41.176Z","has_issues":true}},"sponsorable":false,"topics":["benchmark","machine-learning","deep-neural-networks","video","navigation","vqa","question-answering","visual-reasoning","multimodal","embodied","cross-modality","conditioning","videonavqa"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":137,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Acatalina17%252FVideoNavQA%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/catalina17/VideoNavQA/star":{"post":"xanumg5I9wSE4OqXGNYuch7ssSu6P6mi8DbuNPqOx5ft_kfbEfy7sFg3FAzFgJ5YJU_MnHzjPUdO6PDj9EI-vA"},"/catalina17/VideoNavQA/unstar":{"post":"-1YSsaJg7H-NoRo28Dn5y5nVKiASvOQUesSpwXHXIBvg7QGgQAhffg9zwaiX1_HF7-4I0WlqQ-AKIEbjjWP0Iw"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"6CkwD8e7rnha3A_fQ8eahODJfIdg_YonoZxPcTBNetsyMtC0dhefX-HgHVWccHGBP11XFTRyLLGgvz5M6Vupfg"}}},"title":"Repository search results"}