{"payload":{"header_redesign_enabled":false,"results":[{"id":"451691984","archived":false,"color":"#DA5B0B","followers":4371,"has_funding_file":false,"hl_name":"salesforce/BLIP","hl_trunc_description":"PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation ","language":"Jupyter Notebook","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":451691984,"name":"BLIP","owner_id":453694,"owner_login":"salesforce","updated_at":"2024-05-20T21:21:02.703Z","has_issues":true}},"sponsorable":false,"topics":["image-captioning","visual-reasoning","visual-question-answering","vision-language","vision-language-transformer","image-text-retrieval","vision-and-language-pre-training"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":61,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Asalesforce%252FBLIP%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/salesforce/BLIP/star":{"post":"U4JJaop2tOgwAga6pr2JUg9q3CWTQAN3cYJ7pzcBJq-gzhRnR_aOCxMTdkA-FQ7o3YF0GLDZpXUoHQgVe4G7JQ"},"/salesforce/BLIP/unstar":{"post":"w49VHT1YuQya42vH1oW5zgZq1jVE67LWVuOT6DrMnQnV84Ec8OjH2b6JYvjJP06sK2KliI6tNOOl_kYpVjnyRw"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"1X5ppBknow4XVocLux1hkDSUMTP5wWo02u7DL6SGAChtEri1kBPCgBmjnAbdLnRagzFG-2AUKSkeZjaVlD-Ovw"}}},"title":"Repository search results"}