{"payload":{"header_redesign_enabled":false,"results":[{"id":"280992148","archived":false,"color":"#3572A5","followers":4,"has_funding_file":false,"hl_name":"itsShnik/adaptively-finetuning-transformers","hl_trunc_description":"Adaptively fine tuning transformer based models for multiple domains and multiple tasks","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":280992148,"name":"adaptively-finetuning-transformers","owner_id":43925694,"owner_login":"itsShnik","updated_at":"2023-05-22T22:51:17.203Z","has_issues":true}},"sponsorable":false,"topics":["transformers","pytorch","visual-question-answering","finetuning","vision-and-language","vlbert","lxmert","vqav2","vqacpv2","spottune","blockdrop"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":70,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AitsShnik%252Fadaptively-finetuning-transformers%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/itsShnik/adaptively-finetuning-transformers/star":{"post":"vtzolyERvPidNAMunjE1eglcbQvwk_oteNtw4IPGEysnqLxASdtjZLS6E8GIjCnfBgfChP04xB1704u3_VX3SQ"},"/itsShnik/adaptively-finetuning-transformers/unstar":{"post":"o9pYQBbxa-HN9OO4ZxP0gNPDczd2W9t89KbtPdJmynkUlkgpobsVENKSA7QX-kaebJmJIB2Fbq_F7E-bkr37Ww"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"JQl8iPOt9CemcnAxrs1CYHMrt5bBd1Pdqar_Eu4VLMi8pkFOnhvaP-iy1HhR5phI20ea0L9p73bv5Dz5aeHdWg"}}},"title":"Repository search results"}