{"payload":{"header_redesign_enabled":false,"results":[{"id":"205030175","archived":false,"color":"#3572A5","followers":46,"has_funding_file":false,"hl_name":"bohanli/vae-pretraining-encoder","hl_trunc_description":"PyTorch implementation of A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text (EMNLP 2019)","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":205030175,"name":"vae-pretraining-encoder","owner_id":11835112,"owner_login":"bohanli","updated_at":"2020-02-16T23:22:52.518Z","has_issues":true}},"sponsorable":false,"topics":[],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":61,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Abohanli%252Fvae-pretraining-encoder%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/bohanli/vae-pretraining-encoder/star":{"post":"HtkbC5Dz6iClMr7hSCc1aC0nMv--j_tw354OTkxvkXQzgRRI0fIHhpYqSA11dHHX-b8VQmJywHxcrkqDCIDCyw"},"/bohanli/vae-pretraining-encoder/unstar":{"post":"QnPcS4tIahv8f9Dm-zyLVvwN8XsDvGGe5u8vFqubx4jl-g997mc-dO6q33MYKAE2qcivA5LgZfIE4-BJvsmOCQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"h5lG6tR8cN0FQ_uMOk5_-oGRcBNVpXU_BCide6AOQnlUnlrmfpNvFeq-Auz_IdhHXKNykjCg3LtWclaj0K9uZw"}}},"title":"Repository search results"}