{"payload":{"header_redesign_enabled":false,"results":[{"id":"148729228","archived":false,"color":"#3572A5","followers":16,"has_funding_file":false,"hl_name":"jingyuanz/keras-self-attention-layer","hl_trunc_description":"a simple implementation of self attention layer that outputs flattened sentence embedding matrix, with the Frobenius norm penalty","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":148729228,"name":"keras-self-attention-layer","owner_id":7495778,"owner_login":"jingyuanz","updated_at":"2018-09-14T03:15:41.862Z","has_issues":true}},"sponsorable":false,"topics":[],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":79,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Ajingyuanz%252Fkeras-self-attention-layer%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/jingyuanz/keras-self-attention-layer/star":{"post":"CV4LDNN4tGdCEVJUCWtE1l7G64--eq4mb09DumqBdebJz31P-Bb3mM12iXnk2n7nZUrSmPSieY7RAEOFoQmQNw"},"/jingyuanz/keras-self-attention-layer/unstar":{"post":"vWunYJ88AGrS99DEciAri73FYafbhwJoJuQnaTD8NcjZu8LHN0mye2ESJpRGm0sjhlk-pUkaiNHCOIZisfaefA"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"KntWcp2CMIZ5b1daIDdpV57Vzx53C2fDHj_1NiSeJOlUA_M_mAVM_BzyUoBDopAV1DnXNr5o9Y7vdveWil80xA"}}},"title":"Repository search results"}