{"payload":{"header_redesign_enabled":false,"results":[{"id":"443935262","archived":false,"color":"#DA5B0B","followers":29,"has_funding_file":false,"hl_name":"Heewon-Hailey/multi-armed-bandits-for-recommendation-systems","hl_trunc_description":" implement basic and contextual MAB algorithms for recommendation system","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":443935262,"name":"multi-armed-bandits-for-recommendation-systems","owner_id":58554475,"owner_login":"Heewon-Hailey","updated_at":"2022-01-18T09:20:12.682Z","has_issues":true}},"sponsorable":false,"topics":["python","numpy","scikit-learn","epsilon-greedy","recommendation-system","matplotlib","upper-confidence-bounds","contextual-bandits","multiarmed-bandits"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":85,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AHeewon-Hailey%252Fmulti-armed-bandits-for-recommendation-systems%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/Heewon-Hailey/multi-armed-bandits-for-recommendation-systems/star":{"post":"JPr_tQTtfHcG9NBuUdhlTjsmzrf6_ALZ2h7VupRhUDC2UtboXOeHAsKbp1Mlm-6zEFMml1_IqNZjD9NuBAEibw"},"/Heewon-Hailey/multi-armed-bandits-for-recommendation-systems/unstar":{"post":"C1Wr6CIxzKhwneazA4vgOfmK3UhY1GjYn-a1yE7dvDKYr3j82JpUqBzdZUagUbw2VRyrvd8sBL0Z9Jzwqblzbg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"1DvLIJm3euKl7_mI84zCrKv5iGknDLdbOYTDNeQwAC6JN0VP4BUkfS6AnU5wRGdo15OVdzGKZS9SyHkdvItMyg"}}},"title":"Repository search results"}