{"payload":{"header_redesign_enabled":false,"results":[{"id":"276117807","archived":false,"color":"#DA5B0B","followers":26,"has_funding_file":false,"hl_name":"BY571/Normalized-Advantage-Function-NAF-","hl_trunc_description":"PyTorch implementation of the Q-Learning Algorithm Normalized Advantage Function for continuous control problems + PER and N-step Method","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":276117807,"name":"Normalized-Advantage-Function-NAF-","owner_id":29492081,"owner_login":"BY571","updated_at":"2021-02-16T16:48:30.085Z","has_issues":true}},"sponsorable":false,"topics":["reinforcement-learning","q-learning","dqn","reinforcement-learning-algorithms","continuous-control","naf","ddpg-algorithm","prioritized-experience-replay","normalized-advantage-functions","q-learning-algorithm","n-step-bootstrapping"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":84,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253ABY571%252FNormalized-Advantage-Function-NAF-%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/BY571/Normalized-Advantage-Function-NAF-/star":{"post":"AhOmYpX4ZVZH7bEhEVafbMewPrSNHt5QMBs_th8heD9_fO9lRh6U6o7SAGbl5GT3Ih2PqB3ptARWIhMKqhRWtQ"},"/BY571/Normalized-Advantage-Function-NAF-/unstar":{"post":"Z6v9gCQ3OqyPhTQfInqQHr7YmiJUB5QfRUAtf1CdLY7S9JScjgBwU-K6NmvDG3XmIVfIeuuwPe-ns1WO4L8IDg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"N8bq8VzrC-wr75fhFD7pgfLk7axuWVRoUoFLnlRVp0N3uoGs5WMOWOMXMMONwzvcGA6D60Kr_lytTLzd7J8c2Q"}}},"title":"Repository search results"}