{"payload":{"header_redesign_enabled":false,"results":[{"id":"253995876","archived":false,"color":"#DA5B0B","followers":53,"has_funding_file":false,"hl_name":"kentaroy47/benchmark-FP32-FP16-INT8-with-TensorRT","hl_trunc_description":"Benchmark inference speed of CNNs with various quantization methods in Pytorch+TensorRT with Jetson Nano/Xavier","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":253995876,"name":"benchmark-FP32-FP16-INT8-with-TensorRT","owner_id":39611941,"owner_login":"kentaroy47","updated_at":"2023-05-31T20:13:44.282Z","has_issues":true}},"sponsorable":false,"topics":["hardware","tensorrt","jetson-xavier","jetson-nano","image-inference"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":1,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":92,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Akentaroy47%252Fbenchmark-FP32-FP16-INT8-with-TensorRT%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/kentaroy47/benchmark-FP32-FP16-INT8-with-TensorRT/star":{"post":"AecWsVjHlJjLYpjw1At745dKoALfm61hAIjbf_S4saELP646J9LxMcNR85_V5tWYDlDeyiD6v3dy128o3WRxfw"},"/kentaroy47/benchmark-FP32-FP16-INT8-with-TensorRT/unstar":{"post":"R9MK7kN9UXXWvP23XutFbJ1lW7F6zuTSyCN2W8Mbfjs0CoHx0y3DAzjPRAUrR85ruUsxOaYsvrL0qcV3pSQVhQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"F3_FmK7xF-nk6WsX0WEAKUnaSOodZL5P1aTsZvd8R92167sj3WCOalXZ2rjL5wRqU7S_0b6e_7F3S5GWcOCaNw"}}},"title":"Repository search results"}