{"payload":{"header_redesign_enabled":false,"results":[{"id":"445607550","archived":false,"color":"#DA5B0B","followers":31,"has_funding_file":false,"hl_name":"tezansahu/VQA-With-Multimodal-Transformers","hl_trunc_description":"Exploring multimodal fusion-type transformer models for visual question answering (on DAQUAR dataset)","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":445607550,"name":"VQA-With-Multimodal-Transformers","owner_id":31898274,"owner_login":"tezansahu","updated_at":"2022-01-20T18:42:22.109Z","has_issues":true}},"sponsorable":false,"topics":[],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":68,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Atezansahu%252FVQA-With-Multimodal-Transformers%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/tezansahu/VQA-With-Multimodal-Transformers/star":{"post":"-5-SpEhrE2wCWJLxdfX6zTQG68Frcery1qn1EpOC1THHbXoqpZmJb5DQdrhxZFWRT0PcM7zHzzW1AWmQuUEEcg"},"/tezansahu/VQA-With-Multimodal-Transformers/unstar":{"post":"29m6eQtX4kivnE1aVDjDp8OMtbRukEFysmgxP8JNRPlPSSbtWVcS6qbamALZm8A_lb1cAjZRIdL7IJoX9ETlrw"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"G6NiozgXJsHINh8Z3qViB2pLB4whrXDOqLuQuq9z6dTRecLk1CdKA3lBzVMrQgtiBNeqJggc4GrF-JpwyVT_tQ"}}},"title":"Repository search results"}