{"payload":{"header_redesign_enabled":false,"results":[{"id":"493887055","archived":false,"color":"#DA5B0B","followers":238,"has_funding_file":false,"hl_name":"implus/UM-MAE","hl_trunc_description":"Official Codes for \"Uniform Masking: Enabling MAE Pre-training for Pyramid-based Vision Transformers with Locality\"","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":493887055,"name":"UM-MAE","owner_id":4126718,"owner_login":"implus","updated_at":"2022-12-03T03:45:49.235Z","has_issues":true}},"sponsorable":false,"topics":["coco","mae","ade20k","self-supervised-learning","masked-autoencoder","imagenet-classification","pyramid-vision-transformer","swin-transformer","masked-image-modeling","hierarchical-vision-transformer"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":107,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aimplus%252FUM-MAE%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/implus/UM-MAE/star":{"post":"WrSLesVPAtFVsY0nyzZY8REbI3dZi4a1yOk7oeDdxzC15f0VumA9MJf9suEc47Pcxjpj0jkP_a9OMABizEE2Tg"},"/implus/UM-MAE/unstar":{"post":"V4Xoq06qf4jyAwLeq-As-QKSVVA0G6Bv6ZzpipFQzwaP9zAJJHlyOZGkTvgtzWyHpQxamvL1IFND2_muN1vPAg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"wjshK4_jVlh7AXrsg02nR906E1dnicmv5P0xt2_ay3KykmeejA4JfVtqRr3eQwQoNuXwgDKGXjOBlhZirz_tfA"}}},"title":"Repository search results"}