{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"VLMEvalKit","owner":"open-compass","isFork":false,"description":"Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks","allTopics":["computer-vision","evaluation","pytorch","gemini","openai","vqa","vit","gpt","multi-modal","clip","claude","openai-api","gpt4","large-language-models","llm","chatgpt","llava","qwen","gpt-4v"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":23,"starsCount":1032,"forksCount":145,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,73,30,19,15,26,35,17,26,28,23,10,2,1,10,0,8,13,8,9,8,20,32,14,16,8,22,11,5,3,8,26,15,30,36,59,33,33,11,23,33,20,27,27],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T19:15:55.530Z"}},{"type":"Public","name":"opencompass","owner":"open-compass","isFork":false,"description":"OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.","allTopics":["benchmark","evaluation","openai","llm","chatgpt","large-language-model","llama2","llama3"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":180,"starsCount":3769,"forksCount":403,"license":"Apache License 2.0","participation":[19,0,5,7,5,14,11,17,15,14,7,18,11,12,12,13,11,15,8,15,0,4,7,12,11,9,4,3,4,6,8,19,8,14,6,13,3,2,0,5,4,5,5,13,20,5,5,8,15,5,15,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T10:12:05.049Z"}},{"type":"Public","name":"MMBench","owner":"open-compass","isFork":false,"description":"Official Repo of \"MMBench: Is Your Multi-modal Model an All-around Player?\"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":143,"forksCount":9,"license":"Apache License 2.0","participation":[0,5,0,0,2,0,0,0,1,0,0,1,0,2,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-01T11:44:12.290Z"}},{"type":"Public","name":"hinode","owner":"open-compass","isFork":true,"description":"A clean documentation and blog theme for your Hugo site based on Bootstrap 5","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":51,"license":"MIT License","participation":[35,8,42,79,57,50,57,36,3,19,10,10,21,117,109,41,0,32,25,52,31,19,38,17,13,42,30,24,2,3,9,40,29,24,27,59,47,11,16,25,25,10,27,29,30,9,79,53,31,5,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-01T03:11:24.280Z"}},{"type":"Public","name":"storage","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-18T18:25:33.773Z"}},{"type":"Public","name":"CompassBench","owner":"open-compass","isFork":false,"description":"Demo data of CompassBench","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":2,"starsCount":2,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-07T12:39:22.420Z"}},{"type":"Public","name":"CIBench","owner":"open-compass","isFork":false,"description":"Official Repo of \"CIBench: Evaluation of LLMs as Code Interpreter \"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":5,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-19T04:51:48.990Z"}},{"type":"Public","name":"GAOKAO-Eval","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":2,"starsCount":88,"forksCount":6,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,10,0,0,0,10,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-17T10:58:18.281Z"}},{"type":"Public","name":"ANAH","owner":"open-compass","isFork":false,"description":"[ACL 2024] ANAH: Analytical Annotation of Hallucinations in Large Language Models","allTopics":["acl","gpt","llms","hallucination-detection"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,5,0,0,0,0,2,2,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T18:57:53.120Z"}},{"type":"Public","name":"MathBench","owner":"open-compass","isFork":false,"description":"[ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":5,"starsCount":77,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,4,0,0,0,0,0,0,6,0,0,0,0,0,0,0,0,9,1,1,0,0,0,4,0,2,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T05:17:51.572Z"}},{"type":"Public","name":"GTA","owner":"open-compass","isFork":false,"description":"Official repository for paper \"GTA: A Benchmark for General Tool Agents\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":29,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T03:23:24.351Z"}},{"type":"Public","name":".github","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T02:27:09.928Z"}},{"type":"Public","name":"DevBench","owner":"open-compass","isFork":false,"description":"A Comprehensive Benchmark for Software Development.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":83,"forksCount":4,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-30T13:10:52.870Z"}},{"type":"Public","name":"CodeBench","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T11:38:31.710Z"}},{"type":"Public","name":"Ada-LEval","owner":"open-compass","isFork":false,"description":"The official implementation of \"Ada-LEval: Evaluating long-context LLMs with length-adaptable benchmarks\"","allTopics":["gpt4","llm","long-context"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":49,"forksCount":2,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,41,2,5,3,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-22T09:36:27.993Z"}},{"type":"Public","name":"T-Eval","owner":"open-compass","isFork":false,"description":"[ACL2024] T-Eval: Evaluating Tool Utilization Capability of Large Language Models Step by Step","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":32,"starsCount":209,"forksCount":13,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-03T21:05:37.907Z"}},{"type":"Public","name":"human-eval","owner":"open-compass","isFork":true,"description":"Code for the paper \"Evaluating Large Language Models Trained on Code\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":332,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-14T11:55:53.553Z"}},{"type":"Public","name":"OpenFinData","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":32,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-08T06:32:20.542Z"}},{"type":"Public","name":"CriticBench","owner":"open-compass","isFork":false,"description":"A comprehensive benchmark for evaluating critique ability of LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":25,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T01:31:08.418Z"}},{"type":"Public","name":"code-evaluator","owner":"open-compass","isFork":false,"description":"A multi-language code evaluation tool.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":17,"forksCount":6,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-26T04:12:27.117Z"}},{"type":"Public","name":"evalplus","owner":"open-compass","isFork":true,"description":"EvalPlus for rigourous evaluation of LLM-synthesized code","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":101,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-20T08:13:21.996Z"}},{"type":"Public","name":"MixtralKit","owner":"open-compass","isFork":false,"description":"A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI","allTopics":["moe","mistral","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":762,"forksCount":81,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-15T19:10:55.603Z"}},{"type":"Public","name":"LawBench","owner":"open-compass","isFork":false,"description":"Benchmarking Legal Knowledge of Large Language Models","allTopics":["law","benchmark","llm","chatgpt"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":235,"forksCount":33,"license":"Apache License 2.0","participation":[9,12,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-13T06:42:45.191Z"}},{"type":"Public","name":"BotChat","owner":"open-compass","isFork":false,"description":"Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":132,"forksCount":6,"license":"Apache License 2.0","participation":[11,0,5,4,0,5,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T12:38:19.203Z"}},{"type":"Public","name":"pytorch_sphinx_theme","owner":"open-compass","isFork":true,"description":"Sphinx Theme for OpenCompass - Modified from PyTorch","allTopics":[],"primaryLanguage":{"name":"CSS","color":"#563d7c"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":138,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-30T04:20:35.062Z"}}],"repositoryCount":25,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"open-compass repositories"}