{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"Open-Sora-Plan","owner":"PKU-YuanGroup","isFork":false,"description":"This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":112,"starsCount":10463,"forksCount":940,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,222,53,11,24,58,49,5,3,3,0,1,36],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:09:31.891Z"}},{"type":"Public","name":"MagicTime","owner":"PKU-YuanGroup","isFork":false,"description":"MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators","allTopics":["time-lapse","video-generation","diffusion-models","text-to-video","long-video-generation","time-lapse-dataset","open-sora-plan","metamorphic-video-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":1145,"forksCount":112,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,35,77,0,0,0,2,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T13:20:12.948Z"}},{"type":"Public","name":"Chat-UniVi","owner":"PKU-YuanGroup","isFork":false,"description":"[CVPR 2024 Highlight🔥] Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding","allTopics":["video-understanding","image-understanding","large-language-models","vision-language-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":661,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,12,35,7,7,0,0,1,4,1,0,0,0,0,0,0,1,1,0,0,0,3,1,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T05:05:37.066Z"}},{"type":"Public","name":"Video-LLaVA","owner":"PKU-YuanGroup","isFork":false,"description":"Video-LLaVA: Learning United Visual Representation by Alignment Before Projection","allTopics":["multi-modal","instruction-tuning","large-vision-language-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":71,"starsCount":2533,"forksCount":188,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,10,50,36,5,2,0,0,1,1,0,28,5,2,0,3,0,1,0,0,0,0,0,0,0,0,0,0,3,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T05:28:53.277Z"}},{"type":"Public","name":"Hallucination-Attack","owner":"PKU-YuanGroup","isFork":false,"description":"Attack to induce LLMs within hallucinations","allTopics":["nlp","ai-safety","hallucinations","llm","llm-safety","machine-learning","deep-learning","adversarial-attacks"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":84,"forksCount":11,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T08:48:15.526Z"}},{"type":"Public","name":"Envision3D","owner":"PKU-YuanGroup","isFork":false,"description":"Envision3D: One Image to 3D with Anchor Views Interpolation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":89,"forksCount":8,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T18:46:32.791Z"}},{"type":"Public","name":"MoE-LLaVA","owner":"PKU-YuanGroup","isFork":false,"description":"Mixture-of-Experts for Large Vision-Language Models","allTopics":["moe","multi-modal","mixture-of-experts","large-vision-language-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":47,"starsCount":1755,"forksCount":103,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,10,7,0,38,58,58,18,30,2,0,0,2,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-15T14:50:56.572Z"}},{"type":"Public","name":"ProLLaMA","owner":"PKU-YuanGroup","isFork":false,"description":"A Protein Large Language Model for Multi-Task Protein Language Processing","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":84,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-27T09:51:11.215Z"}},{"type":"Public","name":"Open-Sora-Dataset","owner":"PKU-YuanGroup","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":58,"forksCount":3,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,3,4,0,12,0,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-24T07:13:07.657Z"}},{"type":"Public","name":"LanguageBind","owner":"PKU-YuanGroup","isFork":false,"description":"【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment","allTopics":["multi-modal","zero-shot","pretraining","language-central"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":16,"starsCount":564,"forksCount":46,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-25T12:28:38.429Z"}},{"type":"Public","name":"TaxDiff","owner":"PKU-YuanGroup","isFork":false,"description":"The official code for \"TaxDiff: Taxonomic-Guided Diffusion Model for Protein Sequence Generation\"","allTopics":["protein-sequences","generate-model","meachine-learning","ai4science"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":36,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-04T01:21:18.395Z"}},{"type":"Public","name":"Peer-review-in-LLMs","owner":"PKU-YuanGroup","isFork":false,"description":"Peer-review-in-LLMs: Automatic Evaluation Method for LLMs in Open-environment,https://arxiv.org/pdf/2402.01830.pdf","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":24,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-07T02:30:30.281Z"}},{"type":"Public","name":"Machine-Mindset","owner":"PKU-YuanGroup","isFork":false,"description":"An MBTI Exploration of Large Language Models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":402,"forksCount":18,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-02T02:53:05.571Z"}},{"type":"Public","name":"Video-Bench","owner":"PKU-YuanGroup","isFork":false,"description":"A Comprehensive Benchmark and Toolkit for Evaluating Video-based Large Language Models!","allTopics":["benchmark","toolkit","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":98,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-31T03:41:41.734Z"}}],"repositoryCount":14,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}