{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"MiniGPT-4","owner":"Vision-CAIR","isFork":false,"description":"Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":20,"issueCount":336,"starsCount":25271,"forksCount":2899,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-16T12:33:57.592Z"}},{"type":"Public","name":"MiniGPT4-video","owner":"Vision-CAIR","isFork":false,"description":"Official code for Goldfish model for long video understanding and MiniGPT4-video for short video understanding ","allTopics":["video-understanding","video-retrieval","video-question-answering","long-video-understanding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":32,"starsCount":517,"forksCount":56,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-16T06:47:54.490Z"}},{"type":"Public","name":"Goldfish_website","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"JavaScript","color":"#f1e05a"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-08T05:58:49.042Z"}},{"type":"Public","name":"MiniGPT-Med","owner":"Vision-CAIR","isFork":false,"description":"Open-sourced code of miniGPT-Med","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":63,"forksCount":8,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,3,0,0,0,0,0,0,0,0,1,1,0,2,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-06T07:41:12.968Z"}},{"type":"Public","name":"affectiveVisDial","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":9,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-17T14:02:06.797Z"}},{"type":"Public","name":"3DCoMPaT-v2","owner":"Vision-CAIR","isFork":false,"description":"3DCoMPaT++: An improved large-scale 3D vision dataset for compositional recognition","allTopics":["computer-vision","deep-learning","3d","multimodal-deep-learning","compositional-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":75,"forksCount":7,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T17:02:27.890Z"}},{"type":"Public","name":"InfiniBench","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":1,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-01T08:32:50.030Z"}},{"type":"Public","name":"HomeGPT","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-10T22:22:42.219Z"}},{"type":"Public","name":"FishNet2.0","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T09:09:07.409Z"}},{"type":"Public","name":"MammalNet","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":25,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-15T14:25:15.906Z"}},{"type":"Public","name":"LTVRR","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":35,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-21T04:07:49.676Z"}},{"type":"Public","name":"RelTransformer","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":29,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-04T19:38:03.735Z"}},{"type":"Public","name":"affective-vision-language","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-29T15:45:20.055Z"}},{"type":"Public","name":"VisualGPT","owner":"Vision-CAIR","isFork":false,"description":"VisualGPT, CVPR 2022 Proceeding, GPT as a decoder for vision-language models","allTopics":["image-caption","visualgpt","data-efficient-image-caption"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":3,"starsCount":316,"forksCount":49,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-16T06:13:12.854Z"}},{"type":"Public","name":"saai-factory-tutorial-creative-ai","owner":"Vision-CAIR","isFork":false,"description":"Creative AI for Visual Art and Music slides and demos.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":11,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-02T11:48:44.393Z"}},{"type":"Public","name":"3DCoMPaT","owner":"Vision-CAIR","isFork":false,"description":"Official repository for the 3DCoMPaT dataset (ECCV2022 Oral)","allTopics":["computer-vision","deep-learning","dataset","compositional-data"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":16,"forksCount":1,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-19T20:40:30.692Z"}},{"type":"Public","name":"ChatCaptioner","owner":"Vision-CAIR","isFork":false,"description":"Official Repository of ChatCaptioner","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":5,"starsCount":450,"forksCount":26,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-13T20:05:09.456Z"}},{"type":"Public","name":"AF-Guide","owner":"Vision-CAIR","isFork":false,"description":"Official repository of Action-Free Guide","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":11,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-22T07:54:38.357Z"}},{"type":"Public","name":"artelingo","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":2,"starsCount":3,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-22T08:33:48.627Z"}},{"type":"Public","name":"artemis-v2","owner":"Vision-CAIR","isFork":false,"description":"Code for the paper: It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":2,"starsCount":17,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-27T11:04:58.996Z"}},{"type":"Public","name":"CWAN","owner":"Vision-CAIR","isFork":false,"description":"Creative Walk Adversarial Networks: Novel Art Generation with Probabilistic Random Walk Deviation from Style Norms","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-27T09:55:45.661Z"}},{"type":"Public","name":"Zero-Shot-Learning","owner":"Vision-CAIR","isFork":false,"description":"VisionCAIR Zero-Shot Learning Research","allTopics":["transfer-learning","zero-shot-learning","continual-learning","domain-generalization"],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-10T09:45:30.083Z"}},{"type":"Public","name":"UnlikelihoodMotionForecasting","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-05T12:17:57.705Z"}},{"type":"Public","name":"Affective-and-Creative-AI","owner":"Vision-CAIR","isFork":false,"description":"VisionCAIR Affective and Creative AI Research","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-17T20:36:46.983Z"}},{"type":"Public","name":"cs326-few-shot-classification","owner":"Vision-CAIR","isFork":false,"description":"CS326 Practical assignment #2: few-shot classification","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-07T12:03:11.039Z"}},{"type":"Public","name":"WAGA","owner":"Vision-CAIR","isFork":false,"description":"Code for Wölfflin Affective Generative Analysis paper published in ICCC 2021","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-09-17T17:41:17.596Z"}},{"type":"Public","name":"GRaWD","owner":"Vision-CAIR","isFork":false,"description":"Imaginative Walks: Generative Random Walk Deviation Loss for Improved Unseen Learning Representation. CVPR 2022 Workshop, ICCC 2022.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-09-16T14:40:29.777Z"}},{"type":"Public","name":"HalentNet","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-04-21T18:40:29.326Z"}},{"type":"Public","name":"CIZSLv2","owner":"Vision-CAIR","isFork":false,"description":"CIZSL++: Creativity Inspired Generative Zero-Shot Learning. T-PAMI under review.","allTopics":["attributes","creativity","zero-shot-learning","generative-models","vision-and-language","unseen-classes-understanding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-02-08T07:10:01.625Z"}},{"type":"Public","name":"lifelong_fact_learning","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-08-21T17:05:18.646Z"}}],"repositoryCount":33,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Vision-CAIR repositories"}