{"payload":{"pageCount":3,"repositories":[{"type":"Public","name":"LLaVA-MORE","owner":"aimagelab","isFork":false,"description":"LLaVA-MORE: Enhancing Visual Instruction Tuning with LLaMA 3.1","allTopics":["vision-and-language","llms","llava","multimodal-llms","llama3","llava-llama3","llama3-vision","llama3-1"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":80,"forksCount":5,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-17T16:02:54.924Z"}},{"type":"Public","name":"mammoth","owner":"aimagelab","isFork":false,"description":"An Extendible (General) Continual Learning Framework based on Pytorch - official codebase of Dark Experience for General Continual Learning","allTopics":["deep-learning","knowledge-distillation","neurips2020","dark-experience-replay","pytorch","der","continual-learning","experience-replay"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":532,"forksCount":95,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-12T11:55:04.720Z"}},{"type":"Public","name":"DiCO","owner":"aimagelab","isFork":false,"description":"Revisiting Image Captioning Training Paradigm via Direct CLIP-based Optimization (BMVC 2024 Oral ✨)","allTopics":["image-captioning","captioning-images","vision-and-language","caption-generation","bmvc2024"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-11T12:08:12.274Z"}},{"type":"Public","name":"awesome-human-visual-attention","owner":"aimagelab","isFork":false,"description":"This repository contains a curated list of research papers and resources focusing on saliency and scanpath prediction, human attention, human visual search.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":36,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-10T09:02:01.256Z"}},{"type":"Public","name":"coldfront","owner":"aimagelab","isFork":true,"description":"HPC Resource Allocation System","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":81,"license":"GNU General Public License v3.0","participation":[0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,4,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T13:36:25.838Z"}},{"type":"Public","name":"Alfie","owner":"aimagelab","isFork":false,"description":"Democratising RGBA Image Generation With No $$$ (AI4VA@ECCV24)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":13,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-02T15:01:38.560Z"}},{"type":"Public","name":"MAD","owner":"aimagelab","isFork":false,"description":"Official PyTorch implementation for \"Semantically Coherent Montages by Merging and Splitting Diffusion Paths\", presenting the Merge-Attend-Diffuse operator (ECCV24)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-02T11:59:18.166Z"}},{"type":"Public","name":"FourBi","owner":"aimagelab","isFork":false,"description":"Binarizing Documents by Leveraging both Space and Frequency. (ICDAR 2024)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-02T11:58:26.867Z"}},{"type":"Public","name":"mugat","owner":"aimagelab","isFork":false,"description":"Official implementation of our ECCVW paper \"μgat: Improving Single-Page Document Parsing by Providing Multi-Page Context\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-30T18:06:56.357Z"}},{"type":"Public","name":"freeda","owner":"aimagelab","isFork":false,"description":"FreeDA: Training-Free Open-Vocabulary Segmentation with Offline Diffusion-Augmented Prototype Generation (CVPR 2024)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":20,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-28T12:12:04.913Z"}},{"type":"Public","name":"safe-clip","owner":"aimagelab","isFork":false,"description":"Safe-CLIP: Removing NSFW Concepts from Vision-and-Language Models. ECCV 2024","allTopics":["retrieval","safety","image-to-text","text-to-image","nsfw","vision-and-language","trustworthy-ai","eccv2024"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":34,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-10T10:51:51.390Z"}},{"type":"Public","name":"CoDE","owner":"aimagelab","isFork":false,"description":"[ECCV'24] Contrasting Deepfakes Diffusion via Contrastive Learning and Global-Local Similarities","allTopics":["deepfake-detection","global-local"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-31T11:55:50.912Z"}},{"type":"Public","name":"Alfie_preeccvw","owner":"aimagelab","isFork":false,"description":"Training-Free Illustrations Generation with Diffusion Transformers","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-30T16:19:57.508Z"}},{"type":"Public","name":"bridge-score","owner":"aimagelab","isFork":false,"description":"BRIDGE: Bridging Gaps in Image Captioning Evaluation with Stronger Visual Cues. ECCV 2024","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-17T14:50:22.068Z"}},{"type":"Public","name":"CSL-TAL","owner":"aimagelab","isFork":false,"description":"Pytorch code for ECCVW 2022 paper \"Consistency-based Self-supervised Learning for Temporal Anomaly Localization\"","allTopics":["computer-vision","deep-learning","pytorch","video-anomaly-detection","self-supervised-learning","eccv2022"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":12,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T11:44:55.045Z"}},{"type":"Public","name":"Emuru","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-03T18:46:25.816Z"}},{"type":"Public","name":"CaSpeR","owner":"aimagelab","isFork":false,"description":"Code implementation for \"Latent Spectral Regularization for Continual Learning\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-26T14:36:32.101Z"}},{"type":"Public","name":"mil4wsi","owner":"aimagelab","isFork":false,"description":"DAS-MIL: Distilling Across Scales for MILClassification of Histological WSIs","allTopics":["mil","wsi"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":38,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-26T09:12:39.764Z"}},{"type":"Public","name":"SCAD-LOD-2024","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T14:17:00.411Z"}},{"type":"Public","name":"PMA-Net","owner":"aimagelab","isFork":false,"description":"With a Little Help from your own Past: Prototypical Memory Networks for Image Captioning. ICCV 2023","allTopics":["transformer","image-captioning","captioning-images","captioning","vision-and-language","vision-language","memory-augmented-neural-networks","iccv2023"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":16,"forksCount":2,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T08:52:03.651Z"}},{"type":"Public","name":"cvcs2023","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T06:38:32.256Z"}},{"type":"Public","name":"open-fashion-clip","owner":"aimagelab","isFork":false,"description":"This is the official repository for the paper \"OpenFashionCLIP: Vision-and-Language Contrastive Learning with Open-Source Fashion Data\". ICIAP 2023","allTopics":["clip","vision-and-language","fashionai","contrastive-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":52,"forksCount":5,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T06:34:26.863Z"}},{"type":"Public","name":"unveiling-the-truth","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-30T08:53:04.795Z"}},{"type":"Public","name":"multimodal-garment-designer","owner":"aimagelab","isFork":false,"description":"This is the official repository for the paper \"Multimodal Garment Designer: Human-Centric Latent Diffusion Models for Fashion Image Editing\". ICCV 2023","allTopics":["computer-vision","image-editing","generative-models","dresscode","fashionai","diffusion-models","stable-diffusion","latent-diffusion-models","viton-hd","iccv2023","multimodal-fashion-image-editing"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":405,"forksCount":47,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-28T17:37:43.235Z"}},{"type":"Public","name":"Ti-MGD","owner":"aimagelab","isFork":false,"description":"This is the official repository for the paper \"Multimodal-Conditioned Latent Diffusion Models for Fashion Image Editing\".","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":25,"forksCount":1,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-28T07:51:06.884Z"}},{"type":"Public","name":"Teddy","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-18T10:16:23.815Z"}},{"type":"Public","name":"HWD","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":15,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-14T11:49:24.798Z"}},{"type":"Public","name":"COCOFake","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":1,"license":"Creative Commons Attribution 4.0 International","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-02T17:28:07.462Z"}},{"type":"Public","name":"regesta_OCR","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-07T15:00:31.532Z"}},{"type":"Public","name":"VATr","owner":"aimagelab","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":71,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-29T13:23:26.304Z"}}],"repositoryCount":68,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"aimagelab repositories"}