- [R] Going further under Grounded-Segment-Anything: integrating Whisper and ChatGPT https://github.com/IDEA-Research/Grounded-Segment-Anything 8 comments machinelearning
- [P] Grounded-Segment-Anything: Zero-shot Detection and Segmentation https://github.com/IDEA-Research/Grounded-Segment-Anything 7 comments machinelearning
Linking pages
- GitHub - ShaShekhar/aaiela https://github.com/ShaShekhar/aaiela 30 comments
- GitHub - sail-sg/EditAnything https://github.com/sail-sg/EditAnything 24 comments
- GitHub - IDEA-Research/GroundingDINO: The official implementation of "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection" https://github.com/IDEA-Research/GroundingDINO 6 comments
- GitHub - autodistill/autodistill: Images to inference with no labeling (use foundation models to train supervised models) https://github.com/autodistill/autodistill 5 comments
- Grounded-Segment-Anything/playground/ImageBind_SAM at main · IDEA-Research/Grounded-Segment-Anything · GitHub https://github.com/IDEA-Research/Grounded-Segment-Anything/tree/main/playground/ImageBind_SAM 4 comments
- GitHub - ycheng517/tabletop-handybot: A low-cost AI powered robotic arm assistant https://github.com/ycheng517/tabletop-handybot 1 comment
- GitHub - UX-Decoder/Segment-Everything-Everywhere-All-At-Once https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once 0 comments
- GitHub - Adamdad/Awesome-ComposableAI: A curated list of Composable AI methods: Building AI system by composing modules. https://github.com/Adamdad/Awesome-ComposableAI 0 comments
- GitHub - xinyu1205/Recognize_Anything-Tag2Text: Code for the Recognize Anything Model and Tag2Text Model https://github.com/xinyu1205/Recognize_Anything-Tag2Text 0 comments
- GitHub - haotian-liu/LLaVA: Visual Instruction Tuning: Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities. https://github.com/haotian-liu/LLaVA 0 comments
- GitHub - asFeng/d-edit: The implementation of "An item is Worth a Prompt: Versatile Image Editing with Disentangled Control" https://github.com/asFeng/d-edit 0 comments
Linked pages
- GitHub - microsoft/visual-chatgpt: VisualChatGPT https://github.com/microsoft/visual-chatgpt 229 comments
- Segment Anything | Meta AI https://segment-anything.com/ 157 comments
- GitHub - openai/whisper: Robust Speech Recognition via Large-Scale Weak Supervision https://github.com/openai/whisper/ 126 comments
- LLaVA https://llava-vl.github.io/ 54 comments
- GitHub - sail-sg/EditAnything https://github.com/sail-sg/EditAnything 24 comments
- Segment Anything | Meta AI https://segment-anything.com/demo 9 comments
- GitHub - IDEA-Research/GroundingDINO: The official implementation of "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection" https://github.com/IDEA-Research/GroundingDINO 6 comments
- GitHub - autodistill/autodistill: Images to inference with no labeling (use foundation models to train supervised models) https://github.com/autodistill/autodistill 5 comments
- GitHub - CompVis/stable-diffusion: A latent text-to-image diffusion model https://github.com/CompVis/stable-diffusion 4 comments
- Recognize Anything: A Strong Image Tagging Model https://recognize-anything.github.io/ 4 comments
- Start Locally | PyTorch https://pytorch.org/get-started/locally/ 3 comments
- [2304.08485] Visual Instruction Tuning https://arxiv.org/abs/2304.08485 1 comment
- GitHub - roboflow/notebooks: Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM. https://github.com/roboflow/notebooks 1 comment
- GitHub - ycheng517/tabletop-handybot: A low-cost AI powered robotic arm assistant https://github.com/ycheng517/tabletop-handybot 1 comment
- GitHub - salesforce/LAVIS: LAVIS - A One-stop Library for Language-Vision Intelligence https://github.com/salesforce/LAVIS 0 comments
- [2303.04671] Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models https://arxiv.org/abs/2303.04671 0 comments
- GitHub - UX-Decoder/Segment-Everything-Everywhere-All-At-Once https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once 0 comments
- GitHub - facebookresearch/segment-anything: The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model. https://github.com/facebookresearch/segment-anything 0 comments
- [2304.06718] Segment Everything Everywhere All at Once https://arxiv.org/abs/2304.06718 0 comments
- GitHub - ttengwang/Caption-Anything: Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://github.com/ttengwang/Caption-Anything 0 comments
Would you like to stay up to date with Computer science? Checkout Computer science
Weekly.