groundingLMM

[CVPR 2024 ๐Ÿ”ฅ] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses that are seamlessly integrated with object segmentation masks.

by mbzuai-oryx

60.6
High Trust
🔗 View Source 👤 Claim

🚀 Why groundingLMM?

groundingLMM is a standout agent — gaining momentum with 949 stars, a solid trust score of 60.6/100, and native support for REST.

🎨

Image Processing

Analyze, generate, and transform visuals

🔌 Protocols & Compatibility

REST
⚡ Capabilities
image-understanding

🔧 Technical Specifications

TypeAgent
LanguagePython
Trust Score60.6/100 (High)
Stars★ 949
CategoriesGeneral
ProtocolsREST
Sourcehttps://github.com/mbzuai-oryx/groundingLMM
60.6/100
Trust Score
Well-established with solid validation
949
GitHub Stars
Emerging community interest
Unverified
Not yet claimed or verified
Tags:#foundation-models#llm-agent#lmm#vision-and-language#vision-language-model

🏷️ Embed Badge

Add a trust badge to your README:

Trust Score Stars
[![Fushu](https://fushu.dev/badge/2e39b81072de/trust.svg)](https://fushu.dev/agent/2e39b81072de) click to copy

Get Started with groundingLMM

Install now and integrate into your workflow in minutes.

Share this agent: Twitter / X LinkedIn
← Back to Directory