Opens profile photo
Follow
OpenGVLab
@opengvlab
Shanghai AI Lab, General Purpose Vision Team. We created InternImage, BEVFormer, VideoMAEv2, LLaMA-Adapter-V2, Ask-Anything, and many more!
Shanghai github.com/OpenGVLabJoined January 2023

OpenGVLab’s Tweets

InternGPT can now generate images based on audio input by incorporating 's ImageBind into our pipeline. learn more: github.com/OpenGVLab/Inte
Embedded video
0:57
39 views
Quote Tweet
Image
Very excited to announce our work, InternChat. We can now interact with #ChatGPT using the cursor, bringing human interaction to a whole new level. Demo built with @Gradio No OpenAI API key needed for a limited time! For the demo and more samples: github.com/OpenGVLab/Inte
Show this thread
1
Exciting progress for FocalNet! Wonder if the new method can be applied to other players on the chart too. Our InternImage at 65.4mAP is feeling the heat🔥
Quote Tweet
Our FocalNet is shining again! Combing <700M focalnet-huge and stable-DINO, we achieved 64.8 mAP with only ImageNet-22K and Object365, without any test time augmentation! It beats EVA and only lags behind 3B InternImage, gives you the strongest reproducible object detector! twitter.com/jw2yang4ai/sta…
Show this thread
Image
2
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark #LLMs side-by-side while providing images as inputs. Which model would you like to see being supported? Demo at vlarena.opengvlab.com built with Code: github.com/OpenGVLab/Mult
Quote Tweet
Announcing the Week 2 update for the Chatbot Arena leaderboard! We've added some new models that are showcasing strong performance. Currently, @OpenAI's GPT-4 and @AnthropicAI's Claude lead the pack, with open-source models in hot pursuit. More findings: lmsys.org/blog/2023-05-1
Show this thread
Image
Demo now live! Our latest VideoChat connects video foundation models with #LLMs via a learnable neural interface in an end-to-end manner, exhibiting video deep video understanding such as "Why is this video funny?" Demo & more: github.com/OpenGVLab/Ask-
Embedded video
1:52
42 views
Quote Tweet
Image
VideoChat: Chat-Centric Video Understanding abs: arxiv.org/abs/2305.06355 paper page: huggingface.co/papers/2305.06 github: github.com/OpenGVLab/Ask-
7
Very excited to announce our work, InternChat. We can now interact with #ChatGPT using the cursor, bringing human interaction to a whole new level. Demo built with No OpenAI API key needed for a limited time! For the demo and more samples: github.com/OpenGVLab/Inte
Image
Quote Tweet
Image
InternChat: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language abs: arxiv.org/abs/2305.05662 paper page: huggingface.co/papers/2305.05 github: github.com/OpenGVLab/Inte
1
49
Show this thread
A great recap of recent #LLM progress. We will look into generating the computational performance benchmarks of our LLaMA-Adapter V2 like you mentioned in the blog, thanks
Quote Tweet
It was a great month for open source: So many LLMs came out that it's become quite overwhelming to keep track of it all. So, in this month's Ahead of AI issue, I am sharing resources and research insights on the latest open-source LLMs & datasets! magazine.sebastianraschka.com/p/ahead-of-ai-
LLaMA-Adapter V2: This update allows the 65B model to surpass #ChatGPT on some questions in terms of response quality, while also outperforming #Vicuna Congratulations to our team! Code at github.com/ZrrSkywalker/L Follow us for more🫰
Quote Tweet
🚀65B LLaMA-Adapter-V2 code & checkpoint are NOW ready at github.com/ZrrSkywalker/L! 🛠️Big update enhancing multimodality & chatbot. 🔥LLaMA-Adapter-V2 surpasses #ChatGPT in response quality (102%:100%) & beats #Vicuna in win-tie-lost (50:14). ☕️Thanks to Peng Gao & @opengvlab! 2/2
Show this thread
Embedded video
2:11
13.8K views
6
Thank you . VideoMAE is one of our good works😀
Quote Tweet
Transformers v4.22 is out, and includes the first VIDEO models! 🎥 💥VideoMAE: masked auto-encoders for video 💥X-CLIP: CLIP for video-language Other nice goodies: 💥Swin Transformer v2 💥Pegasus-X 💥Donut 💥MobileViT ... and MacOS support (device="mps")!
Show this thread
Image
1