Monitor the performance of OpenAI's GPT-4V model over time.
-
Updated
May 19, 2024 - HTML
Monitor the performance of OpenAI's GPT-4V model over time.
Vision utilities for web interaction agents 👀
Control Any Computer Using LLMs
AI Voiceover with GPT4V
The ultimate sketch to code app made using GPT4 vision. Choose your desired framework (React, Next, React Native, Flutter) for your app. It will instantly generate code and preview (sandbox) from a simple hand drawn sketch on paper captured from webcam
Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception
Explore the rich flavors of Indian desserts with TunedLlavaDelights. Utilizing the in Llava fine-tuning, our project unveils detailed nutritional profiles, taste notes, and optimal consumption times for beloved sweets. Dive into a fusion of AI innovation and culinary tradition
Multi-Modal Multi-Embodied Hivemind-like Iteration of RTX-2
Implementation of MambaByte in "MambaByte: Token-free Selective State Space Model" in Pytorch and Zeta
Convert different model APIs into the OpenAI API format out of the box.
Vision-Assisted Camera Orientation
I GAVE GPT-4 EYES!
[READ-ONLY] Describe images and generate alt tags for visually impaired users.
Add a description, image, and links to the gpt4v topic page so that developers can more easily learn about it.
To associate your repository with the gpt4v topic, visit your repo's landing page and select "manage topics."