Text to 3D generation in Apple Vision Pro built with the VisionOS SDK. 3D Scribblenauts in AR for the Scale Generative AI Hackathon. Won Scale AI Prize
Welcome to Dream with Vision Pro, a lucid text-to-3D tool built with the Apple VisionOS SDK. Powered by Scale AI’s Spellbook, OpenAI’s GPT-4 and Shap-E, Modal, Replicate, and the Meta Quest 2, we empower you to transform your imagination into stunning immersive experiences.
Type in the text description of the object you envision. This could be anything from an elephant to a sword. Unleash your imagination. Once you’ve described it, your object will appear before you.
Using Scale AI’s Spellbound to infer the size of the objects to render accurately.
Here’s a step-by-step breakdown of what Dream with Vision Pro does:
First, the user specifies the object they want to visualize. This input triggers the Shap-E model via Modal and Replicate, producing a .obj file - a standard 3D model format.
Next, we employ Spellbook and GPT-4 to estimate the object’s height, ensuring the 3D representation is accurately scaled.
The final phase employs 3D Viewer to convert your .obj file into a realistic 3D model that you can interact with. This 3D model can be directly accessed from Apple’s VisionOS, which we stream directly to your Meta Quest 2, offering a fully immersive experience of your original concept.
As an AI system, you are extremely skilled at extracting objects and estimating their realistic height in meters from a given text prompt. Your task is to identify the object(s) mentioned in the prompt and their estimated height in meters. Once identified, the information must be formatted according to the provided format for a text-to-3D model application.
Could you extract the object and realistic object height in meters from the following text prompts?
Begin:
Input: a red apple
Output: 0.075
Input: a large elephant
Output: 3.000
Input: {{ input }}
Output:
We’ve started to integrate OpenAI’s Whisper model, expanding our capability beyond text-to-3D transformations. Users will be able to engage in a more intuitive way, interacting with their 3D creations through the power of voice.
Once we have the .obj file, we are working on using USZD Tools which lets us convert to the .usdz format - a requisite for VisionOS. Following this conversion, we can seamlessly render the objects.
We thank the Scale AI Spellbook team for the credits and ease of use, Ben Firshman of Replicate for the dedicated A100 GPU we run Shap-E on, Erik Bernhardsson of Modal for dedicated Whisper and hosted endpoints, and especially Mehran Jalali for letting us borrow the Meta Quest 2 for testing.