Multimodal AI — systems that can understand and respond through text, images, and voice — is redefining how we interact with technology. In this lab, we’ll show how to build a multimodal agent that seamlessly blends language, vision, and voice inputs — and also speaks back — creating a fully natural human-AI conversation flow.
Developers & ML engineers adding voice/vision to AI
Product & CX leads rethinking support with agents
Business leaders automating workflows with AI
Ready to slash development time and ship your next AI idea lightning-fast?
→ Save my seat now
Matt Paige is VP of Strategy & Marketing at HatchWorks AI, where he leads go-to-market strategy, education, and community around AI-native software. As host of the Talking AI podcast and founder of HatchWorks AI Labs, he’s known for turning complex AI concepts into practical playbooks. With a background in business strategy and a passion for hands-on training, Matt helps teams—from startups to enterprises—build real capability in the age of intelligent software.
Avoid pitfalls and keep full control over your strategy.
A round-up of industry stats, research, and insights to understand where AI stands, how it got here, and where it’s going.