Vision language models as a design material

Learn to directly shape AI for products that “see,” so you can ship the right thing, influence engineering decisions, and demonstrate your value.

Vision language models (VLMs) are AI systems that can “interpret” images and respond to them. If AI-powered “seeing” is a core part of what you’re building, this workshop helps you get hands-on enough to stay in the room when the hard decisions get made.

(When is this in other time zones?)



You’ll leave with outcomes you can use immediately

You’ll learn how to:

And, you’ll take home:



Don’t get left out of key product decisions

When the core of a product is highly technical — AI vision, sensors, “smart” features — designers with primarily 2D practice (UX, interaction, visual design, research) often get boxed out of the parts that determine what’s actually possible.

Your usual design artifacts — flows, mockups, prototypes, storyboards — are great at communicating intent. But, with vision language models and other complex technologies, they often leave critical questions unanswered, forcing engineering to guess, compromise, and ship avoidable failure modes.

This workshop gives you a way to generate empirical, defensible knowledge about the model’s behavior, so you can:

I was surprised to see how this applies to industrial design — and humbled by the amount of considerations it takes to launch a physical product.

Participant

Book workshop, $225



A proven method

The design techniques in this workshop were formalized in the late 2000s, when designers and academics developed methods and practices for non-engineers to make sense of technologies, from NFC tags and GPS signals, to mobile phone sensors and physical AI. I began teaching it in 2012.

For over a decade, this framework has helped designers explore sensors and prototype smart devices without needing to learn to code or become electrical engineers. Now, I’ve updated this rigorous, material-based approach specifically for the current generation of generative AI.

Watch the material exploration process in this ~2m excerpt from a previous workshop recording, as a vision language model’s constraints are tested with two different images. Note the different number of turkeys and fauna being identified as we change how the model is prompted.

Book workshop, $225



Proof you can show your org (and future employers)

If your organization values formal, legible learning: you’ll receive a digital certificate of completion for the public workshop.

A sample digital certificate issued for completing the workshop.

Book workshop, $225



Inside the workshop

There will be two 5m breaks.

The third hour is entirely practice with the vision language model.

Watch an entire previous run of this workshop in under 60 seconds.

Book workshop, $225



What model will we use?

We’ll use Perceptron Isaac 0.1, a “perceptive-language” model trained to “understand” the physical world.

It can:

The point isn’t to become an “Isaac expert.” The point is to learn a method you can apply to any vision-capable assistant or VLM.

Watch the Issac 0.1 model identify all the objects in this scene, frame-by-frame, when prompted in a naïve fashion. You’ll learn many of the levers you can use to adjust this output, and get to experiment with them directly.

Book workshop, $225



Prerequisites

No technical experience necessary.

Ensure your Zoom app is updated to the latest version.

Visit training.tertile.one from the computer and internet connection you will be participating from ahead of the workshop, and contact with any issues for troubleshooting steps: .



Who it’s for

This is for you if you’re a:

This is especially useful if:

Book workshop, $225



About the instructor

A public workshop from Tertile, LLC, led by Vitorio Miliano.

Vitorio Miliano has trained designers to work with advanced technologies as design materials for over a decade, from small shops to the Fortune 5. He brings experience across product management, software development, research, and design, with a track record of evidence-based decision-making at program and product levels.

Prior to Tertile, his past work includes a healthcare news briefing on the Amazon Alexa platform, a 3D environment used by NASA to visualize the International Space Station, and building research and developer-relations programs in industry.



Alternative scheduling (private groups)

For private trainings for teams/groups of 4+ (including corporate, on-site in-person, or full-day sessions), contact: , .



Frequently asked questions

Billing

Public workshops booked by credit card are billed immediately. Refunds are available by request until one business day before the workshop. Contact for refunds or alternative billing arrangements: , .

Materials

All original materials provided to participants are licensed by Tertile, LLC for their private, personal use, and are not to be shared or redistributed. Licenses are revoked in the event of a post-training refund or chargeback, and all materials must be destroyed.

Terms of service and privacy policy

Payment for and participation in training is governed by Tertile, LLC’s terms of service (PDF) and privacy policy (PDF).

Book workshop, $225