Cover Image for Fine-Tune Friday: Qwen 3.5 Goes Multimodal 🎬👀 – Understanding Images & Videos to Detect Actors in Scenes!
Cover Image for Fine-Tune Friday: Qwen 3.5 Goes Multimodal 🎬👀 – Understanding Images & Videos to Detect Actors in Scenes!
Avatar for Oxen.ai
Presented by
Oxen.ai
Iterate on prompts, collaborate on datasets, fine-tune custom models, and deploy at scale. All in one platform.
133 Went

Fine-Tune Friday: Qwen 3.5 Goes Multimodal 🎬👀 – Understanding Images & Videos to Detect Actors in Scenes!

Zoom
Registration
Past Event
Welcome! To join the event, please register below.
About Event

​Welcome to this week’s edition of our favorite series from the Oxen.ai Herd, Fine-Tune Fridays. Each week, we take an open-source model and put it head-to-head with a closed-source foundation model on a specialized task.

​​We share practical, end-to-end examples, including reference data, model weights, and the full infrastructure needed to reproduce the experiments on your own.

​​This Week
We’re diving into the new Qwen 3.5 series with a focus on multimodality. We’ll test how fast and capable the smaller models are for image and video tasks, like detecting actors in a scene, and see how the larger models perform across text, image, and video understanding.

​We’ll also walk through our fine-tuning and deployment pipeline and show how you can reproduce the experiments and fine-tune these powerful models yourself using Oxen.ai.

​​Looking forward to seeing you there!

Avatar for Oxen.ai
Presented by
Oxen.ai
Iterate on prompts, collaborate on datasets, fine-tune custom models, and deploy at scale. All in one platform.
133 Went