Industry Insight
Featured

Alberto Gimeno Participates as Judge at YC x Google DeepMind Multimodal AI Hackathon

Thuy Vi Nguyen

Inbound Sales Marketing

3.10.2026

min read

Challenge

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

Solution

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

Result

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

Recently, Invofox CEO Alberto Gimeno participated as a judge at the YC x Google DeepMind Multimodal Frontier Hackathon, an in-person event held in San Francisco that brought together founders, engineers, and researchers working at the cutting edge of multimodal AI.

The event, hosted by Y Combinator in collaboration with Google DeepMind, focused on exploring new types of applications enabled by the latest generation of multimodal AI systems. Participants were encouraged to build products that combine audio, video, and image generation capabilities, moving beyond the traditional “chatbot” paradigm that has dominated much of the current AI landscape.

More details about the event can be found on the official page:

https://events.ycombinator.com/deepmind-march26

Alberto Gimeno with participants during the YC x Google DeepMind Multimodal Frontier Hackathon in San Francisco.

Evaluating the Next Generation of Multimodal AI Projects

As part of the judging process, Alberto joined a group of founders and engineers, including Y Combinator founders and Google DeepMind engineers, to review projects built during the hackathon.

In the initial evaluation round, Alberto reviewed and scored several submitted projects across a number of criteria, including:

  • Technical feasibility
  • Innovation and novelty
  • Real-world applicability
  • Market potential and fundability

As part of the evaluation process, Alberto provided independent scoring and feedback on the projects based on their technical execution, originality, and potential real-world impact.

This early review helped determine which teams would move forward to the final stage of the competition. The final judging panel consisted exclusively of Google engineers, who selected the overall winners from the shortlisted projects.

During the preliminary judging round, Alberto evaluated several emerging ideas exploring new applications of multimodal AI, including:

  • AI systems that generate cinematic video narratives from user-provided images
  • Tools for dynamically generating alternative story outcomes for media content
  • AI-powered contract interpretation and legal analysis
  • Robotics-focused applications
  • Accessibility tools designed to support navigation for visually impaired users

These projects demonstrated how quickly multimodal AI capabilities are evolving and how developers are experimenting with entirely new categories of software built on top of these models.

Selected Finalist: Solus Forge

Based on the preliminary evaluation process, Solus Forge was selected as a finalist to advance to the final round of judging.

The project stood out for its strong combination of technical feasibility, innovation, and real-world applicability, and was among the top-scoring projects in Alberto’s evaluation.

Exploring the Next Generation of Multimodal AI

The hackathon encouraged teams to experiment with advanced multimodal AI technologies developed by Google DeepMind, including:

  • Gemini 3.1, featuring expanded long-context reasoning and native agentic vision capabilities
  • Lyria, DeepMind’s model for high-fidelity music and audio generation
  • NanoBanana 2, designed for advanced image composition, character consistency, and detailed text rendering

By building with these tools, participants explored new forms of multimodal applications that combine visual, audio, and language-based reasoning.

For industry practitioners like Alberto — who works closely with AI systems that extract and structure information from complex documents at Invofox — participating in events like this provides a valuable opportunity to evaluate emerging ideas and contribute expertise to the broader AI ecosystem.

Supporting Innovation in Applied AI

Hackathons like the YC x Google DeepMind Multimodal Frontier Hackathon serve as important environments for experimentation and collaboration between founders, engineers, and researchers.

By bringing together experts from leading organizations and early-stage builders, these events help accelerate the development of new AI applications and highlight the expanding possibilities of multimodal systems.

For Invofox, staying closely connected to these developments helps inform how next-generation AI capabilities may shape the future of document automation and data extraction technologies.

Thuy Vi Nguyen

Guest Contributor

Thuy Vi Nguyen is an Inbound Sales Marketing Specialist at Invofox, where she focuses on growth, demand generation, and go-to-market strategy. She has over a decade of experience in B2B SaaS across marketing, sales, and customer experience, and has led marketing initiatives for multiple technology companies.

Table of Contents

Subscribe to Our Blog

Subscribe for tips and insights from Invofox — the intelligent document processing (IDP) platform that helps businesses automate invoices, receipts, and more.

Industry Insight
Featured

Alberto Gimeno Participates as Judge at YC x Google DeepMind Multimodal AI Hackathon

Thuy Vi Nguyen

Inbound Sales Marketing

3.10.2026

min read

Recently, Invofox CEO Alberto Gimeno participated as a judge at the YC x Google DeepMind Multimodal Frontier Hackathon, an in-person event held in San Francisco that brought together founders, engineers, and researchers working at the cutting edge of multimodal AI.

The event, hosted by Y Combinator in collaboration with Google DeepMind, focused on exploring new types of applications enabled by the latest generation of multimodal AI systems. Participants were encouraged to build products that combine audio, video, and image generation capabilities, moving beyond the traditional “chatbot” paradigm that has dominated much of the current AI landscape.

More details about the event can be found on the official page:

https://events.ycombinator.com/deepmind-march26

Alberto Gimeno with participants during the YC x Google DeepMind Multimodal Frontier Hackathon in San Francisco.

Evaluating the Next Generation of Multimodal AI Projects

As part of the judging process, Alberto joined a group of founders and engineers, including Y Combinator founders and Google DeepMind engineers, to review projects built during the hackathon.

In the initial evaluation round, Alberto reviewed and scored several submitted projects across a number of criteria, including:

  • Technical feasibility
  • Innovation and novelty
  • Real-world applicability
  • Market potential and fundability

As part of the evaluation process, Alberto provided independent scoring and feedback on the projects based on their technical execution, originality, and potential real-world impact.

This early review helped determine which teams would move forward to the final stage of the competition. The final judging panel consisted exclusively of Google engineers, who selected the overall winners from the shortlisted projects.

During the preliminary judging round, Alberto evaluated several emerging ideas exploring new applications of multimodal AI, including:

  • AI systems that generate cinematic video narratives from user-provided images
  • Tools for dynamically generating alternative story outcomes for media content
  • AI-powered contract interpretation and legal analysis
  • Robotics-focused applications
  • Accessibility tools designed to support navigation for visually impaired users

These projects demonstrated how quickly multimodal AI capabilities are evolving and how developers are experimenting with entirely new categories of software built on top of these models.

Selected Finalist: Solus Forge

Based on the preliminary evaluation process, Solus Forge was selected as a finalist to advance to the final round of judging.

The project stood out for its strong combination of technical feasibility, innovation, and real-world applicability, and was among the top-scoring projects in Alberto’s evaluation.

Exploring the Next Generation of Multimodal AI

The hackathon encouraged teams to experiment with advanced multimodal AI technologies developed by Google DeepMind, including:

  • Gemini 3.1, featuring expanded long-context reasoning and native agentic vision capabilities
  • Lyria, DeepMind’s model for high-fidelity music and audio generation
  • NanoBanana 2, designed for advanced image composition, character consistency, and detailed text rendering

By building with these tools, participants explored new forms of multimodal applications that combine visual, audio, and language-based reasoning.

For industry practitioners like Alberto — who works closely with AI systems that extract and structure information from complex documents at Invofox — participating in events like this provides a valuable opportunity to evaluate emerging ideas and contribute expertise to the broader AI ecosystem.

Supporting Innovation in Applied AI

Hackathons like the YC x Google DeepMind Multimodal Frontier Hackathon serve as important environments for experimentation and collaboration between founders, engineers, and researchers.

By bringing together experts from leading organizations and early-stage builders, these events help accelerate the development of new AI applications and highlight the expanding possibilities of multimodal systems.

For Invofox, staying closely connected to these developments helps inform how next-generation AI capabilities may shape the future of document automation and data extraction technologies.

Thuy Vi Nguyen

Guest Contributor

Thuy Vi Nguyen is an Inbound Sales Marketing Specialist at Invofox, where she focuses on growth, demand generation, and go-to-market strategy. She has over a decade of experience in B2B SaaS across marketing, sales, and customer experience, and has led marketing initiatives for multiple technology companies.

Table of Contents

Subscribe to Our Blog

Subscribe for tips and insights from Invofox — the intelligent document processing (IDP) platform that helps businesses automate invoices, receipts, and more.

Ready to see it in action?

Utilizado por más de 150 empresas.

Invofox LinkedIn link
ISO 27001 certified document processing API ensuring information security managementSOC 2 compliant API audited by AICPA for secure and reliable service operationsHIPAA compliant document parsing API for handling healthcare data securelyHIPAA compliant document parsing API for handling healthcare data securely
Product Hunt widget - Invofox is the number 1 SaaS product of the week