Job Description
We're seeking an exceptional senior to staff level frontend engineer who can architect and build new ways of editing video with AI. We're not looking for specialists, rather driven problem solvers who can move seamlessly between crafting intuitive UIs and engineering browser capabilities to their limits. You'll work directly with the creators of Wav2lip, tackling challenges from real-time video processing to making tools people can't live without. Our creative suite is built with TypeScript and React, with NextJS and tRPC powering our API infrastructure. You'll own the development of our core video processing engine, bringing real-time editing capabilities directly to the browser through WebGL and WebAssembly, and maximally leveraging our models through our core AI platform. - Create fluid, responsive interfaces that make complex AI capabilities feel intuitive - Build high-performance video processing pipelines that run entirely in the browser - Ship features that define new standards for AI video creation - Work closely with research to rapidly prototype and productize new model capabilities - Drive improvements in frontend architecture and development practices - 5+ years building complex web applications, with deep React/TypeScript expertise - Strong background in video/media technologies (WebGL, Canvas, WebAssembly) - Track record of shipping 0-to-1 features that users love - Experience optimizing performance in demanding frontend applications - Ability to move fast, iterate from user feedback, and execute on tight deadlines - Obsession with craft and user experience - Strong team player who thrives in highly collaborative, fast-moving environments
We’re a team of artists, engineers, and researchers building controllable AI video editing tools to unbound human creative potential. Our research team build AI video models to understand and affect fine-grained, controllable edits over any human in any video. Our product team makes these models accessible to editors, animators, developers, and businesses to edit and repurpose any video for any audience. Our technology is used to automate lip-dubbing in localization processes in entertainment, create dynamic marketing campaigns personalized to individuals or communities, animate new characters to life in minutes instead of days, affect word-level edits in studio-grade videos to fix mistakes in post-production avoiding having to rerecord entire scenes, and more. Our models are used by everyday people, prosumers, developers, and businesses large and small to tell outstanding stories. In just the last year we graduated at the top of our YC batch (W24), raised a $5.5M seed backed by GV, won the AI grant from Nat Friedman and Daniel Gross, scaled to millions in revenue – and this is only the beginning.