Vision, Language and Learning Lab
banner
vislang.ai
Vision, Language and Learning Lab
@vislang.ai
Vision, Language and Learning group at Rice University. Check our work at https://vislang.ai
Reposted by Vision, Language and Learning Lab
Can pretrained diffusion models be connected for cross-modal generation?

📢 Introducing AV-Link ♾️

Bridging unimodal diffusion models in one self-contained framework to enable:
📽️ ➡️ 🔊 Video-to-Audio generation.
🔊 ➡️ 📽️ Audio-to-Video generation.

🌐: snap-research.github.io/AVLink/

⤵️ Results
January 14, 2025 at 6:13 PM
Reposted by Vision, Language and Learning Lab
La Era de la Inteligencia Artificial a short documentary produced by Telemundo Houston won a Lone Star Emmy in the Science category www.telemundohouston.com/noticias/tec...
La Era de la Inteligencia Artificial
Nos adentramos a la revolución de la inteligencia artificial con la historia detrás de Ameca, el robot humanoide más avanzado del mundo
www.telemundohouston.com
November 26, 2024 at 3:23 AM
Reposted by Vision, Language and Learning Lab
Common software development advice that says we should write tests before implementations also applies to LLM-generated code.

PropTest: Automatic Property Testing for Improved Visual Programming.
arxiv.org/pdf/2403.16921
November 19, 2024 at 5:08 PM