Gilberto Ochoa-Ruiz
banner
gochoa-ruiz.bsky.social
Gilberto Ochoa-Ruiz
@gochoa-ruiz.bsky.social
🇲🇽 Researcher/Assist. Prof. Tec de Monterrey | CV-INSIDE Lab head | Board SMIA MX | Chair LatinX in AI | Sponsor Chair MIUA 2025 (Leeds 🇬🇧 ) | 🇪🇺 EU Alumni | 🏳️‍🌈
I run the small, mobile and independent research CV-inside lab: https://beton-ochoa.github.io
Reposted by Gilberto Ochoa-Ruiz
This is a fantastic oral history of the last 10 years of NLP and AI. www.quantamagazine.org/when-chatgpt...
When ChatGPT Broke an Entire Field: An Oral History | Quanta Magazine
Researchers in “natural language processing” tried to tame human language. Then came the transformer.
www.quantamagazine.org
May 1, 2025 at 11:55 AM
Reposted by Gilberto Ochoa-Ruiz
Have you ever wondered what makes two models different?
We all know the ViT-Large performs better than the Resnet-50, but what visual concepts drive this difference? Our new ICLR 2025 paper addresses this question! nkondapa.github.io/rsvc-page/
April 11, 2025 at 4:11 PM
Reposted by Gilberto Ochoa-Ruiz
[1/10] Is scene understanding solved?

Models today can label pixels and detect objects with high accuracy. But does that mean they truly understand scenes?

Super excited to share our new paper and a new task in computer vision: Visual Jenga!

📄 arxiv.org/abs/2503.21770
🔗 visualjenga.github.io
March 29, 2025 at 7:36 PM
Reposted by Gilberto Ochoa-Ruiz
“Inside arXiv—the Most Transformative Platform in All of Science”

www.wired.com/story/inside...
March 27, 2025 at 10:59 PM
Reposted by Gilberto Ochoa-Ruiz
Same here… not going back. So currently just hanging around and hopeful that my students will catch any fancy trends … btw … interesting paper arxiv.org/abs/2503.20680 🙃
Vision as LoRA
We introduce Vision as LoRA (VoRA), a novel paradigm for transforming an LLM into an MLLM. Unlike prevalent MLLM architectures that rely on external vision modules for vision encoding, VoRA internaliz...
arxiv.org
March 27, 2025 at 6:23 PM
Reposted by Gilberto Ochoa-Ruiz
My lab works hard to do human participant evaluations of our Explainable RL systems. I’m rather proud of experiments we ran in this paper: arxiv.org/abs/2210.04723 (coming soon as a journal publication)
Experiential Explanations for Reinforcement Learning
Reinforcement Learning (RL) systems can be complex and non-interpretable, making it challenging for non-AI experts to understand or intervene in their decisions. This is due in part to the sequential ...
arxiv.org
February 8, 2025 at 12:24 AM
Reposted by Gilberto Ochoa-Ruiz
Scalable Benchmarking and Robust Learning for Noise-Free Ego-Motion and 3D Reconstruction from Noisy Video

Xiaohao Xu, @tianyizhang.bsky.social, Shibo Zhao, Xiang Li, Sibo Wang, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-Roberson, Sebastian Scherer, Xiaonan Huang

arxiv.org/abs/2501.14319
January 30, 2025 at 3:42 AM
We are looking for motivated PhD students to join my group CVINSIDE lab!
January 29, 2025 at 3:28 PM
We are super happy as our special track "Trustworthy AI for Computer Assisted Diagnosis and Intervention" has been accepted for the 2025 edition of MIUA or the Conference on Medical Image Understanding and Analysis! 😊
January 28, 2025 at 4:09 AM
Reposted by Gilberto Ochoa-Ruiz
Yiqing Liang, Abhishek Badki, Hang Su, James Tompkin, Orazio Gallo
Zero-Shot Monocular Scene Flow Estimation in the Wild
https://arxiv.org/abs/2501.10357
January 20, 2025 at 5:17 AM
Working from this nice little corner today 😁
January 15, 2025 at 6:03 PM
Reposted by Gilberto Ochoa-Ruiz
A generalist #AI medical LLM model that improves accuracy of physician diagnosis across 8 specialities for common and rare diseases
nature.com/articles/s41...
A generalist medical language model for disease diagnosis assistance - Nature Medicine
Trained on a large corpus of medical text and patient records and tested across diseases, with specific focus on rare presentations, an open-source medical language model demonstrates higher accuracy ...
nature.com
January 8, 2025 at 2:48 PM
Reposted by Gilberto Ochoa-Ruiz
Teaching my #DeepLearning in #ComputerVision course next semester. Looking for blind spots in my content. Share your favourite deep learning course in the thread.

Course: www.eecs.yorku.ca/~kosta/Cours...
December 23, 2024 at 12:33 AM
Reposted by Gilberto Ochoa-Ruiz
#roboticsurgery #ai #viral
#fyp #robotics #robots #foryou #airobots
Robotic-assisted spine surgery is a minimally invasive surgical technique that uses a robotic arm to assist the surgeon in performing the delicate tasks involved in spinal fusion surgery.
www.online-sciences.com/robotics/use...
Use of robotics in minimally invasive spine surgery and benefits from robotic assistance - Science online
Robotic-assisted spine surgery can perform spine fixation procedures. It can be used to treat a variety of spinal conditions, including degenerative disc disease, spinal stenosis, scoliosis, spondylol...
www.online-sciences.com
May 24, 2024 at 3:45 AM
Reposted by Gilberto Ochoa-Ruiz
MegaSynth: Scaling Up 3D Scene Reconstruction with Synthesized Data

Hanwen Jiang, Zexiang Xu, @desaixie.bsky.social, Ziwen Chen, @haian-jin.bsky.social, Fujun Luan, Zhixin Shu, Kai Zhang, Sai Bi, Xin Sun, Jiuxiang Gu, Qixing Huang, Georgios Pavlakos, Hao Tan

arxiv.org/abs/2412.14166
December 19, 2024 at 6:43 AM
Reposted by Gilberto Ochoa-Ruiz
🚨 Call for papers still open! 🚨

Special Issue on Innovations in Artificial Intelligence for Medicine and Healthcare

Title: Innovations in Artificial Intelligence for Medicine and Healthcare

Journal Health Information Science and Systems (Springer)
Submit your research!
December 7, 2024 at 4:14 PM
Reposted by Gilberto Ochoa-Ruiz
Your figures when using raster rather than vector graphics 🫣
December 18, 2024 at 5:18 PM
Reposted by Gilberto Ochoa-Ruiz
I was waiting for this one! Uncalibrated SLAM with changing focal length, very nice results :)
Introducing MASt3R-SLAM, the first real-time monocular dense SLAM with MASt3R as a foundation.

Easy to use like DUSt3R/MASt3R, from an uncalibrated RGB video it recovers accurate, globally consistent poses & a dense map.

With @ericdexheimer.bsky.social* @ajdavison.bsky.social (*Equal Contribution)
December 16, 2024 at 6:07 PM
Reposted by Gilberto Ochoa-Ruiz
I'm doing another AI Art Gallery with @cvprconference.bsky.social in Nashville this year 😍🤖🎨

Submissions for artworks using / about computer vision are open until 9th March 2025

More details: bit.ly/CVPRAIArt25

#CVPR2025 #creativeAI #AIart
December 13, 2024 at 3:28 PM
Reposted by Gilberto Ochoa-Ruiz
#MICCAI2024 participants will have received an invite to submit their feedback on how things went.

We'd love to hear from you on what we did well, and what we could have done even better - this will inform next year's chairs.

Organising #MICCAI2024 came with challenges but also great rewards!
December 11, 2024 at 7:45 PM
Reposted by Gilberto Ochoa-Ruiz
A very necessary systematic review of the use of large language models in healthcare.

jamanetwork.com/journals/jam...
Testing and Evaluation of Health Care Applications of Large Language Models
This systematic review characterizes the current performance of large language models in evaluating clinical health care settings, including uniformity, thoroughness, and robustness and proposes a fra...
jamanetwork.com
December 10, 2024 at 9:38 PM
🚨 Call for papers still open! 🚨

Special Issue on Innovations in Artificial Intelligence for Medicine and Healthcare

Title: Innovations in Artificial Intelligence for Medicine and Healthcare

Journal Health Information Science and Systems (Springer)
Submit your research!
December 7, 2024 at 4:14 PM
Reposted by Gilberto Ochoa-Ruiz
Introducing MegaSaM!

Accurate, fast, & robust structure + camera estimation from casual monocular videos of dynamic scenes!

MegaSaM outputs camera parameters and consistent video depth, scaling to long videos with unconstrained camera paths and complex scene dynamics!
December 6, 2024 at 5:43 PM
It seems that I’m going to Seattle in January (my 3rd visit) At least i was able to get some rest during December (I wanted to travel a bit less in 2025 😂)
December 6, 2024 at 5:21 AM
Reposted by Gilberto Ochoa-Ruiz
LLM-based agentic systems in medicine and healthcare - Nature Medicine Intelligence by Dr. Eric Topol @erictopol.bsky.social and team.

www.nature.com/articles/s42...
LLM-based agentic systems in medicine and healthcare - Nature Machine Intelligence
Large language model-based agentic systems can process input information, plan and decide, recall and reflect, interact and collaborate, leverage various tools and act. This opens up a wealth of oppor...
www.nature.com
December 5, 2024 at 8:01 PM