Bilawal Sidhu
banner
bilawal.bsky.social
Bilawal Sidhu
@bilawal.bsky.social
🪄 Blending Realities | 🎙️ Host, TED AI Show | 🚀 Scout, A16z | 🎬 1.4M+ Subs & 450M+ Views | 🌎 Ex-Google PM, 3D Maps & AR/VR 🥽 https://spatialintelligence.ai https://bilawal.ai
Eric Schmidt got a standing ovation from the TED audience this morning.

An absolute pleasure to interview him on the red circle.

We dove into the big questions—superintelligence, national strategy, open source, and what it means to be human in the age of AI.

One for the books.
April 12, 2025 at 12:52 AM
TikTok ban imminent, yet funny how things change.

>2020: Stressed about TikTok drama at 120K subs.

>2024: Sitting at 994K and completely unfazed.

Ban it? Cool, I’ll build elsewhere. Keep it? Roger that, I’ll double down.

The game is bigger than any one app. Who cares about vanity metrics.
January 14, 2025 at 2:47 AM
Merry Christmas y’all! 🎄

Pictured: 3d scan vs. ground truth of the feast to follow
December 25, 2024 at 9:24 PM
Omnidirectional 3D video of reality — damn near teleportation in a VR headset.

This $17,000 VR camera released in 2017 was ahead of its time. 17 cameras → cloud stitching → 8K x 8K stereo VR video.

The moment is ripe for a new 4d capture rig optimized for dynamic 3d gaussians. Anyone building one?
December 16, 2024 at 12:15 AM
And given we're dealing with real stereoscopic content, results are notably better than synthetic data, giving you a faithful rendition of the real-world with a diverse set of subject matter.
December 15, 2024 at 2:29 PM
They're using it to train this model called DynaDUSt3R that can predict both 3D structure and motion from video frames. Which means it tracks how objects move between frames while simultaneously reconstructing their 3D shape.
December 15, 2024 at 2:29 PM
It was always clear that stereo datasets would be valuable -- and we launched some cool VR tools with it back in 2017 (link below). But the game changer now in 2024 is the scale -- they're providing 110K clips :-) That's the kind of massive, real-world dataset that was just a dream in those days!
December 15, 2024 at 2:29 PM
Check out this Stereo4D paper from Google DeepMind. It's a pretty clever approach to a persistent problem in computer vision -- getting good training data for how things move in 3D. The key insight is using VR180 videos -- those stereo fisheye videos we launched back in 2017 for YouTubeVR 🧵
December 15, 2024 at 2:29 PM
The future isn't just virtual or augmented – it's ambient and intelligent

The Google XR unlocked event in NYC
December 14, 2024 at 7:36 PM
MKBHD dropped his OpenAI Sora review (after a week of testing) the much hyped AI video model.

5 immediate observations:
December 9, 2024 at 4:52 PM
Wav2lip can FINALLY rest in peace. Being able to retarget the facial performance of characters in *existing* live action & CG video makes Act-One an extremely useful tool for all types of creators.

Nicely done RunwayML!
December 6, 2024 at 4:25 PM
Imagine making 2D concept art for a game world –pressing a button – and suddenly you can walk around an interactive 3D world. That's what Google DeepMind's new paper Genie 2 can do – simulate virtual worlds, including the consequences of any action (e.g. unlock door, jump, swim etc).
December 4, 2024 at 5:07 PM
Tencent’s open weights Hunyuan Video 13B model looks impressive — oh, and image-to-video and facial performance? They’re coming too.

If 2024 was the year open-source LLMs caught up with closed-source AI — 2025 will be the year open-source video catches up.
December 3, 2024 at 4:58 PM
World Labs first demo dropped, and it’s consistent 3D worlds from a single 2D image.

Decent volume size to move around in — def a big step up from the RGB + depth 360 environments we’re used to e.g. Blockade Labs.

Stylized results look good; i’d love to see more photorealistic AI generations!
December 2, 2024 at 5:21 PM
Augmented reality x-ray vision to “see through” concrete.

Your infrastructure won’t just be scanned — it’ll be anchored to reality.

Demo: Pix4D reality capture with precise geospatial localization.
November 29, 2024 at 12:54 AM
> Sora API leaks on HF
> Yep, it’s real
> New Turbo variant; way faster
> Artists declare war: “Unpaid R&D, PR puppets”
> OAI Discord goes nuclear—topic banned
> All before my first coffee
November 26, 2024 at 5:41 PM
When making city-scale 3D models radiance fields are cool and all — but tried & tested photogrammetry just hits different. Microsoft Flight Simulator 2024 is a damn good case in point. Gosh I love geospatial 3D imagery — especially in virtual reality 😍

Video demo: u/lunchanddinner
November 24, 2024 at 2:22 PM
Just read a neat AI paper called SAMURAI -- it takes SAM 2 (Meta's "segment anything" model) and makes it way better at tracking objects in videos.

Basic problem is SAM 2 gets confused when things move fast or there's a crowd of similar objects (big problem for VFX and video intelligence alike).
November 23, 2024 at 4:00 PM
Ironic that OpenAI’s o1 model hides its chain-of-thought reasoning, while the Chinese DeepSeek-R1 makes it transparent to users. Shouldn’t it be the other way around?
November 21, 2024 at 5:43 PM
If 10% of engineers at top tech companies are “ghost engineers” you have to wonder — what’s that ratio in bureaucracies around the world?

It’s likely flipped — only 20% of people doing the real work. Pareto principle strikes again.
November 21, 2024 at 3:57 PM
Happy GIS day! Linked up with Array Labs -- they're a YC startup using satellite sensors to build a near real-time 3d map of the world.

Here's a peek into a beta reconstruction of San Francisco made entirely from satellite imagery. For the initiated, their goal is essentially Vricon 3D on steroids.
November 21, 2024 at 1:31 AM
Mixing photogrammetry, procedural 3D maps and BIM models is a helluva combo.

Imagine a world where every city has a digital twin like this -- letting us model not just buildings, but every piece of infrastructure underneath. The ultimate mirror world.
November 20, 2024 at 3:30 AM
The blueprint
April 28, 2023 at 4:26 AM