https://github.com/huntorochi/HoGS
3Dprinting (175) A.I. (751) animation (339) blender (197) colour (229) commercials (49) composition (152) cool (360) design (636) Featured (69) hardware (308) IOS (109) jokes (134) lighting (282) modeling (127) music (186) photogrammetry (177) photography (751) production (1252) python (87) quotes (490) reference (310) software (1336) trailers (296) ves (538) VR (219)
https://gapszju.github.io/RGBAvatar
A method for reconstructing photorealistic, animatable head avatars at speeds sufficient for on-the-fly reconstruction. Unlike prior approaches that utilize linear bases from 3D morphable models (3DMM) to model Gaussian blendshapes, our method maps tracked 3DMM parameters into reduced blendshape weights with an MLP, leading to a compact set of blendshape bases.
https://github.com/gapszju/RGBAvatar
For decades, LiDAR and 3D sensing systems have relied on mechanical mirrors and bulky optics to direct light and measure distance. But at CES 2025, Lumotive unveiled a breakthrough—a semiconductor-based programmable optic that removes the need for moving parts altogether.
LiDAR and 3D sensing systems work by sending out light and measuring when it returns, creating a precise depth map of the environment. However, traditional systems have relied on physically moving mirrors and lenses, which introduce several limitations:
To bring high-resolution depth sensing to wearables, smart devices, and autonomous systems, a new approach is needed.
Lumotive’s Light Control Metasurface (LCM) replaces mechanical mirrors with a semiconductor-based optical chip. This allows LiDAR and 3D sensing systems to steer light electronically, just like a processor manages data. The advantages are game-changing:
LCM technology works by controlling how light is directed using programmable metasurfaces. Unlike traditional optics that require physical movement, Lumotive’s approach enables light to be redirected with software-controlled precision.
This means:
At CES 2025, Lumotive showcased how their LCM-enabled sensor can scan a room in real time, creating an instant 3D point cloud. Unlike traditional LiDAR, which has a fixed scan pattern, this system can dynamically adjust to track people, objects, and even gestures on the fly.
This is a huge leap forward for AI-powered perception systems, allowing cameras and sensors to interpret their environment more intelligently than ever before.
Lumotive’s programmable optics have the potential to disrupt multiple industries, including:
Lumotive’s Light Control Metasurface represents a fundamental shift in how we think about optics and 3D sensing. By bringing programmability to light steering, it opens up new possibilities for faster, smarter, and more efficient depth-sensing technologies.
With traditional LiDAR now facing a serious challenge, the question is: Who will be the first to integrate programmable optics into their designs?
https://tobias-kirschstein.github.io/avat3r
Avat3r takes 4 input images of a person’s face and generates an animatable 3D head avatar in a single forward pass. The resulting 3D head representation can be animated at interactive rates. The entire creation process of the 3D avatar, from taking 4 smartphone pictures to the final result, can be executed within minutes.
https://www.uploadvr.com/meta-researchers-generate-photorealistic-avatars-from-just-four-selfies
https://arcanous98.github.io/projectPages/gaussianVolumes.html
We propose a compact and efficient alternative to existing volumetric representations for rendering such as voxel grids.
SPAR3D is a fast single-image 3D reconstructor with intermediate point cloud generation, which allows for interactive user edits and achieves state-of-the-art performance.
https://github.com/Stability-AI/stable-point-aware-3d
https://stability.ai/news/stable-point-aware-3d?utm_source=x&utm_medium=social&utm_campaign=SPAR3D
https://sonsang.github.io/dmesh2-project
An efficient differentiable mesh-based method that can effectively handle complex 2D and 3D shapes. For instance, it can be used for reconstructing complex shapes from point clouds and multi-view images.
https://radiancefields.com/gaussian-splatting-in-nuke
https://aescripts.com/gaussian-splatting-for-nuke
Odyssey and a Pixar co-founder, Ed Catmull (also an investor), just dropped Explorer, a revolutionary 3D world generator that turns any image into an editable 3D world.
https://odyssey.systems/introducing-explorer
https://jixuan-fan.github.io/Momentum-GS_Page
https://github.com/Jixuan-Fan/Momentum-GS
A novel approach that leverages momentum-based self-distillation to promote consistency and accuracy across the blocks while decoupling the number of blocks from the physical GPU count.
A feed-forward model capable of reconstructing 3D scenes parameterized by 3D Gaussians from unposed sparse multi-view images.
https://scaniverse.com/news/spz-gaussian-splat-open-source-file-format
https://github.com/nianticlabs/spz
• Slashes file sizes by 90% (250MB → 25MB) with virtually zero quality loss
• Lightning-fast uploads/downloads, especially on mobile
• Dramatically reduced memory footprint
• Enables real-time processing right on your phone
Tech breakthrough:
• Smart compression of position, rotation, color & scale data
• Column-based organization for maximum efficiency
• Innovative fixed-point quantization & log encoding
https://www.8thwall.com/products/niantic-studio
COLLECTIONS
| Featured AI
| Design And Composition
| Explore posts
POPULAR SEARCHES
unreal | pipeline | virtual production | free | learn | photoshop | 360 | macro | google | nvidia | resolution | open source | hdri | real-time | photography basics | nuke
FEATURED POSTS
Social Links
DISCLAIMER – Links and images on this website may be protected by the respective owners’ copyright. All data submitted by users through this site shall be treated as freely available to share.