The model is fast, producing a 2.25-megapixel depth map in 0.3 seconds on a standard GPU.
https://github.com/apple/ml-depth-pro
https://arxiv.org/pdf/2410.02073
The model is fast, producing a 2.25-megapixel depth map in 0.3 seconds on a standard GPU.
https://github.com/apple/ml-depth-pro
https://arxiv.org/pdf/2410.02073
In the next couple of decades, we will be able to do things that would have seemed like magic to our grandparents.
This phenomenon is not new, but it will be newly accelerated. People have become dramatically more capable over time; we can already accomplish things now that our predecessors would have believed to be impossible.
We are more capable not because of genetic change, but because we benefit from the infrastructure of society being way smarter and more capable than any one of us; in an important sense, society itself is a form of advanced intelligence. Our grandparents – and the generations that came before them – built and achieved great things. They contributed to the scaffolding of human progress that we all benefit from. AI will give people tools to solve hard problems and help us add new struts to that scaffolding that we couldn’t have figured out on our own. The story of progress will continue, and our children will be able to do things we can’t.
https://www.bbc.com/future/article/20240912-what-riddles-teach-us-about-the-human-mind
“As human beings, it’s very easy for us to have common sense, and apply it at the right time and adapt it to new problems,” says Ilievski, who describes his branch of computer science as “common sense AI”. But right now, AI has a “general lack of grounding in the world”, which makes that kind of basic, flexible reasoning a struggle.
AI excels at pattern recognition, “but it tends to be worse than humans at questions that require more abstract thinking”, says Xaq Pitkow, an associate professor at Carnegie Mellon University in the US, who studies the intersection of AI and neuroscience. In many cases, though, it depends on the problem.
A bizarre truth about AI is we have no idea how it works. The same is true about the brain.
That’s why the best systems may come from a combination of AI and human work; we can play to the machine’s strengths, Ilievski says.
The goal is to reduce costs by replacing traditional storyboard artists and VFX crews with AI-generated “cinematic video.” Lionsgate hopes to use this technology for both pre- and post-production processes. While the company promotes the cost-saving potential, the creative community has raised concerns, as Runway is currently facing a lawsuit over copyright infringement.
https://depthcrafter.github.io/
We innovate DepthCrafter, a novel video depth estimation approach, by leveraging video diffusion models. It can generate temporally consistent long depth sequences with fine-grained details for open-world videos, without requiring additional information such as camera poses or optical flow.
https://loopyavatar.github.io/
Loopy supports various visual and audio styles. It can generate vivid motion details from audio alone, such as non-speech movements like sighing, emotion-driven eyebrow and eye movements, and natural head movements.
Built on top of a series of fine-tunes of Stable Diffusion, Flux, and Magnific-built models.
https://www.freepik.com/ai/image-generator
https://www.anthropic.com/news/artifacts
Artifacts turn conversations with Claude into a more creative and collaborative experience.
It connects Nuke with the ComfyUI server, any plugin that comes out in ComfyUI can be used in nuke, rotos with sam, rescaling, image generation, inpaintins, normal generator, the nodes are IPAdapter, ControlNet, AnimateDiff, Flux etc.
https://github.com/vinavfx/nuke_comfyui
https://www.rankred.com/convert-2d-images-to-3d/