Users can generate videos up to 1080p resolution, up to 20 sec long, and in widescreen, vertical or square aspect ratios. You can bring your own assets to extend, remix, and blend, or generate entirely new content from text.

4 Followers

We’ve discovered neurons in CLIP that respond to the same concept whether presented literally, symbolically, or conceptually. This may explain CLIP’s accuracy in classifying surprising visual renditions of concepts, and is also an important step toward understanding the associations and biases that CLIP and similar models learn. Fifteen years ago, Quiroga et al.1 discovered that the human brain possesses multimodal neurons. These neurons respond to clusters of abstract concepts centered around a common high-level theme, rather than any specific visual feature. The most famous of these was the “Halle Berry” neuron, a neuron featured in both Scientific American⁠(opens in a new window) and The New York Times⁠(opens in a new window), that responds to photographs, sketches, and the text “Halle Berry” (but not other names). Two months ago, OpenAI announced CLIP⁠, a general-purpose vision system that matches the performance of a ResNet-50,2 but outperforms existing vision systems on some of the most challenging datasets. Each of these challenge datasets, ObjectNet, ImageNet Rendition, and ImageNet Sketch, stress tests the model’s robustness to not recognizing not just simple distortions or changes in lighting or pose, but also to complete abstraction and reconstruction—sketches, cartoons, and even statues of the objects. Now, we’re releasing our discovery of the presence of multimodal neurons in CLIP. One such neuron, for example, is a “Spider-Man” neuron (bearing a remarkable resemblance to the “Halle Berry” neuron) that responds to an image of a spider, an image of the text “spider,” and the comic book character “Spider-Man” either in costume or illustrated. Our discovery of multimodal neurons in CLIP gives us a clue as to what may be a common mechanism of both synthetic and natural vision systems—abstraction. We discover that the highest layers of CLIP organize images as a loose semantic collection of ideas, providing a simple explanation for both the model’s versatility and the representation’s compactness.

Solution Builders

4 Followers

Since 1995, we've been working with businesses just like yours providing Managed IT Services to solve typical IT and technology issues. We do this by using our proven and scalable process which aligns your technology needs and resources with our trusted outsourced IT support services resulting in a stable and scalable IT infrastructure for your business. Solutions Builders’ Managed IT Services are designed for companies and non-profits with between 5 and 200 employees who want to be free from the burden of managing their technology. We manage our client’s Information technology toward a predictable result for a fixed monthly fee allowing them to be experts at their business, not in managing technology. Our client’s results over the years speak volumes about our commitment to their success and as one of the longest lasting IT managed services companies in the Minneapolis and St. Paul areas, we are ready to serve our newest clients' needs just as well as we have served the old ones.

Corndog Blind

2 Followers

Simplifying what the eye doctor told you, (or didn't tell you). How to choose glasses so your face isn't 10x smaller and people can see around your head. Why you have to be dilated and or your pressure checked at every visit. What's the difference between an optometrist and an ophthalmologist and who you should schedule with and why. Helping you be an intelligent patient and give you some power in understanding how your eyes work. Some humor and a LOT of dorky antics because your host is a freakin NERD!