780 points by lnyan 2 weeks ago | 72 comments
vipshek 2 weeks ago
I think the utility of generating vectors is far, far greater than all the raster generation that's been a big focus thus far (DALL-E, Midjourney, etc). Those efforts have been incredibly impressive, of course, but raster outputs are so much more difficult to work with. You're forced to "upscale" or "inpaint" the rasters using subsequent generative AI calls to actually iterate towards something useful.
By contrast, generated vectors are inherently scalable and easy to edit. These outputs in particular seem to be low-complexity, with each shape composed of as few points as possible. This is a boon for "human-in-the-loop" editing experiences.
When it comes to generative visuals, creating simplified representations is much harder (and, IMO, more valuable) than creating highly intricate, messy representations.
gwern 2 weeks ago
esperent 2 weeks ago
https://www.recraft.ai/ai-image-vectorizer
The quality does look quite amazing at first glance. How are the vectors to work with? Can you just open them in illustrator and start editing?
gwern 2 weeks ago
(The editing quality of the vectorized ones were not great, but it is hard to see how they could be good given their raster-style appearance. I can't speak to the editing quality of the native-generated ones, either in the old obsolete Recraft models or the newer ones, because the old ones were too ugly to want to use, and I haven't done much with the new one yet.)
brown_martin 1 week ago
gwern 1 week ago
brown_martin 1 week ago
Lerc 2 weeks ago
datadrivenangel 2 weeks ago
SillyUsername 2 weeks ago
zidad 1 week ago
spyder 1 week ago
janalsncm 2 weeks ago
I’d also like to see this in music generation. Tools like Suno are cool but I would much rather have something that generates MIDIs and instrument configurations instead.
Maybe this is a good lesson for generative tools. It’s possible to generate something that’s a good starting point. But what people actually want is long tail, so including the capability of precision modification is the difference between a canned demo and a powerful tool.
> Code coming soon
The examples are quite nice but I have no idea how reproducible they are.
kadushka 2 weeks ago
Sounds like you're looking for something like https://www.aiva.ai
janalsncm 2 weeks ago
I guess I’m hoping for something better. It’s also closed source, the web ui doesn’t have editing functionality, and the output is pretty disjointed. Maybe if I messed around with it enough the result would be decent.
kadushka 2 weeks ago
bufferoverflow 2 weeks ago
chaosprint 1 week ago
few days ago I was thinking about restarting this project with Glicol
https://github.com/chaosprint/RaveForce
RaveForce - An OpenAI Gym style toolkit for music generation experiments.
Love suno but eventually I need midi or xml or some lossless samples to work with
gexaha 2 weeks ago
scosman 2 weeks ago
Fun example: https://gist.github.com/scosman/701275e737331aaab6a2acf74a52...
astrodude 1 week ago
intalentive 2 weeks ago
Once you have your IR, modify and render. Once you have your render, apply a final coat of AI pixie dust.
Maybe generative models will get so powerful that fine-grained control can be achieved through natural language. But until then, this method would have the advantages of controllability, interoperability with existing tools (like Intellisense, image editors), and probably smaller, cheaper models that don’t have to accommodate high dimensional pixel space.