IDK, I was playing with Claude yesterday/this morning and before I hit the free tier context limit it managed to create a speech-to-phoneme VQ-VAE contraption with a sliding window for longer audio clips and some sort of "attention to capture relationships between neighboring windows" that I don't quite understand. That last part was due to a suggestion it provided where I was like "umm, ok..."
Seems pretty useful to me where I've read a bunch of papers on different variational autoencoder but never spent the time to learn the torch API or how to set up a project on the google.
In fact, it was so useful I was looking into paying for a subscription as I have a bunch of half-finished projects that could use some love.
Seems pretty useful to me where I've read a bunch of papers on different variational autoencoder but never spent the time to learn the torch API or how to set up a project on the google.
In fact, it was so useful I was looking into paying for a subscription as I have a bunch of half-finished projects that could use some love.