Atom_101
Atom_101 t1_j3voh2m wrote
Checkout tortoise-tts
Atom_101 OP t1_j3b2vmr wrote
Reply to comment by mkthabet in [D] Best way to package Pytorch models as a standalone application by Atom_101
From what I read it doesn't package cuda, only python dependencies. Were you able to get cuda inside your pyinstaller executable? Also is it even possible to package cuda inside an executable? Cuda needs to go into a specific folder right? And it needs to be added to the path variable for pytorch or other libraries to see it. For example in Linux it goes inside /usr/local/cuda
Atom_101 OP t1_j3b2hnc wrote
Reply to comment by nins_ in [D] Best way to package Pytorch models as a standalone application by Atom_101
Thanks. This might be more useful for someone doing mobile or edge deployment since ncnn seems to be cpu only. I need my model to run on GPUs.
Atom_101 OP t1_j3b291o wrote
Reply to comment by dr-pork in [D] Best way to package Pytorch models as a standalone application by Atom_101
It seems you can't lock a container. If the end user has root access they will be able to ssh into the container and see your source code. The solution seems to be to obfuscate your code using something like pyarmor, so that even if the user accesses the docker image, they won't easily figure out your source code.
Atom_101 OP t1_j3b20ok wrote
Reply to comment by CyberDainz in [D] Best way to package Pytorch models as a standalone application by Atom_101
So you are manually setting up cuda and windows environment variables in that script? I'll see if I can get it to work like that. Thanks!
Atom_101 OP t1_j384ogh wrote
Reply to comment by robertknight2 in [D] Best way to package Pytorch models as a standalone application by Atom_101
I haven't used onnx before but have worked with torchscript. With torchscript I have had to change the models quite a bit with to make it scriptable. If onnx requires similar amount of effort I don't think it will be useful.
I don't want to go through the hassle of scripting because we might change the model architectures soon. I need a quick and possibly inefficient (space wise, not perf wise) way to package the models without exposing source code.
Atom_101 t1_iy2pta1 wrote
Unsupervised contrastive learning is sort of like dimensionality reduction. So something like a CLIP Huge can be considered sota I guess.
Atom_101 t1_it1fzij wrote
Reply to [Research] Scholars Program by ml_magic_
Where can I find Cohere's publications? Basically a way to get an idea of the things you work on.
>In fact, our one criterion for selection is that you cannot have published a machine learning research paper previously.
This is from the blog. Is this criterion strictly enforced. What if someone has publications but no first author publications? What if they have published in lower tier conferences? Why exclude such people, since many of them will be interested in the program.
Atom_101 t1_is7ldte wrote
Reply to comment by Quaxi_ in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
I think VAEs are weak not because of scaling issues but , because of an overly strong bias that the latent manifold has to be a Gaussian distribution with a diagonal covariance matrix. This problem is reduced using things like variational quantization. Dalle-1 actually used this, before DMs came to be. But even then, I believe they are too underpowered. Another technique of image generation is normalising flows which also require heavy restrictions on model architecture. GANs and DMs are much more unrestricted and can model arbitrary data distributions.
Can you point to an example where you see GANs perform visibly worse? Although we can't really compare quality between sota GANs and sota DMs. The difference in scale is just too huge. There was a tweet thread recently, regarding Google imagen iirc, which showed that increasing model size drastically improves image quality for text-to-image DMs. Going from 1B to 10B params showed visible improvements. But if you compare photorealistic faces generated by stable diffusion and say stylegan3, I am not sure you would be able to see differences.
Atom_101 t1_is6wyys wrote
Reply to comment by pm_me_your_ensembles in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
I see. Thanks!
Atom_101 t1_is6igij wrote
Reply to comment by SleekEagle in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
Just when I thought I understand the math behind DMs, they went ahead and added freaking DEs to it? Guess I should have paid more attention to math in college.
Atom_101 t1_is61h1l wrote
Reply to comment by Quaxi_ in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
I doubt it's anywhere close to diffusion models though. Haven't worked with ttur and feature matching. But have tried spectral norm and wgan+gp. They can be unstable in weird ways. In fact, while wasserstein loss is definitely more stable, it massively slows down convergence compared to standard dcgan loss.
The biggan paper by Google tried to scale up GANs by throwing every known stabilization trick at them. They observed that even with these tricks you can't train beyond a point. BigGANs start degrading when trained too much. Granted it came out in 2018, but if this didn't hold true today we would have 100B parameter GANs already. I think the main advantage with DMs is that you can keep training them for an eternity without worrying about performance degradation.
Atom_101 t1_is5z87v wrote
Reply to comment by pm_me_your_ensembles in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
How so?
Atom_101 t1_is4b95t wrote
Reply to comment by M4xM9450 in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
Diffusion is inherently slower than GANs. It takes N forward passes vs only 1 for GANs. You can use tricks to make it faster, like latent diffusion which does N forward passes with a small part of the model and 1 forward pass with the rest. But as a method diffusion is slower.
Atom_101 t1_j3vwoid wrote
Reply to comment by CeFurkan in [D] Any model like VALL-E available currently? by CeFurkan
Yeah. It supports zero shot voice cloning using a reference clip.