Replies: 2 comments 2 replies
-
Just got merged! :) https://github.com/lucidrains/audiolm-pytorch/blob/main/audiolm_pytorch_demo.ipynb |
Beta Was this translation helpful? Give feedback.
-
Hi, However, you seem to have collected some valuable information: 10 days on a NVIDIA RTX A6000 and 23GB of data set is actually a number to go #1 and #2. (So it is not like Stable Diffusion where I throw 10 images in for half an hour and finish training!) But a training without annotation doesn't make sense. If I want to be able to use the Text Conditioning, I need annotated files so in the end I can use more than just 1 word for my 23GB of training data. The A.I. needs to know which words correspond to which sound files, or it can only produce unguided recompilations. All I want to say is that, the examples are awesome, but still hard to follow if you are new, maybe @lucidrains reads this discussion and adds the things! |
Beta Was this translation helpful? Give feedback.
-
Hello,
is it possible to add to the readme an example on how to use this? A line like:
Beta Was this translation helpful? Give feedback.
All reactions