We’ve been playing with Alibaba’s WAN2.1 text-to-video model lately. Like most image and video generation models, Wan has a lot of input parameters, and each of them can have a profound impact on the quality of the generated output. What happens when you tweak those mysterious inputs? Let’s find out. The experiment We wanted to

Something went wrong, but don’t fret — let’s g

[Submitted on 14 Oct 2022] Download PDF Abstract: With the ever-growing size of pre-trained models (PMs), fine-tuning them has become more expensive and resource-hungry. As a remedy, low-rank adapters (LoRA) keep the main pre-trained weights of the model frozen and just introduce some learnable truncated SVD modules (so-called LoRA blocks) to the model. While LoRA