Odd that the page doesn't seem to link to either,
LLaSA is a simple framework for speech synthesis that employs a single-layer vector quantizer (VQ) codec and a single Transformer architecture to fully align with standard LLMs such as LLaMA.
I can't wait see this integrated into Open WebUI! These sound amazing.
the long 'uuuuhhhhhhh' from some of the lesser models is killing me.
> employs a single-layer vector quantizer (VQ) codec and a single Transformer architecture to fully align
I really wish when new models were released that they would draw a diagram of all the layers and the tensor input and output sizes at each layer, with zoom in/out capabilities if needed using D3.js or whatever visualization framework if needed. Every single layer should be on there with its input and output sizes.
These one-sentence descriptions, and approximate block diagrams with arrows pointing at each other are never enough to understand how something is actually implemented.