If this is your first visit, be sure to
check out the FAQ by clicking the
link above. You may have to register
before you can post: click the register link above to proceed. To start viewing messages,
select the forum that you want to visit from the selection below.
I am not getting along with Google's build system: bazel, which is needed to compile this. Feels like somebody wanted a non-Java maven, and this got slapped together. The current Lyra source requires a specific checkout of clang and is Ubuntu-aligned for building. I'm going to go use podman with Ubuntu 20.04 on my Fedora 34 to try to play with this.
Google's definition of open source, means providing sources that doesn't build, and a binary they promise is build from it.
google reinventing the wheel again...
opus is great for this, no need to make a new one
At these tiny bitrates, OPUS is not very suited for speech. The question is, is there a reason to go that low since UDP/TCP overhead is on the same level as the lyra bitstream...
From the github project page:
Please note that there is a closed-source kernel used for math operations that is linked via a shared object called libsparse_inference.so. We provide the libsparse_inference.so library to be linked, but are unable to provide source for it. This is the reason that a specific toolchain/compiler is required.
Lyra harnesses the power of new natural-sounding generative models to maintain the low bitrate of parametric codecs while achieving high quality
For adhering to the principles of Open/Libre Source, they could also provide the source code and the assets of these "generative models" jointly with the tooling and the knowledge needed to build and improve them.
Also, the audio source samples used to validate the performance and accuracy of the codec are absolutely indispensable for the codec development.
Without these parts, this release consists of a partial source code release because one is only able to build and integrate it. Improving and constructively criticize it is practically denied to very talented people existing in the open-source community.
I hope that this situation would improve as the developers have more time to evolve their code, the math, and the assets that are not included in the Github repository.
At these tiny bitrates, OPUS is not very suited for speech. The question is, is there a reason to go that low since UDP/TCP overhead is on the same level as the lyra bitstream...
Good questions!
Opus already was developed as a "combo" from other 2 codecs:
Skype SILK codec for low bitrates and voice
Xiph.Org’s CELT codec for music and high-quality sound needs.
I know that Lyra is a different kind of codec based on speech synthesis instead of the traditional DCT/Quantization/Entropy Coding scheme. It is similar to Jean-Marc Valin's LPCNet codec.
What I'm thinking and wondering is:
If the machine learning models or the audio assets used to create and train Lyra couldn't be used to further improve Opus?
If a parametric codec like Lyra, WaveNet, or LPCNet couldn't be mixed in the Opus combo for achieving even lower bitrates?
Comment