@griffinboy some guides on this kind of stuff would be amazing. I'd love to have a crack at some of this too. When you have the time it would be awesome.
Best posts made by ccbl
-
RE: 3rd party Header Files in Hise?
-
RE: Help me understand Waveshaping
@clevername27 I appreciate your sentiment, but I'm just a guy with a normal day job with an interest in plugins. Not looking to make any money here, just releasing my work for free as and when inspiration strikes.
-
RE: Help me understand Waveshaping
@griffinboy ha ha yeah. To be honest I wasn't expecting to make some UAD level simulation but just learn the basics of HISE a bit more.
I'm already very familiar with AI though through making a lot of NAM profiles and being one of the co-founders of ToneHunt.
Really looking forward to implementing quite a few ideas once that process is ironed out.
-
RE: Simple ML neural network
@griffinboy Here's the link to their discord channel for RTNeural
-
RE: Simple ML neural network
@resonant that's awesome. Would love to pick their brains and see if we can get it up and running for the rest of us.
-
RE: Simple ML neural network
@ccbl for instance, how I plan to use HISE is to create plugins where I use a NN to model various non-linear components such as transformers, tubes, fet preamps etc, and then use the regular DSP in between. I'm just a hobbiest who plans to release everything FOSS though, so I'll have to wait and see what you much more clever folks come up with.
-
RE: Simple ML neural network
I realise we already hashed this discussion out, and people might be sick of it. But IMO the NAM trainer has a really intuitive GUI trainer which allows for different sized networks, at various sample rates. It also has a very defined output model format, which seems to be a sticking point with RTNeural.
Given the existence of the core C++ library https://github.com/sdatkinson/NeuralAmpModelerCore
Might it be easier to implement this instead, given many people want to use ML Networks for non-linnear processing for the most part?
-
RE: Simple ML neural network
@Christoph-Hart so for now I decided to try and make a very simple audio processor using scriptnode. I followed the tutorial project adding the weights for the math.neural module to pickup, and it does see the object, however when I select it HISE crashes, complete CTD.
It's an LSTM network generated by a RTNeural based project (guitarML automatedguitarampmodelling pipline).
Not sure what the issue is.
-
RE: Simple ML neural network
@Christoph-Hart said in Simple ML neural network:
@ccbl And can't you just convert the models to work in RTNeural? In the end it's just running maths and I'm not super interested in adding the same thing but in 5 variations.
I understand why you feel that way. There's no point chasing your tail every time a new NN comes on the scene. What I would say though is that since NAM has been released, I've seen at least 4 different companies bring out their own NN capture tech. And not a single one of those has taken off, all combined they have less captures publicly available than one comprehensive NAM pack on ToneHunt .
Multiple commercial companies have incorporated NAM including Melda, Amplifire, Two Notes, Audio Assualt, and Solemn Tones amongst others. AnalogueX now uses NAM to capture preamps and compressors. Past of Future also does this.
Essentially NAM has become the industry standard format for Analogue gear capture using neural networks at this point. So personally I think it would be worth implementing the NAM core tech as there's a far higher chance people are going to want to use a NAM model than any of the other tech.
Beggars can't be choosers though. I don't mean this to come across as a demand so much as trying to present a compelling case. I don't have the skills to implement this myself. I understand there's only some much time and will power a person has to dedicate to a project like this.
So I'm highly grateful for any way to incorporate NAM into the signal processing using HISE. Whether that involves conversion or native processing.
-
RE: Simple ML neural network
@Christoph-Hart A few things I guess. The number of NAM captures currently dwarfs all the other that are supported by RTNeural (I have used guitar ML in the past for instance) and it's only growing (here for example https://tonehunt.org/). Not just 1000s of amp snap shots but people are really getting into studio gear captures too. There's a huge group of people for support in generating good captures and technical training support. And of course it is probably the best sounding in terms of accuracy right now.
I'm interested enough in using a neural net that I'm willing to use RTNeural, it's still a great system. NAM is becoming a defacto standard in a lot of NN capture spaces currently though. So for the future it seems like a good addition to the code base. And on a personal level I've already created over 1000 NAM models.
Maybe once I've learned more of the basics, if someone is willing to help me with it I would appreciate it.
Latest posts made by ccbl
-
Soft Bypass CPU Usage
Pretty simple question I guess, when using SoftBypass, do the blocks contained within stop using CPU power?
For instance, if I used on of the bypass blocks that can switch between multiple signal chains, are the unused chain using CPU power?
-
RE: Hardcoded Neural Network does not work as expected
@Dan-Korneff I followed all the steps in your Gitlab to get the Aida-X trainer up and running, but when I get to the actual training part, it reads all the configs and starts the training process but then fails with
"RuntimeError: cuDNN error: CUDNN_STATUS_NOT_SUPPORTED. This error may appear if you passed in a non-contiguous input.
I tried with both 24bit and 32bit float input files.
-
RE: Hardcoded Neural Network does not work as expected
@Dan-Korneff are there any flags I need to enable when compiling HISE to get the RT_Neural stuff compiling into a plugin properly?
-
RE: Hardcoded Neural Network does not work as expected
@ustk so basically every NN model uses a number of weights. This results in a certain number of parameters being able to be tweaked by the AI essentially. The more parameters the more CPU is required for processing.
With simpler circuits, or even say individual components, you could probably use very small models. I am actually planning to use the same approach as you, using NNs just for the non linnear stuff. I'm hoping to get it all working at higher sample rates though to reduce aliasing which would stack up quickly if you are chaining NNs together.
-
RE: Apple has finally included 16GB of RAM in the base model! :)
@bendurso If all you want is the base model then yeah, I think this is a great step up and an awesome deal. As usual though with Apple products, if you want anything above the base. Yikes, hope you're willing to pay almost 5x market prices for RAM and NAND upgrades.
-
RE: Hardcoded Neural Network does not work as expected
@Dan-Korneff I'm not sure if the tolerance refers to an absolute loss/ESR value, or whether it actually refers to a threshold of progress between Epochs.
Also, depending on the size of the model, there's basically a hard limit to how close the model can be to the original. The Rat is quite harmonically rich, so it will be on the harder end to model.
-
RE: Help me understand Waveshaping
@clevername27 I appreciate your sentiment, but I'm just a guy with a normal day job with an interest in plugins. Not looking to make any money here, just releasing my work for free as and when inspiration strikes.
-
RE: Help me understand Waveshaping
@griffinboy ha ha yeah. To be honest I wasn't expecting to make some UAD level simulation but just learn the basics of HISE a bit more.
I'm already very familiar with AI though through making a lot of NAM profiles and being one of the co-founders of ToneHunt.
Really looking forward to implementing quite a few ideas once that process is ironed out.
-
RE: Help me understand Waveshaping
@griffinboy Second question I guess. Is this all affected by the sample rate. -1/1 at 24bit has a much tighter ranger than -1/1 at 32bit float right?