Hi everyone,
Once again, we have exciting news!
We are thrilled to announce the release of the Neural Amp Modeler (NAM), developed by Steven Atkinson, following the successful launch of the AIDA-X neural model player in March.
NAM was initially designed for desktop use, with a strong emphasis on model accuracy and user-friendly model training. It has quickly gained popularity within the guitarist community, as mentioned previously in the forum, and has led to the creation of a comprehensive model database called ToneHunt, which already hosts thousands of models.
In the beginning, the earlier versions of NAM and its models were too resource-intensive to run on embedded devices like ours. However, with continuous software evolution and optimization efforts, the NAM team has recently introduced a new training setting that produces lightweight models suitable for the MOD Dwarf and Duo X devices.
We have collaborated with @MikeOliphant to bring his LV2 implementation of NAM to our platform, and now everyone can download it from the Plugin Store. Thank you, Mike!
Please note that you must have MOD OS 1.13.2 or a higher version installed, as it includes the NAM section in the File Manager. This is where you should place your NAM model files.
Regarding the management of model files, we strongly recommend utilizing ToneHunt as the central repository for NAM models. It is crucial that we refrain from using our Forum to share NAM models. The AIDA team is already in contact with the ToneHunt developers, and we hope that soon it will also be possible to share and download AIDA-X models from there, rendering the Forum unnecessary as a model exchange tool.
Previously, NAM models were heavier and unable to run on our devices. However, the current standard workflow allows for models that can be run on our devices, although not optimally yet. These lighter models are referred to as “nano” weights.
Both projects are actively evolving, and developments can occur rapidly. For the time being, let’s adhere to the following guidelines:
When searching for a model on ToneHunt, look for the “nano-mdl” tag. These models are specifically designed to be lightweight and compatible with our devices. You can directly visit: #nano-mdl Models | ToneHunt
If you have previously trained an AIDA model and still have the captures, it is easy to retrain them for NAM and share them on ToneHunt. Kais has provided detailed instructions in this post: Training for NAM (Neural Amp Modeler)
To contribute effectively to the communities and facilitate meaningful discussions and comparisons among different models and weights, let’s provide NAM files, both LSTM and WaveNet, to ToneHunt users. This will not only support their amazing work but also offer valuable resources for users.
As our interaction with ToneHunt gains traction, they will undoubtedly work towards better supporting and filtering different model structures and weights.
That concludes the information dump for now!
I am extremely excited about this release and anticipate an influx of new users who will undoubtedly be enthralled by this neural technology.
Below, you’ll find an appendix with my understanding of the technical details about the two projects. It provides further insight into the reasons behind the guidelines mentioned above and offers a glimpse into what lies ahead. Brace yourself for an abundance of acronyms!
Best wishes and happy neural rocking!
Gianfranco, aka The MOD Father
APPENDIX WARNING: TECHNICAL DIGRESSION ABOUT A.I. MADE BY A MECHANICAL ENGINEER
When training a model, we essentially capture the behaviour of a physical device. These captures can be made using different numerical structures, each with its own advantages and disadvantages, much like using JPEG and PNG for images. The three structures commonly used are LSTM, GRU, and WaveNet. Generally, LSTM and GRU require less CPU than WaveNet but sacrifice accuracy to achieve that efficiency.
In addition to the model structure, we have the concept of model size, which refers to the numerical density within the structure. This is comparable to the quality setting of a JPEG or the bitrate of an MP3. The model size directly impacts CPU usage when using the model, regardless of the structure employed. Larger models require higher CPU resources. The model size is defined during training and can range from lighter to heavier settings. It is a parameter set during training, independent of whether LSTM, GRU, or WaveNet structures are used.
As with other areas, discussions about the quality outcomes of these parameters can be endless. While using a powerful Intel or AMD CPU allows for maximum sizes to achieve the best possible quality, many of us are willing to make compromises in exchange for convenience and practicality. Similar to the popularity of JPEGs and MP3s, “good enough” quality does not require bloated and heavy files.
Furthermore, we have seen in practice that proper capturing and training speak much more than model formats in regards to the resulting sound quality and, if the capture is crap, there is no WaveNet model capable of sounding good.
If you go now to the beta shop and search for “AIDA” you will see three plugins that are going to be released to stable very soon. Those are all using LSTM / GRU and I am still to hear better sounding models than these jaw dropping plugin, thanks to the great work made by @madmaxwell and @spunktsch
NAM’s primary mission is accuracy, and its standard approach is to use WaveNet structures during training. However, NAM can also train LSTM models, albeit with some additional steps instead of using the user-friendly training colab notebook. NAM does not support GRU. The NAM plugin can seamlessly load both LSTM and WaveNet models, as long as they were trained using NAM.
On the other hand, AIDA primarily uses LSTM as the standard structure when training with the simple online colab notebook, but it can also train GRU models when using the local training method. AIDA-X can load both LSTM and GRU models interchangeably. AIDA-X does not support WaveNet.
Currently, the lightest NAM models available are the nano-weighted WaveNet models. These models consume 66% of the CPU on a Dwarf device. However, by utilising local training, you can produce an LSTM model that sounds very similar to the nano WaveNet while consuming only 34% of the CPU on the Dwarf.
By providing ToneHunt with different formats for the same capture, we can offer their community not only the models themselves but also a basis for objective comparisons, ultimately helping to popularize LSTM and GRU models.
In the near future, it is highly likely that both players will support each other’s formats, simplifying the user experience.
In an ideal scenario, we will at one point have a unified training method that combines the best of both AIDA and NAM, allowing users to create heavy or light models according to their preferences.
As the great @harryhaaren used to say, “The future is a big place”
END OF DIGRESSION