This tool has been developed using both LM Studio and Ollama as LLM providers. The idea behind using a local LLM, like Google's Gemma-3 1B, is data privacy and low cost. In addition, with a good LLM a ...
Currently optimized for standard Transformer architectures (Llama, Mistral, Gemma, BERT). MoE and custom architectures may have experimental support. Fetching the model's configuration from the ...