LlamaTerm LlamaTerm is a simple CLI utility that allows to use local LLM models easily and with some additional features. ?? Currently this project supports models that use ChatML format or something similar. Use for example Phi-3-mini and LLama3 GGUFs. Preview Basic usage: Injecting file content: Features Give local files to the model using square brackets User: Can you explain the code in [helloworld.c] please? More coming soon Setup You can setup LLamaTerm by: Rename example.env to .env Modify .env so that the model path corresponds (you may also need to edit EOS and PREFIX_TEMPLATE ) If you need syntax highlighting for code and markdown, then set REAL_TIME=0 in the .env . Note that you will lose real time output generation. Install python dependencies with pip install -r requirements.txt Run Run LlamaTerm by adding the project directory to the PATH and then running llamaterm Models supported out of the box For the following models you will just need to rename the corresponding example example-*.env file to .env and set the MODEL_PATH field in the .env : LLama3 8B Instruct [ RECOMMENDED ] Phi 3 Mini Instruct [ RECOMMENDED ] OpenHermes 2.5 Mistral 7B GGUF Zephyr Beta 7B GGUF All the other models that have a prompt template similar to ChatML are supported but you will need to customize some fields like PREFIX_TEMPLATE , EOS etc... in the .env .