LLMFarm is an iOS and MacOS app to work with large language models (LLM). It allows you to load different LLMs with certain parameters. # Features * Various inferences * Various sampling methods * Metal * Model setting templates * LoRA adapters support * LoRA FineTune and Export # Inferences * LLaMA * GPTNeoX * Replit * GPT2 + Cerebras * Starcoder(Santacoder) * RWKV (20B tokenizer) * Falcon * MPT * Bloom * StableLM-3b-4e1t * Qwen * Gemma * Phi * Mamba * Others # Multimodal * LLaVA 1.5 models * Obsidian * MobileVLM 1.7B/3B models Note: For Falcon, Alpaca, GPT4All, Chinese LLaMA / Alpaca and Chinese LLaMA-2 / Alpaca-2, Vigogne (French), Vicuna, Koala, OpenBuddy (Multilingual), Pygmalion/Metharme, WizardLM, Baichuan 1 & 2 + derivations, Aquila 1 & 2, Mistral AI v0.1, Refact, Persimmon 8B, MPT, Bloom select llama inferece in model settings. Sampling methods * Temperature (temp, tok-k, top-p) * Tail Free Sampling (TFS) * Locally Typical Sampling * Mirostat * Greedy * Grammar (dont work with GGJTv3)