The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
localLLM provides an easy-to-use interface to run
large language models (LLMs) directly in R. It uses the performant
llama.cpp library as the backend and allows you to generate
text and analyze data with LLM. Everything runs locally on your own
machine, completely free. It also ensures reproducibility by default.
Our goal is to develop it into a reliable toolkit for scientific
research.
Tutorial: https://www.eddieyang.net/software/localLLM
Getting started requires two simple steps: installing the R package
from CRAN and then downloading the backend C++ library that handles the
heavy computations. The install_localLLM() function
automatically detects your operating system (Windows, macOS, Linux) to
download the appropriate pre-compiled library.
# 1. Install the R package from CRAN
install.packages("localLLM")
# 2. Load the package and install the backend library
library(localLLM)
install_localLLM()You can start running an LLM using quick_llama().
library(localLLM)
# Ask a question and get a response
response <- quick_llama('Classify whether the sentiment of the tweet is Positive
or Negative.\n\nTweet: "This paper is amazing! I really like it."')
cat(response) # Output: The sentiment of this tweet is Positive.By default , all generation functions in localLLM
(quick_llama(), generate(), and
generate_parallel()) use deterministic greedy decoding.
Even when temperature > 0, results are reproducibile.
response1 <- quick_llama('Classify whether the sentiment of the tweet is Positive
or Negative.\n\nTweet: "This paper is amazing! I really like it."',
temperature=0.9, seed=92092)
response2 <- quick_llama('Classify whether the sentiment of the tweet is Positive
or Negative.\n\nTweet: "This paper is amazing! I really like it."',
temperature=0.9, seed=92092)
print(response1==response2)Please report bugs to xu2009@purdue.edu with your sample code and data file. Much appreciated!
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.