The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
The {shiny.telemetry}
package can be used with any Shiny
application and in this guide we will show how to use it with different
databases backend.
The following databases are supported by
{shiny.telemetry}
:
A requirements to use {shiny.telemetry}
with external
databases in a production environment is to have the database server
running and a user with the necessary permissions to insert. A minimal
setup should have a user that only has write/insert permissions to the
{shiny.telemetry}
table storing the events. The read
permission is only necessary for processing the data, such as the
default analytics dashboard that we provide with the package (see
analytics_app()
). This database setup can be done by an
infrastructure team or the database administrator.
We provide example applications for each database backend with
necessary R code to run both the application and the analytics server.
This is further supported with a docker-container.yml
to
help users quickly setup and test the apps locally. It requires Docker
(docker compose up -d
) or Podman
(podman-compose up -d
) installed to run the containers.
These applications are available under the
inst/examples/
folder or via the GitHub
link.
Each data storage backend will create the necessary tables (or “Collection” in the case of MongoDB) with the respective schema when needed.
The arguments to create an data storage instance vary, as different databases require their own options. However, once the data storage object is created, the read and write operations have the same API.
Below you find chunks to create a data storage object for each supported database.
Sys.setenv("POSTGRES_USER" = "postgres", "POSTGRES_PASS" = "mysecretpassword")
data_storage <- DataStoragePostgreSQL$new(
user = Sys.getenv("POSTGRES_USER"),
password = Sys.getenv("POSTGRES_PASS"),
hostname = "127.0.0.1",
port = 5432,
dbname = "shiny_telemetry",
driver = "RPostgreSQL"
)
notes:
dbname
database needs to be created before running
the application with {shiny.telemetry}
;driver
allows users to use either
{RPostgreSQL}
or {RPostgres}
R packages;.Renviron
file is the default way in R of setting
up environment variables (instead of Sys.setenv()
as
shown above for convenience).To run PostgreSQL in a container locally, you can use the following
Docker compose file: inst/examples/postgresql/docker-compose.yml
.
Sys.setenv("MARIADB_USER" = "mariadb", "MARIADB_PASS" = "mysecretpassword")
data_storage <- DataStorageMariaDB$new(
user = Sys.getenv("MARIADB_USER"),
password = Sys.getenv("MARIADB_PASS"),
hostname = "127.0.0.1",
port = 3306,
dbname = "shiny_telemetry"
)
notes:
dbname
database needs to be created before running
the application with {shiny.telemetry}
;.Renviron
file is the default way in R of setting
up environment variables (instead of Sys.setenv()
as
shown above for convenience).To run MariaDB in a container locally, you can use the following
Docker compose file: inst/examples/mariadb/docker-compose.yml
.
Sys.setenv(MSSQL_USER = "sa", MSSQL_PASS = "my-Secr3t_Password")
data_storage <- DataStorageMSSQLServer$new(
user = Sys.getenv("MSSQL_USER"),
password = Sys.getenv("MSSQL_PASS"),
hostname = "127.0.0.1",
port = 1433,
dbname = "my_db",
driver = "ODBC Driver 18 for SQL Server",
trust_server_certificate = "YES"
)
notes:
dbname
database needs to be created before running
the application with {shiny.telemetry}
;.Renviron
file is the default way in R of setting
up environment variables (instead of Sys.setenv()
as
shown above for convenience).To run Microsoft SQL Server in a container locally, you can use the
following Docker compose file: inst/examples/mssql/docker-compose.yml
.
data_storage <- DataStorageMongoDB$new(
host = "localhost",
dbname = "test",
authdb = NULL,
options = NULL,
ssl_options = mongolite::ssl_options()
)
To run MongoDB in a container locally, you can use the following Docker compose file: [`inst/examples/mssql/docker-compose.yml`](https://github.com/Appsilon/shiny.telemetry/blob/main/inst/examples/mongodb/docker-compose.yml).
{shiny.telemetry}
The data storage API to read and write events for
{shiny.telemetry}
is consistent across all backends, which
allows the developer to implement and test the package with the most
convenient backend and then easily migrate to an external database.
Therefore, once it is initialized it can be used to create the
Telemetry
object and start a session.
# data_storage variable is initialized with one of the previous code chunks.
telemetry <- Telemetry$new(data_storage = data_storage) # 1. Initialize telemetry with object created above
shinyApp(
ui = fluidPage(
use_telemetry(), # 2. Add necessary javascript to Shiny
numericInput("n", "n", 1),
plotOutput('plot')
),
server = function(input, output) {
telemetry$start_session() # 3. Minimal setup to track events
output$plot <- renderPlot({ hist(runif(input$n)) })
}
)
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.