The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.
R-Package Knowledge Base (RPKB)
This database is the official pre-computed ChromaDB vector database for the paper: DARE: Aligning LLM Agents with the R Statistical Ecosystem via Distribution-Aware Retrieval.
It contains 8,191 high-quality R functions meticulously curated from CRAN, complete with extracted statistical metadata (Data Profiles) and pre-computed embeddings generated by the DARE model.
π Database Overview
- Database Engine: ChromaDB
- Total Documents: 8,191 R functions
- Embedding Model:
Stephen-SMJ/DARE-R-Retriever - Primary Use Case: Tool retrieval for LLM Agents executing data science and statistical workflows in R.
π How to Use (Plug-and-Play)
You can easily download and load this database into your own Agentic workflows using the huggingface_hub and chromadb libraries.
1. Install Dependencies
pip install huggingface_hub chromadb sentence-transformers
2. Download RPKB and Connect
from huggingface_hub import snapshot_download
import chromadb
# 1. Download the database folder from Hugging Face
db_path = snapshot_download(
repo_id="Stephen-SMJ/RPKB",
repo_type="dataset",
allow_patterns="RPKB/*" # Adjust this if your folder name is different
)
# 2. Connect to the local ChromaDB instance
client = chromadb.PersistentClient(path=f"{db_path}/RPKB")
# 3. Access the specific collection
collection = client.get_collection(name="inference")
print(f"β
Loaded {collection.count()} R functions ready for conditional retrieval!")
3. Perform a R Pakcage Retrieval
To retrieve the best function, make sure you encode your query using the DARE model.
from sentence_transformers import SentenceTransformer
# Load the DARE embedding model
model = SentenceTransformer("Stephen-SMJ/DARE-R-Retriever")
# Formulate the query with data constraints
user_query = "I have a high-dimensional genomic dataset named hidra_ex_1_2000.csv in my environment. I need to identify driver elements by estimating regulatory scores based on the counts provided
in the data. Please set the random seed to 123 at the start. I need to filter for fragment lengths between 150 and 600 bp and use a DNA count filter of 5. For my evaluation, please print the
first value of the estimated scores (est_a) for the very first region identified."
# Generate embedding
query_embedding = model.encode(user_query).tolist()
# Search in the database with Hard Filters
results = collection.query(
query_embeddings=[query_embedding],
n_results=3,
include=["metadatas", "distances", "documents"]
)
# Display Top-1 Result
print("Top-1 Function:", results["metadatas"][0][0]["package_name"], "::", results["metadatas"][0][0]["function_name"])
- Downloads last month
- 5