OnPrem.LLM: Launch Autonomous AI Agents with Sandboxed Execution in 2 Lines of Code
OnPrem.LLM's AgentExecutor enables launching autonomous AI agents with 9 built-in tools (file ops, shell, web search) that work with both cloud and local models, featuring sandboxed execution and working directory isolation.
OnPrem.LLM has introduced the AgentExecutor — a pipeline for creating autonomous AI agents that can execute complex tasks using a variety of tools, with support for both cloud and local models.
How It Works
The AgentExecutor is implemented using their coding agent, PatchPal, and provides a simple interface for launching AI agents with sandboxed execution:
from onprem.pipelines import AgentExecutor
# Full access (all tools including shell):
executor = AgentExecutor(model='anthropic/claude-sonnet-4-5')
# Safer mode (no shell access):
executor = AgentExecutor(model='openai/gpt-5-mini', disable_shell=True)
Supported Models
The pipeline works with any LiteLLM-supported model that supports tool-calling:
- Cloud: openai/gpt-5.2-codex, anthropic/claude-sonnet-4-5, gemini/gemini-1.5-pro
- Local: Ollama (ollama/llama3.1), vLLM (hosted_vllm/), llama.cpp (via OpenAI interface)
9 Built-in Tools
- read_file — Read complete file contents
- read_lines — Read specific line ranges from files
- edit_file — Edit files via find/replace
- write_file — Write complete file contents
- grep — Search for patterns in files
- find — Find files by glob pattern
- run_shell — Execute shell commands
- web_search — Search the web for information
- web_fetch — Fetch and read content from URLs
Security Features
- Working directory isolation — agents cannot read or write outside their specified working directory
- Shell access is optional — can be disabled for safer deployments
- Tool whitelisting — users can specify exactly which tools an agent can access
- Sandboxed execution — agents operate within defined boundaries
Use Cases
The system is designed for tasks like autonomous code generation, file manipulation, web research, and complex multi-step workflows — all from a local environment without sending data to external services (when using local models).
Source: OnPrem.LLM Documentation | GitHub | HN: 48pts
← Previous: Python 3.15's JIT Compiler Is Back on Track — 11% Faster on ARMNext: Google DeepMind Proposes Cognitive Framework for Measuring Progress Toward AGI →
0