Multi-app workflow: ask Gemini about droidclaw.ai, copy response,
switch to WhatsApp, find contact, paste and send.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Groq (free tier, API key) is the primary setup path.
Ollama (local, no key) shown as secondary option.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Site and README now show both curl one-liner and manual steps.
Explicit note that bun is required (node/npm won't work).
Install script improvements: version check, node detection warning,
curl check, Windows detection, .git dir validation, shallow clone,
fallback on bun install --silent, post-install missing dep summary.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
One-liner install: curl -fsSL https://droidclaw.ai/install.sh | sh
Auto-installs bun and adb, clones repo, sets up .env.
Updated site meta tags and canonical URL to droidclaw.ai.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Reuses OpenAIProvider via Ollama's OpenAI-compatible API at localhost:11434.
No API key needed - just install Ollama, pull a model, and set LLM_PROVIDER=ollama.
Vision models (llava, llama3.2-vision) supported for screenshot fallback.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- Update all GitHub links from thisuxhq to unitedbyai
- Add contributors section with sanju sivalingam and somasundaram mahesh
- Update meta author and footer attribution to unitedby.ai
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
New --flow mode executes scripted YAML steps without LLM, mapping 17
commands (tap, type, swipe, scroll, etc.) to existing actions. Element
finding uses accessibility tree text/hint/id matching.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- New actions: open_url, switch_app, notifications, pull_file, push_file, keyevent, open_settings
- Workflow system: runWorkflow() for multi-app sub-goal sequences with --workflow CLI flag
- Export runAgent() with {success, stepsUsed} return for workflow integration
- Fix clipboard_set shell escaping (single-quote wrapping matching skills.ts)
- Improve type action escaping for backticks, $, !, ?, brackets, braces
- Move parseJsonResponse to llm-providers.ts and export it
- Update SYSTEM_PROMPT and Zod schema for 22 total actions
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Removes the unnecessary nesting — all source, config, and docs now live
at the project root for simpler paths and commands.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- Auto-detect screen resolution and compute dynamic swipe coordinates
- Detect foreground app each step via dumpsys activity
- Smart element filtering: deduplicate by position, score by relevance, compact to essentials
- Session logging with crash-safe .partial.json writes and final summary
- Real multimodal vision: send base64 screenshots to LLMs (off/fallback/always modes)
- Multi-turn conversation memory: maintain full chat history across steps with trimming
- Multi-step planning: think/plan/planProgress fields on every LLM decision
- Streaming responses for all 4 providers (OpenAI, Groq, OpenRouter, Bedrock)
- Comprehensive README with examples, architecture docs, and troubleshooting
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>