Make groq the default provider, ollama as alternative
Groq (free tier, API key) is the primary setup path. Ollama (local, no key) shown as secondary option. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
16
README.md
16
README.md
@@ -53,17 +53,19 @@ cp .env.example .env
|
|||||||
|
|
||||||
> **note:** droidclaw requires [bun](https://bun.sh), not node/npm. it uses bun-specific apis (`Bun.spawnSync`, native `.env` loading) that don't exist in node.
|
> **note:** droidclaw requires [bun](https://bun.sh), not node/npm. it uses bun-specific apis (`Bun.spawnSync`, native `.env` loading) that don't exist in node.
|
||||||
|
|
||||||
edit `.env` - fastest way to start is with ollama (fully local, no api key):
|
edit `.env` - fastest way to start is with groq (free tier):
|
||||||
|
|
||||||
|
```bash
|
||||||
|
LLM_PROVIDER=groq
|
||||||
|
GROQ_API_KEY=gsk_your_key_here
|
||||||
|
```
|
||||||
|
|
||||||
|
or run fully local with [ollama](https://ollama.com) (no api key needed):
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# option a: local with ollama (no api key needed)
|
|
||||||
ollama pull llama3.2
|
ollama pull llama3.2
|
||||||
LLM_PROVIDER=ollama
|
LLM_PROVIDER=ollama
|
||||||
OLLAMA_MODEL=llama3.2
|
OLLAMA_MODEL=llama3.2
|
||||||
|
|
||||||
# option b: cloud with groq (free tier)
|
|
||||||
LLM_PROVIDER=groq
|
|
||||||
GROQ_API_KEY=gsk_your_key_here
|
|
||||||
```
|
```
|
||||||
|
|
||||||
connect your phone (usb debugging on):
|
connect your phone (usb debugging on):
|
||||||
@@ -206,8 +208,8 @@ name: Send WhatsApp Message
|
|||||||
|
|
||||||
| provider | cost | vision | notes |
|
| provider | cost | vision | notes |
|
||||||
|---|---|---|---|
|
|---|---|---|---|
|
||||||
|
| groq | free tier | no | fastest to start |
|
||||||
| ollama | free (local) | yes* | no api key, runs on your machine |
|
| ollama | free (local) | yes* | no api key, runs on your machine |
|
||||||
| groq | free tier | no | fastest cloud option |
|
|
||||||
| openrouter | per token | yes | 200+ models |
|
| openrouter | per token | yes | 200+ models |
|
||||||
| openai | per token | yes | gpt-4o |
|
| openai | per token | yes | gpt-4o |
|
||||||
| bedrock | per token | yes | claude on aws |
|
| bedrock | per token | yes | claude on aws |
|
||||||
|
|||||||
@@ -793,19 +793,18 @@ cp .env.example .env</pre>
|
|||||||
<div class="stepper-step">
|
<div class="stepper-step">
|
||||||
<span class="stepper-num">2</span>
|
<span class="stepper-num">2</span>
|
||||||
<h3>configure an llm provider</h3>
|
<h3>configure an llm provider</h3>
|
||||||
<p>edit <code>.env</code> - fastest way is ollama (fully local, no api key):</p>
|
<p>edit <code>.env</code> - fastest way to start is groq (free tier):</p>
|
||||||
<pre># local (no api key needed)
|
<pre>LLM_PROVIDER=groq
|
||||||
ollama pull llama3.2
|
GROQ_API_KEY=gsk_your_key_here
|
||||||
LLM_PROVIDER=ollama
|
|
||||||
|
|
||||||
# or cloud (free tier)
|
# or run fully local with ollama (no api key)
|
||||||
LLM_PROVIDER=groq
|
# ollama pull llama3.2
|
||||||
GROQ_API_KEY=gsk_your_key_here</pre>
|
# LLM_PROVIDER=ollama</pre>
|
||||||
<table>
|
<table>
|
||||||
<thead><tr><th>provider</th><th>cost</th><th>vision</th><th>notes</th></tr></thead>
|
<thead><tr><th>provider</th><th>cost</th><th>vision</th><th>notes</th></tr></thead>
|
||||||
<tbody>
|
<tbody>
|
||||||
|
<tr><td>groq</td><td>free</td><td>no</td><td>fastest to start</td></tr>
|
||||||
<tr><td>ollama</td><td>free (local)</td><td>yes*</td><td>no api key, runs on your machine</td></tr>
|
<tr><td>ollama</td><td>free (local)</td><td>yes*</td><td>no api key, runs on your machine</td></tr>
|
||||||
<tr><td>groq</td><td>free</td><td>no</td><td>fastest cloud option</td></tr>
|
|
||||||
<tr><td>openrouter</td><td>per token</td><td>yes</td><td>200+ models</td></tr>
|
<tr><td>openrouter</td><td>per token</td><td>yes</td><td>200+ models</td></tr>
|
||||||
<tr><td>openai</td><td>per token</td><td>yes</td><td>gpt-4o</td></tr>
|
<tr><td>openai</td><td>per token</td><td>yes</td><td>gpt-4o</td></tr>
|
||||||
<tr><td>bedrock</td><td>per token</td><td>yes</td><td>claude on aws</td></tr>
|
<tr><td>bedrock</td><td>per token</td><td>yes</td><td>claude on aws</td></tr>
|
||||||
|
|||||||
@@ -168,12 +168,12 @@ fi
|
|||||||
|
|
||||||
printf "next steps:\n\n"
|
printf "next steps:\n\n"
|
||||||
printf " ${BOLD}1.${RESET} configure an llm provider in ${CYAN}droidclaw/.env${RESET}\n\n"
|
printf " ${BOLD}1.${RESET} configure an llm provider in ${CYAN}droidclaw/.env${RESET}\n\n"
|
||||||
printf " ${DIM}# local with ollama (no api key needed)${RESET}\n"
|
printf " ${DIM}# groq (free tier, fastest to start)${RESET}\n"
|
||||||
printf " ollama pull llama3.2\n"
|
|
||||||
printf " ${DIM}# set in .env:${RESET} LLM_PROVIDER=ollama\n\n"
|
|
||||||
printf " ${DIM}# or cloud with groq (free tier)${RESET}\n"
|
|
||||||
printf " ${DIM}# set in .env:${RESET} LLM_PROVIDER=groq\n"
|
printf " ${DIM}# set in .env:${RESET} LLM_PROVIDER=groq\n"
|
||||||
printf " ${DIM}# ${RESET} GROQ_API_KEY=gsk_...\n\n"
|
printf " ${DIM}# ${RESET} GROQ_API_KEY=gsk_...\n\n"
|
||||||
|
printf " ${DIM}# or local with ollama (no api key needed)${RESET}\n"
|
||||||
|
printf " ollama pull llama3.2\n"
|
||||||
|
printf " ${DIM}# set in .env:${RESET} LLM_PROVIDER=ollama\n\n"
|
||||||
printf " ${BOLD}2.${RESET} connect your android phone (usb debugging on)\n\n"
|
printf " ${BOLD}2.${RESET} connect your android phone (usb debugging on)\n\n"
|
||||||
printf " adb devices\n\n"
|
printf " adb devices\n\n"
|
||||||
printf " ${BOLD}3.${RESET} run it\n\n"
|
printf " ${BOLD}3.${RESET} run it\n\n"
|
||||||
|
|||||||
Reference in New Issue
Block a user