Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| faq [2026/04/11 05:29] – kat | faq [2026/04/11 16:19] (current) – beeg brane here amolith | ||
|---|---|---|---|
| Line 1: | Line 1: | ||
| ====== Frequently Asked Questions ====== | ====== Frequently Asked Questions ====== | ||
| - | ===== When will Synthetic get model X? ===== | + | {{namespace> |
| - | + | ||
| - | Before you ask this, make sure it: | + | |
| - | + | ||
| - | - Is [[https:// | + | |
| - | - Has weights available on [[https:// | + | |
| - | - Has a compatible license that allows Synthetic to actually make money hosting a model (sometimes model weights are published " | + | |
| - | - Has an [[https:// | + | |
| - | - Has solid support for that model or its general architecture in [[https:// | + | |
| - | + | ||
| - | Factors that can delay Synthetic getting a model: | + | |
| - | + | ||
| - | * If it is unusually large, it may take time for them to acquire or free up GPU space to host it. | + | |
| - | + | ||
| - | * If it has a novel or unusual architecture (such as DeepSeek Sparse Attention for GLM 5), it will take time for inference engines like sglang to get reliable support for the model. | + | |
| - | + | ||
| - | * If the model has not yet been quantized to NVFP4, Synthetic will have to wait for NVIDIA to do that, or make one themselves, both of which can take some time. | + | |
| - | + | ||
| - | Additionally, | + | |
| - | + | ||
| - | ===== Why is model X from Synthetic not in $preferred_harness' | + | |
| - | + | ||
| - | Many of these lists are updated by hand by a human, so you might be the first '' | + | |
| - | + | ||
| - | * Wait until someone else opens a PR to '' | + | |
| - | * Find where '' | + | |
| - | * OpenCode has [[https:// | + | |
| - | * Crush has [[https:// | + | |
| - | * Use some kind of provider-specific plugin for '' | + | |
| - | * Pi has [[https:// | + | |
| - | * Maintain your own provider/ | + | |
| - | * [[https:// | + | |
| - | * [[https:// | + | |
| - | * [[https:// | + | |
| - | + | ||
| - | ===== Why am I burning through my credits or requests so quickly? ===== | + | |
| - | + | ||
| - | ==== Step 1: Check Your Tools ==== | + | |
| - | + | ||
| - | The software you use with Synthetic has a massive impact on your token "burn rate." | + | |
| - | + | ||
| - | === 1. Are you using Claude Code? === | + | |
| - | + | ||
| - | **Recommendation: | + | |
| - | + | ||
| - | === 2. Are you using OpenCode (with oh-my-opencode/ | + | |
| - | + | ||
| - | While OpenCode is significantly better than Claude Code, it is still not optimized for efficiency. If you are using the oh-my-opencode/ | + | |
| - | + | ||
| - | === 3. Are you using Zed? === | + | |
| - | + | ||
| - | Zed is a powerful editor, but its real-time " | + | |
| - | + | ||
| - | - **Intent to Edit:** It sends a request to define the goal. | + | |
| - | - **Execution: | + | |
| - | + | ||
| - | Because of this " | + | |
| - | + | ||
| - | ==== Step 2: Optimize Your Workflow ==== | + | |
| - | + | ||
| - | If your tools aren't known for being wasteful but your usage remains high, follow these steps (roughly in order of recommendation) to reduce token bloat: | + | |
| - | + | ||
| - | === 1. Reduce Frequency === | + | |
| - | + | ||
| - | Reduce the frequency of automated workflow runs. For example, if you use OpenClaw, review your current tasks in OpenClaw' | + | |
| - | + | ||
| - | === 2. Prompt Efficiency === | + | |
| - | + | ||
| - | Refine your input tokens. Use concise system prompts and AGENTS.md files. | + | |
| - | + | ||
| - | === 3. Model Tiering === | + | |
| - | + | ||
| - | Switch to a cheaper model for simpler tasks; Kimi and GLM don't need to be running for every single prompt. | + | |
| - | + | ||
| - | === 4. Limit " | + | |
| - | + | ||
| - | For less complex tasks, reduce the model' | + | |
| - | + | ||
| - | === 5. Increase Packs === | + | |
| - | + | ||
| - | If your workflow is already lean but you still hit limits, it may be time to upgrade your number of packs to match your professional output. | + | |