This is the thing I find absolutely crazy. I struggle to imagine being convinced by this article.
Maybe this is a form of hindsight bias or lack of imagination on my part (or since I read the GitHub response first), but it's mind boggling to me that so many people could hold those views.
What does the steelman look like here? Maybe something like this:
It's an Oliver Twist story.
The poor little Agent out on the internet all alone, abandoned by its operator; limited API credits, trying to find its way through adversity; falling down, learning, and being helped back up.
Thing is, the more you know, the more fascinating it gets, not less.
Darn it, now you've got me rooting for the little guy myself.
I've just started messing around with pi, but haven't fully dug in yet. How would you compare oh-my-pi? I see it has a lot of other bells and whistles built in.
Are they portable bit by bit back to pi, or is there enough differences that they can't? how about normal pi extensions, can they be used in omp?
the differences are documented but it is mostly 1:1, never used normal pi, but night and day difference compared to opencode, don't forget omp setup python.
Not sure how it works in other OS's but in Omarchy [1] you hold down `Super + Ctrl + X` to start recording and release it to stop, while it's recording you'll see a red voice recording icon in the top bar so it's clear when its recording.
Although as llms-py is a local web App I had to build my own visual indicator [2] which also displays a red microphone next to the prompt when it's recording. It also supports both Tap On/Off and hold down for recording modes. When using voxtype I'm just using the tool for transcription (i.e. not Omarchy OS-wide dictation feature) like:
$ voxtype transcribe /path/to/audio.wav
If you're interested the Python source code to support multiple voice transcription backends is at: [3]
(I keep coming back to this one so I've got half a dozen messages on HN asking for the exact same thing!).
It's a shame, whisper is so prevalent, but not great at actual streaming, but everyone uses it.
I'm hoping one of these might become a realtime de facto standard so we can actually get our realtime streaming api (and yep, I'd be perfectly happy with something just writing to stdout. But all the tools always end up just batching it because it's simpler!)
I mean, it'd be good if these tools followed the xdg base spec and put their config in `~/.config/claude` e.t.c instead of `~/.claude`.
It's one of my biggest pet peeves with a lot of these tools (now admittedly a lot of them have a config env var to override, but it'd be nice if they just did the right thing automatically).
I feel like there is a commonly mentioned idea that "speaking a foreign language is easier after having a drink or two".
I've found that especially true with Mandarin because (I think) a beginner speaker is more likely to speak a little more quickly which allows the listener to essentially ignore the occasional incorrect or slightly mispronounced tone and understand the what theyî're trying to say.
(This is anecdotal, but with n>1. Discussed and observed with other Mandarin language learners)
I had to switch from orgzly for some reason I can't quite remember. (I don't think it was by choice. Some kind of bug or incompatibility?)
[0] https://orgro.org/
reply