Hacker Newsnew | past | comments | ask | show | jobs | submit | Squarex's commentslogin

And what cross platform code editor does that nowadays? vscode is electron, jetbrains has swing, ...

Maybe what I'm saying is that people shouldn't use cross-platform UI solutions. (write-once-suck-everywhere)

The old ones, and the ones that use Qt.

I hope someone will create a lightweight version without AI and code editing stuff. The terminal experience is the best, but I don't have any use for the agentic stuff while having claude code, opencode, codex and plenty other options.

It's good feedback. We've tried to make it so there is a single "turn off all the AI stuff" button (and you can opt into plain old terminal during onboarding as well, with no login, etc). Curious if this does the trick?

~And where is it? I am a long time signed in user, so no onboarding for me.~ How would you make money from users like me?

edit: nevermind, it was quite discoverable...


the rpm available for download (warp-terminal-0:v0.2026.04.27.15.32.stable_03-1) doesn't seem to work without login. after going through all the setup i got stuck at a prompt asking me to log in with no option to skip it.

Hey Zach - one thing I'm really missing is the ability for this to be toggled on/off per device - whilst I love it on my personal devices and want to use AI there, I also want to be able to use and log into warp at work without having to toggle it off, as I can't use AI there.

From other comments in the thread, login is no longer required so you should be able to just not login on your work machine

https://github.com/warpdotdev/warp


I originally got into Warp because they made a terminal where my normal text input keyboard shortcuts work.

As they've been scrambling to find a way to monetize and riding the AI train, it feels more bloated than ever and the constant pushing for me to use "agents" and whatnot really put my off using it. Plus with all the privacy concerns I can't with good conscience use it on my work machine.

So yes, I'd like a non-tracking, no-AI version of Warp too.


100,000% I loved it when it was JUST good at terminal rendering. Ghostty alternative. Not a crazy bloatware AI play.

This is a pretty good used case for vibecoding. “Claude, take this project and rip out all the obnoxious monetization and vendor lock in.” It just might do the trick. I’ve been to get rid of a fair bit of paid software by just cloning the parts I want with little more than a high-level description.

I don’t think the approach of open source as a substitute for a quality program is going to last.


can you share more about what makes it so great. this is the first i am hearing about it , so i am curious.

i currently use tmux and ghostty for my workflow


The way it goes beyond just emulating terminal. Multiline input that works like text editor, separated input and output blocks, wrapped shells that keeps the same ux with local and remote shells, the polish.

You could probably few-shot this yourself by pointing at the repo. I'm 95% sure it can be done in a day end-to-end.

Even more so, questions and user answers from agents were not charged as separate requests.

And when you make your harness ask you for next steps in a tool call, the journey continues forever, yeehaa

I would say all benchmarks are inherently subjective. How is yours better? It seems to produce a little bit strange results. Opus 4.6 being worse than 4.5 for example. Or chinese models being rated too high. Kimi, Deepseek or GLM are all great in open source world, but I don't believe they are ahead of SOTA models from Anthropic, OpenAI or Google.

No, some benchmarks are definitely objective, but most can be easily gamed. For example, most of the benchmarks on the model cards: they have measurable answers that don't rely on a human judge (a human made the question, but the answers are measuring some uncontroversial knowledge or capability). But because there is a single, correct answer, and those answer leak (or are randomly discovered and optimized for in training), they lose value over time, and regardless, they have a ceiling on the intelligence they can measure.

Others are purely subjective, like LMArena, which really only measures the personality and style preferences of the masses at this point, because frontier LLM technical answers are too hard for the average person to judge.

Then there are some interesting one-off benchmarks, but they lack enough rigor, breadth, and samples to draw larger conclusions from.

So we designed our benchmark with 3 goals: objective measurements (individual submissions not dependent on a human or LLM judge), no known correct answer (so simulations can scale to much higher levels of intelligence), and enough variety over important aspects of intelligence. We do this by running multiple models in cooperative/competitive environments with very complex action spaces and objective scoring, where model performance is relative and affected by the actions of other participants.

And yeah, there are some interesting results when you have a more objective benchmark. It should raise eyebrows when every single sub-release of every company's model is better across the board than its predecessor -- that isn't reality.


The word "objective" just seems too authoritative to me.

I agree that benchmarks are inherently subjective.

but the fact that you cite your brief as your main argument is funny - you don't even have any inherently subjective numbers to justify what you believe, you only have "I don't believe".


Sure, I have mixed up two things together. I don't think this benchmark is bad, I just did not like it is presented as the ultimate objective truth. The other thing I have mentioned is that it delivers different results from other benchmarks, so the "believe" stems from other benchmarks.

you are arguing with your belief instead of an objective truth. benchmark is more objective, if you don't agree with it, come up with a better one. but what you believe doesn't matter.

It was not a confrontational take. But all benchmarks are designed by humans, we are not that great at measuring intelligence. So it is somewhat subjective. I was just arguing with the word "objective". Not with the results per se.

If the benchmark has a correct answer, the benchmark itself is an objective measure (but of what?). The "of what" may well be subjective

Only if the benchmark is private and done properly on relevant tasks, which is rarely the case. I can guarantee that you have a ton of blind spots if you look at it through the lens of a ranking ladder in some generic tasks.

The rumor was that the 5.5 is a brand new pretrain. But who knows, it's 2x as expensive as 5.4, so it would check out.

If so that would be big, they haven’t been able to successfully pretrain in close to two years (since 4o).

As a europe federalist, I would think it is more likely EU would implement these restrictions itself instead of step against Spain.


In theory, we should already be protected against this via the various "Net neutrality" directives, but as the US currently is showing us, laws and regulations are only worth as much as you're willing to enforce them ultimately. But things like these are supposed to be worth at least something:

> Regulation 2015/2120 also states that access providers “shall treat all traffic equally, when providing internet access services, without discrimination, restriction or interference, and irrespective of the sender and receiver, the content accessed or distributed, the applications or services used or provided, or the terminal equipment used,” although they are permitted to apply “reasonable traffic management measures.” In any case, those measures must be “transparent, non-discriminatory and proportionate, and shall not be based on commercial considerations but on objectively different technical quality of service requirements of specific categories of traffic” (Article 3.3) - https://www.cuatrecasas.com/en/global/intellectual-property/...

Remains to be seen if something/someone will put a stop to La Liga's shenanigans, judges have seem unwilling so far, and not a big enough problem for the average person to really care about it (yet?).


The regulation has an opt out for court orders though, which these are.


Codex and gemini cli are open source already. And plenty of other agents. I don't think there is any moat in claude code source.


Well, Claude does boast an absolutely cursed (and very buggy) React-based TUI renderer that I think the others lack! What if someone steals it and builds their own buggy TUI app?


Your favorite LLM is great at building a super buggy renderer, so that's no longer a moat


Gemini-cli is much worse in my experience but I agree


I think that’s a real problem now. In our parliament (Czech) almost every politician is a lawyer or a doctor. Almost no other profession is represented.


It is behind paywall, but the question itself seems like trivial.


It is clearly not. Why would you think so?


the ux feels extremely similar down to the elicitation ... but I did some more research ... they were started independently in april 2025. Therefore, one being a fork of the other is almost impossible and there is no evidence for it. Also, opencode is in go and gemini is in typescript.

Sadly my above misinformation can no longer be edited.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: