Hacker Newsnew | past | comments | ask | show | jobs | submit | more mudkipdev's commentslogin

I'm getting a "failed to verify your browser" error on this article


And, ironically, it's hosted on vercel :D


Why do you need an API key to tokenize the text? Isn't it supposed to be a cheap step that everything else in the model relies on?


I'd guess it's because they don't want people to reverse engineer it.

Note that they're the only provider which doesn't make their tokenizer available offline as a library (i.e. the only provider whose tokenizer is secret).


Anthropic is somewhat becoming the Apple of AI in terms of closed ecosystem. Not saying I blame them, I just don't like it as a customer.

The fact that it's impossible to get the actual thinking tokens anymore, but we have to do with a rewritten summary, is extremely off-putting. I understand that it's necessary for users, but when writing agentic applications yourself, it's super annoying not to have the actual reasoning of the agent to understand failure modes.


It's _not_ that it's necessary for users. It's that Anthropic got Opus 4.6 ripped off so hard by MiniMax that they no longer want to expose true thinking tokens to random developers. If you're one of the blessed class, you can still get real thinking tokens, but you need to be a major enterprise customer, like the companies that they gave Mythos access.


To prevent abuse? It's a completely free endpoint so I don't understand your complaint.


It may be free, but it cannot be used without credits.

  Error: {"type":"error","error":{"type":"invalid_request_error","message":"Your credit balance is too low to access the Anthropic API. Please go to Plans & Billing to upgrade or purchase credits."},"request_id":"req_011CaGaBf6uTHfbmdZ39nx1Z"}


Again it is to help prevent abuse I don't really see how this is a valid concern? Tokenization is actually fairly CPU intensive


I'd love it if that API (which I do not believe Anthropic charge anything for) worked without an API key.


Grayish dark themes are underrated


The Claude prompt is already quite bloated, around 7,000 tokens excluding tools.


If anyone has a better workflow for creating lots of captions in kdenlive please let me know. I had to duplicate each title to the media library and drag it into the timeline, because if I simply copy/pasted then the text content/styling would be shared across instances


Re-read that


You should. 3.5 MoE was worse than 3.5 dense, so expecting 3.6 MoE to be superior than 3.5 dense is questionable, one could argue that 3.6 dense (not yet released) to be superior than 3.5 dense.


Ok but you made a claim about the new model by stating a fact about the old model. It's easy to see how you appeared to be talking about different things. As for the claim, Qwen do indeed say that their new 3.6 MoE model is on a par with the old 3.5 dense model:

> Despite its efficiency, Qwen3.6-35B-A3B delivers outstanding agentic coding performance, surpassing its predecessor Qwen3.5-35B-A3B by a wide margin and rivaling much larger dense models such as Qwen3.5-27B.

https://qwen.ai/blog?id=qwen3.6-35b-a3b


This says a slightly different thing:

https://x.com/alibaba_qwen/status/2044768734234243427?s=48&t...

If you look, at many benchmarks the old dense model is still ahead but in couple benchmarks the new 35B demolishes the old 27B. "rivaling" so YMMV.


Does the large system prompt work fine for this model? If needed, you could use a lightweight CLI like Pi, which only comes with 4 tools by default


I built a Claude-inspired UI for Ollama/llama.cpp

https://github.com/mudkipdev/chat


People have made toki pona translation models before, not exclusively trained though


It's strange that my iPhone 14 is at regular temperature when using the E2B model. But also it's a lot slower (not sure how to measure the exact tokens per second, ~12 if I had to guess)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: