Hacker Newsnew | past | comments | ask | show | jobs | submit | simlevesque's commentslogin

Empty ? Why call it that? It's proactive.

Also it's naive to think they announce their intention to move somewhere. They try to cover it and never tell a soul until it's a done deal.


I've been making skills from arxiv papers for a while. I have a one for multi-object tracking for example. It has a SKILL.md describing all important papers (over 30) on the subject and a folder with each paper's full content as reStructuredText.

To feed Arxiv papers to LLMs I found that RST gives the best token count/fidelity ratio. Markdown lacks precision. LateX is too verbose. I have a script with the paper's urls, name and date that downloads the LateX zips from Arxiv, extracts it, transforms them to RST and then adds them to the right folder. Then I ask a LLM to make a summary from the full text, then I give other LLMs the full paper again with the summary and ask them to improve on and and proofread them. While this goes on I read the papers myself and at the end I read the summaries and if I approve them I add it to the skill. I also add for each paper info on how well the algorithms described do in common benchmarks.

I highly recommend doing something similar if you're working in a cutting-edge domain. Also I'd like to know if anyone has recommendations to improve what I do.


I've been working on ctoth/research-papers-plugin, the pipeline to actually get LLMs to extract the notes. I really like your insight re RST over Markdown! It sounds like we're working on similar stuff and I'll absolutely reach out :)

Another format that's worth investigating is Asciidoc. It supports the richness of Docbook XML but has fewer quirks than rST in my eyes.

I'm gonna look at your plugin. My email is in my profile.

Honestly I think that Markdown with LateX code blocks would be the most efficient representation but when doing it with Pandoc I kept having issues with loss of information and sometimes even syntax error.


I am surprised you found RST better than markdown.

This sounds like it would work, but honestly if you've already read all 30 papers fully, what do you still need to llm to do for you? Just the boilerplate?

I'm trying to make a go library that implements a wide ranges of MOT algorithms and can gather metrics for all of them.

Reading all the papers once isn't the same as this. I find it very useful.

I can ask an LLM to do the basic implementations, then I can refine them (make the code better, faster, cut on memory use), then I can ask the LLM if I'm still implementing the algorithms as they're described in the paper.


> then I can ask the LLM if I'm still implementing the algorithms as they're described in the paper.

Unit testing would save on tokens... unit testing is perfect for validating refactors, or when re-writing a project from one language to the next, build unit tests first.


It lets you filter out interesting papers more quickly.

sounds similar to "LLM Knowledge Bases" https://xcancel.com/karpathy/status/2039805659525644595

I’ve been meaning to build something similar. I will report back once I have something to show.

Thanks for sharing!


Does that even fit in the context? It seems like 30 papers worth of content would just overflow it.

For each paper, have your agent extract a three sentence description, create a description.md, then concat those with the paper names into an INDEX.md which it should consult to find appropriate papers. Also: have your agent tag papers, then autogenerate your tagged collection on the filesystem. Then you get nice things like https://github.com/ctoth/Qlatt/tree/master/papers/tagged

Then something in your {CLAUDE,AGENTS}.md that says: when working on something with relevant context supplied by papers, read the papers before doing the work. You can find all papers plus their descriptions in ./papers/INDEX.md and papers by tag in ./papers/tagged


What is RST?


I really don't like OpenCode. One thing that really irritated me is that on mouse hover it selects options when you're given a set of choices.

It shows issues now. Probably not when the person you're replying to wrote their comment.

11 minutes elapsed between the comments. There is going to be some actual time before a report and the status page being live in a breaking system.

You can use almost any model with Claude Code.

that doesnt make sense. how?

Here's how to use MiniMax v2.7 for example: https://platform.minimax.io/docs/token-plan/claude-code

You just add this to your ~/.claude/settings.json:

  {
    "env": {
      "DISABLE_AUTOUPDATER": "1",
      "ANTHROPIC_BASE_URL": "https://api.minimax.io/anthropic",
      "ANTHROPIC_AUTH_TOKEN": "YOUR_SECRET_KEY",
      "API_TIMEOUT_MS": "3000000",
      "CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC": 1,
      "ANTHROPIC_MODEL": "MiniMax-M2.7-highspeed",
      "ANTHROPIC_SMALL_FAST_MODEL": "MiniMax-M2.7-highspeed",
      "ANTHROPIC_DEFAULT_SONNET_MODEL": "MiniMax-M2.7-highspeed",
      "ANTHROPIC_DEFAULT_OPUS_MODEL": "MiniMax-M2.7-highspeed",
      "ANTHROPIC_DEFAULT_HAIKU_MODEL": "MiniMax-M2.7-highspeed"
    }
  }

ah 'almost' . i want to use codex.

Well, then something's wrong. I click on different pages in the documentation and the whole page gets rerendered. Seems like it's not delivering what's promised.

I did it correctly with a mouse and was rejected. Then I drew a straight line and was rejected. Then I closed the tab.

It could even be both % and offline hours per year. To me the percentage is simpler to understand.

Yeah, I hate this trend. AWS did it too with CDK: https://constructs.dev/


Beware: this might totally freeze your computer like it did for me.


Runs fine on my iphone


WebGPU moment (have same issue on Firefox/Linux).


Works fine on my phone, Firefox+GrapheneOS.


Had to break out Chromium for this one - Firefox+Linux does not like webgpu (my whole DE started flickering).


I was amazed that it run smoothly on Firefox mac without WebGPU.


Yeah, it seems fine on my iPhone 13 running Safari 18. It's not warming up.

Some ball shadows look kind of grainy but moving my finger around moves the balls around.


Works great on my M3 MacBook Air under Safari. GPU core temps got into the 130-160 degrees F. range. Fun demo!


definitely needs a lot of computing power


Runs smoothly and without crashes on my Pixel phone


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: