Imagine if Cargo was not first-party, but a third-party tool belonging to a vc startup with zero revenue.
Then that startup makes rustup, rustfmt and rust-analyzer.
Great, but I would be more comfortable with the ecosystem if at least the rust-analyzer and rustfmt parts had competitive alternatives.
If you find it works for you, then that’s great! This post is mostly from our learnings from getting it to solve hard problems in complex brownfield codebases where auto generation is almost never sufficient.
I asked it this in a conversation where it referenced my city (I never mentioned it) and it conveniently left out the location in the metadata response, which was shrewd. I started a new conversation and asked the same thing and this time it did include approximate location as "United States" (no mention of city though).
I just tried it out and docling finished in 20s (with pretty good results) the same document which in Tensorlake is still pending for 10 minutes. I won't even wait for the results.
There was an unusual traffic spike around that time, if you try now it should be a lot faster. We were calling up but there was not enough GPU capacity at that time.
LLM provides inputs to your system like any human would, so you have to validate it. Something like pydantic or Django forms are good for this.
reply