Hacker Newsnew | past | comments | ask | show | jobs | submit | spyder's commentslogin

District 9 vibes :-)

Nice, but it's weird that no "language" or "English" is mentioned on the github page, and only from the "Release multilingual TTS" Roadmap item could I guess it's probably English only for now.

Using physical analogs for virtual things is not the best choice, for example: Would you give a copy of your bike, or copy of your food to your poor neighbor kid if you could copy it as easily and as cheaply as digital products?


For you... But the results are different for different users.

For me Google shows the .net site first the github one as second.

Asking chatgpt 5.2 (Auto mode) to search for the nanoclaw site, it says the same, first links the .net site and shows the github as an optional page. When I try to give it a hint by asking "are you sure?" it still even hallucinates that it's linked from the github:

"Yes — nanoclaw.net is the official documentation/site for the NanoClaw project, in the sense that it’s the project’s published homepage and is directly linked from its canonical open-source repository. It describes the project, features, installation steps, and links to the source code on GitHub, which is the authoritative source for the project’s codebase."

Chatgpt 5.2 (Thinking mode) and Claude gets it right the first try, they asnwer with the official .dev page first and claude shows the .net second as "another site covering the project".


I was surprised by what you said, so I used a browser that's not logged in to a Google account, to compare. Indeed the fake site ranks #1! Dang!

I guess Google has my account in an autism bucket, so biases GitHub links higher ;)


Yea, even in the case they could match human level stereo depth perception with AI, why would they say "no" to superhuman lidar capabilities. Cost could be a somewhat acceptable answer if there wouldn't be problems with the camera only approach but there are still examples of silly failures of it. And if I remember correctly they also removed their other superhuman radar in their newer models, the one which in certain conditions was capable of sensing multiple cars ahead by bouncing the signal below other cars.


Because they don't have superhuman LIDAR. They never did. Nobody ever did. LIDAR input is not completely reliable so what do you do then?



Not the great answer you think it is.


It all depends on how cheap they can get. And another interesting thought: what if you could stack them? For example you have a base model module, then new ones come out that can work together with the old ones and expanding their capabilities.


It sounds similar, but doesn't sound the same to me. Also how would you determine the similarity allowed? Maybe if we would have such a measure they could use that in voice model training to not allow that much similarity to a single voice, but if we don't have an agreed upon value for that than it's a subjective "sounds the same to me" rule then it's hard to follow that. Ok, they can say that don't train on their voice, but it's very likely that a blend of voices from an "allowed" set could produce a very similar voice to his.


Yea, LLMs have prompt-, harness- and even random seed variability, and it leaves you wonder maybe with a better prompt or system instruction a model could perform better. Too bad most benchmarks don't report that variability, because it could reveal that the model may only perform well if it's prompted in the style of their training data and not generalize well to unseen prompt styles. Also it could explain some of the benchmark vs real world usage gaps.

I remember some papers about earlier models having around 15% prompt variability, and with different tool use sometimes there are even more significant jumps. And if I remember correctly the reasoning models improve some of these because lot of the early prompting tricks is included in them like "thinking step-by-step", "think carefully" and some other "magic" methods. Also another trick is to ask the models to rephrase the prompt with their own words because that may produce prompt that better align with their training prompts. For sure the big model developers are aware of these and constantly improving it, I just don't see too much discussion or numbers about it.


I haven't been able to find it again, but a few years ago I read a paper that found that certain prompts massively improved the performance of some LLMs on benchmarks. But the same prompt massively reduced the performance of some other LLMs. I assume this is still true, though perhaps not as dramatically as before.


with a high speed camera any vibrating reflective object like a potato chips bag can become a weak microphone if you have line of sight even behind a soundproof window: https://www.youtube.com/watch?v=FKXOucXB4a8


Yes it's in a controlled environment not in a real world noisy environment. But this is more stealthy than a camera and could potentially work with non-line-of-sight or even through walls.

And based on that I could imagine with a combination of a camera and this method, you could train the model on data where both the camera and this method is seeing the individual and then continue to track them with the wifi sensing + the trained model even where the camera cannot see them anymore.

But yea real world is noisy, so it could be very challenging.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: