Hacker Newsnew | past | comments | ask | show | jobs | submit | billylo's commentslogin

We learned computer languages so we can ask computers to do work for us. It was out of necessity because there are no other ways.

If we can instruct computers with natural language 50% of the time, that's 50% less translation work for our human brains. I have no problem with not needing to write instructions in computer languages (no regex, no sed/awk, no python even) for day-to-day stuff.

Critical thinking, reasoning are another story. We can't let those skills atropied.


It's a strange new world to get used to.

Time it takes to go from 100m users to 5 billions: Internet (25 years), Smartphone (13), AI? (tracking to achieve that in ~6 years!)

https://evergreen-labs.org/assets/Acceleration.jpg


We should ask how the traders manage this. It's essentially 24/7 markets in the world. For them, the FOMO effects are even stronger... actual money earning opportunity.

That's why benchmarks are useful. We all suffer from the shortcomings of human perception.


Benchmarks shortcomings are no worse... they inevitably measure something that is only close to the thing you actually care about, not the thing you actually care about. It's entirely plausible that this decreased benchmark score is because Anthropic's initial prompting of the model was overtuned to the benchmark and as they're gaining more experience with real world use they are changing the prompt to do better at that and consequentially worse at the benchmark.


I wonder how best we can measure the usefulness of models going forward.

Thumbs up or down? (could be useful for trends) Usage growth from the same user over time? (as an approximation) Tone of user responses? (Don't do this... this is the wrong path... etc.)


Benchmarks measure what they measure. But your subjective experience also matters.


xcode has been getting better bit-by-bit. No major regression.


Windows and macOS does come with a small model for generating text completion. You can write a wrapper for your own TUI to access them platform agnostically.

For consistent LLM behaviour, you can use ollama api with your model of choice to generate. https://docs.ollama.com/api/generate

Chrome has a built-in Gemini Nano too. But there isn't an official way to use it outside chrome yet.


Do you know what it’s called, at least on Windows? I’m struggling to find API docs.

When I asked AI it said no such inbuilt model exists (possibly a knowledge date cutoff issue.)



Thankyou!


Yes. I am not aware of a model shipping with Windows nor announced plans to do so. Microsoft’s been focused on cloud based LLM services.


This thread is full of hallucinations ;)


These are the on-device model APIs for apple: https://developer.apple.com/documentation/foundationmodels


Is there a Linux-y standard brewing?


Each distro is doing their own thing. If you are targeting Linux mainly, I would suggest to code it on top of ollama or LiteLLM


Windows doesn't?


Please let us know when and which LLM changes its "minds". This is a cool experiment. I wish there are more time-bound datasets that we can experiment with to get a better sense on how LLMs are influenced.


I did a fact check on the article and it seems to check out. I am happy to help billy@evergreen-labs.org.

I'd also suggest to reach out to the author of the NYT article and look for ideas. They took the time to study the subject and will likely have some insights on what could work (tech or non-tech approaches.)


Same here. Agents has allowed me to take on more experiments because the cost for testing ideas is now much much lower.


Compared to moving to assembly coding to high level languages, which change do you think is more dramatic? I am curious.


If you are curious about doing something similar with TPU, Google has an article. https://developers.googleblog.com/train-gpt2-model-with-jax-...


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: