Hacker Newsnew | past | comments | ask | show | jobs | submit | chid's commentslogin

Did I see this on X first?

just bought a pre-release Apple. the value was insane.

interestingly that's what comes up when you search for pi-coding-agent rather than the new domain.


utterly unrelated, the RSP had nothing to do with their usage terms and was entirely about research and release of high-capability models.


Given the high bar of entry 160VRAM GPU - is there anything practical one can use this for?


The model being 32B could run in <20GB VRAM with Q4 quantization (minimal loss of quality), or 80GB unquantized at full fidelity. The quoted 160GB is for their recommended evaluation settings.

There's a few pre-quantized options[0] or you can quantize it yourself with llama.cpp[1]. You can run the resulting gguf with llama.cpp `llama-cli` or `llama-server`, with LM Studio or with Ollama.

0: https://huggingface.co/models?search=cwm%20q4%20gguf

1: https://huggingface.co/spaces/ggml-org/gguf-my-repo


I see, still a fair more VRAM than I have access to. Thanks for sharing that information.


Interesting though one would think this is also an obvious finding.

Quantifying this would be interesting though.


it definitely feels like it.


Took quite a while for it to show up oddly.


I can't think of any other than potentially oil and gas (though they probably use a lot of it in head office type environment).


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: