Shakahs

๐Ÿ“… Joined in 2013

๐Ÿ”ผ 592 Karma

โœ๏ธ 106 posts

๐ŸŒ€
15 latest posts

Load

(Replying to PARENT post)

I use it via the Kilo Code extension for VSCode, which is invoking Qwen3-Coder via a Cerebras Code subscription.

https://github.com/Kilo-Org/kilocode https://www.cerebras.ai/blog/introducing-cerebras-code

๐Ÿ‘คShakahs๐Ÿ•‘1mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Sonnet/Claude Code may technically be "smarter", but Qwen3-Coder on Cerebras is often more productive for me because it's just so incredibly fast. Even if it takes more LLM calls to complete a task, those calls are all happening in a fraction of the time.
๐Ÿ‘คShakahs๐Ÿ•‘1mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

SlateDB offers different durability levels for writes. By default writes are buffered locally and flushed to S3 when the buffer is full or the client invokes flush().

https://slatedb.io/docs/design/writes/

๐Ÿ‘คShakahs๐Ÿ•‘1mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

While your technical analysis is excellent, making judgements about workload suitability based on a Preview release is premature. Preview services have historically had significantly lower performance quotas than GA releases. Lambda for example was limited to 50 concurrent executions during Preview, raised to 100 at GA, and now the default limit is 1,000.
๐Ÿ‘คShakahs๐Ÿ•‘1mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Grok are the first models I am boycotting on purely environmental grounds. They built their datacenter without sufficient local power supply and have been illegally powering it with unpermitted gas turbine generators until that capacity gets built, to the significant detriment of the local population.

https://www.datacenterdynamics.com/en/news/elon-musk-xai-gas...

๐Ÿ‘คShakahs๐Ÿ•‘2mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Reading about mainframes feels very much like reading science fiction. Truly awesome technology that exists on a completely different plane of computing than anything else.
๐Ÿ‘คShakahs๐Ÿ•‘6mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

This thinly veiled advertisement claims it's a waste of time to understand the tradeoffs in the models you're using, and you should instead pay them to make those decisions for you. No thank you.
๐Ÿ‘คShakahs๐Ÿ•‘7mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Forcing Twitter to make public posts easily accessible is not at all the same as compelling Apple to hand over your private messages.
๐Ÿ‘คShakahs๐Ÿ•‘8mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Per the ongoing Freedesktop discussion, AWS offered to host but Freedesktop is leaning towards self-hosting on Hetzner so they can control their own destiny and sponsors can contribute cash towards the bill instead of donating hardware.

> https://gitlab.freedesktop.org/freedesktop/freedesktop/-/iss...

๐Ÿ‘คShakahs๐Ÿ•‘8mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Same for me, I use "ar <tab>" in Intellij all the time.
๐Ÿ‘คShakahs๐Ÿ•‘8mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Because they own the CDN and most of the bandwidth is from peering, so it essentially costs them nothing. Netlify on the other hand has to pay per GB to AWS.
๐Ÿ‘คShakahs๐Ÿ•‘9mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Note: This is type stripping only, no transpilation / code generation is done.
๐Ÿ‘คShakahs๐Ÿ•‘9mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0
๐Ÿ‘คShakahs๐Ÿ•‘9mo๐Ÿ”ผ17๐Ÿ—จ๏ธ7

(Replying to PARENT post)

Both these issues can be resolved by adding some sample code to context to influence the LLM to do the desired thing.

As the op says, LLMs are going to be biased towards doing the "average" thing based on their training data. There's more old backend code on the internet than new backend code, and Tailwind is pretty dominant for frontend styling these days, so that's where the average lands.

๐Ÿ‘คShakahs๐Ÿ•‘9mo๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Counterpoint: I also used to fix Lexmark printers as a field tech servicing pharmacies. It was routine to see Lexmark MS711dn printers with page counts in the millions. They did not need more than basic maintenance.
๐Ÿ‘คShakahs๐Ÿ•‘1y๐Ÿ”ผ0๐Ÿ—จ๏ธ0