Hacker Newsnew | past | comments | ask | show | jobs | submit | mks_shuffle's commentslogin

How much Codex and Claude Code are different from each other? I have been using Codex for few weeks doing experiments related to data analysis and training models with some architecture modifications. I wouldn't say I have used it extensively, but so far my experience has been good. Only annoying part has been not able to use GPU in the Codex without using `--sandbox danger-full-access` flag. Today, I started using Claude Code, and ran similar experiments as Codex. I find the interface is quite similar to Codex. However, I hit the limit quite quickly in Claude Code. I will be exploring its features further. I would appreciate if anyone can share their experience of using both tools.

Codex is heavily inspired by Claude Code. They aren't that different. They might diverge more in future.

They are quite different, Claude Code with Opus 4.5 is miles better than Codex

What is it you find makes it much better?

What could be the potential impact on smartphone and tablet prices in the coming months or years? I am assuming that laptop prices will start increasing next year.


I think the parent comment is referring to "Attention is All You Need", famous transformer paper.


And I couldn’t fit any unreasonable effectiveness in my reply, so I had to be contrarian.


Does anyone have insights on the best approaches to compare reasoning models? It is often recommended to use a higher temperature for more creative answers and lower temperature values for more logical and deterministic outputs. However, I am not sure how applicable this advice is for reasoning models. For example, Deepseek-R1 and QwQ-32b recommend a temperature around 0.6, rather than lower values like 0.1–0.3. The Qwen3 blog provides performance comparisons between multiple reasoning models, and I am interested in knowing what configurations they used. However, the paper is not available yet. If anyone has links to papers focused on this topic, please share them here. Also, please feel free to correct me if I’m mistaken about anything. Thanks!


Oh really? Should I adjust the temp to 0,6 on QwA-32B? Where did you get these numbers from?


These are recommendations provided on huggingface page under usage guidelines QwQ-32b: https://huggingface.co/Qwen/QwQ-32B DeepSeek-R1: https://huggingface.co/deepseek-ai/DeepSeek-R1


Can you please elaborate a bit more about your issues with PyTorch on M4 mac. I read PyTorch has some support for Mac GPU with MPS backend, but not sure how extensive it is. I am looking for a new machine, and use of PyTorch and LLM inference are one of the main uses. Sorry for being a bit off-topic from the thread. Thanks.


I couldn't get pytorch to use the mac GPU at all. I didn't spend ages on it - like about a day. In general I find the build environment on mac really annoying compared to say Linux so it may be that with more patience it could be done. IT was also on my work computer and setting up Xcode command line tools on my work mac is super annoying for whatever dumb reason, but basically getting any kind of prebuilt python version (like say the ones that ship with rye) to talk to your Xcode if it's not in the same path as the one that built the python version is annoying, so then you're looking at building the whole of python which is not that hard but (when I tried it) meant it broke the next time Xcode upgraded. I tried symlinks to the sdk etc but that was no bueno for reasons I don't quite remember. And then even having done that and patched pytorch it didn't actually run GPU accelerated.

So bottom line is it might be possible to make it work but in my brief attempt I couldn't get it to work.

If you're at all linux-capable I would say it's hands down a better dev experience in every way (and that's coming from someone who's used macs for years and years in both a professional and private capacity).


Another option with Python is

uploadserver: https://github.com/Densaugeo/uploadserver

python built-in: python -m http.server <port> (does not support upload)


You can try Groq API for faster inference. They use custom hardware to speed up the inference. Supported open models can be found here: https://console.groq.com/docs/models (includes llama-70b)


thanks, tried this to some mixed results. seems like they have caps on speed/rate limits etc if you havent spoken to them so might reach out


For ML/DL papers you can check https://paperswithcode.com/


Do they release yearly/periodic bundled datasets or you will have to use their api to make dataset?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: