New ask Hacker News story: Ask HN: Any informed guesses on the actual size/architecture of GPT-5.4 etc.?
Ask HN: Any informed guesses on the actual size/architecture of GPT-5.4 etc.? 2 by dsrtslnd23 | 0 comments on Hacker News. Does anyone have decent intuitions or hard clues on how big models like GPT-5.4, Gemini 3.1, and Opus 4.6 actually are, and how they compare to the best open models like GLM-5? Are they all roughly in the same range now (for example around 1T params, maybe MoE), or are the closed models still much bigger? Also curious about “pro” versions like GPT-5.4 Pro - is that likely a different model, or mostly the same model with more inference-time compute / longer reasoning / better orchestration?