WeSearch

Beyond Memorization: Do Larger Models Know More, or Just Better?

· 0 reactions · 0 comments · 1 view

Just read 2 papers: 1. Incompressible Knowledge Probes 2. Densing Law of LLMs densing laws suggest for every 3 months you will get a new model that does same things in half the parameter. These ppl (IKPs) argue that better architecture and training methods only affect instruction following, reasoning abilities and stuff but not factual knowledge(this is still parameter dependent and scaling rules still apply). This kinda leads to the question like how much factual info is enough? like we can do

Original article
OpenAI
Read full at OpenAI →
Anonymous · no account needed
Share 𝕏 Facebook Reddit LinkedIn Threads WhatsApp Bluesky Mastodon Email

Discussion

0 comments

More from OpenAI