The week ahead: 20250929
-
[Announcement] ChatGPT Pulse. ChatGPT can now proactively push personalized content.
-
[Paper] RL’s Razor. KL divergence between trained and base models on the trained task is a strong indicator for forgetting; On-policy training causes less forgetting than SFT because it implicitly favors KL-minimal solutions by sampling from the model’s own distribution.
-
[Paper] Continued Pretraining with agentic abilities in mind.
-
[GitHub] Lots of algorithms in Python.
-
[GitHub] Spec-kit: YANC (Yet Another No-Code).
-
[Paper] Is In-Context Learning learning?
-
[Article] Investment lessons from the past - should one swim upstream or fish downstream?
-
[Blog] Onboarding best practices.
-
[Tweet] David Haber: The best opportunities are usually the hardest to classify.
- [採訪]Plaud的成功:是市場驅動產品,還是產品驅動市場?
Have a great week!
Comments