It was thrilling to watch AI compete against some of the best human competitive programmers at AtCoder World Finals Heuristics yesterday. Check out @andresnds ‘s thread on how the AI solutions improved throughout the 10h contest. Congrats to @FakePsyho on 1st place!
It was thrilling to watch AI compete against some of the best human competitive programmers at AtCoder World Finals Heuristics yesterday. Check out @andresnds ‘s thread on how the AI solutions improved throughout the 10h contest. Congrats to @FakePsyho on 1st place!
Two important points from our new technical report:
1. Scaling continues to work and the bitter lesson still holds
2. Recent AI models are strong at reasoning tasks and are rapidly becoming stronger — 4o was released less than a year ago, o1 less than six months ago
i generally feel super grateful that i get to work with such exceptionally skilled and kind people on reasoning research. the sprint for IOI in particular was special though. IOI 2024 gold @ 10k submissions; 49th percentile of competitors under real contest conditions
i generally feel super grateful that i get to work with such exceptionally skilled and kind people on reasoning research. the sprint for IOI in particular was special though. IOI 2024 gold @ 10k submissions; 49th percentile of competitors under real contest conditions
Today, I’m excited to share with you all the fruit of our effort at @OpenAI to create AI models capable of truly general reasoning: OpenAI's new o1 model series! (aka 🍓) Let me explain 🧵 1/
But the ELO can ultimately become bounded by the difficulty of the prompts (i.e. can’t achieve arbitrarily high win rates on the prompt: “what’s up”). We find on harder prompt sets — and in particular coding — there is an even larger gap: GPT-4o achieves a +100 ELO over our prior…
new preprint
"ReLU to the Rescue: Improve your On-policy Actor-Critic with Positive Advantages"
shockingly simple changes to A3C can give a cautious RL algorithm more effective than PPO
in some settings, just adding a ReLU is enough!
arxiv.org/abs/2306.01460
The Google DeepMind alignment team is looking for research scientists and research engineers to help us work towards safe AGI.
I think this is a very pressing problem, and it's a nice place to work. Please apply and help take our work to the next level.
boards.greenhouse.io/deepmind/jobs/…
With more powerful AI systems comes more responsibility to identify novel capabilities in models. 🔍
Our new research looks at evaluating future 𝘦𝘹𝘵𝘳𝘦𝘮𝘦 risks, which may cause harm through misuse or misalignment.
Here’s a snapshot of the work. 🧵 dpmd.ai/novel-ai-risks
1.9M Followers 27K FollowingYes, I can see some risk that your threat to jail Internet company executives for not censorsing aggressively enough could backfire.
443 Followers 283 FollowingCognitive Neuroscientist @ Harvard, AI Researcher @ Motional
Models of human & robot decision making in complex environments, e.g. video games & urban driving
662 Followers 729 FollowingUp scaling my SaaS current MRR - 6K$/m | UG @ BitsPilani | ex Machine Alchemist @ InfotechSolutions |
I like to read research papers.
1K Followers 1K FollowingDirector, Core AI, IBM. Chief Architect https://t.co/lo3JOcuDFA . Founder, Red Hat AI Innovation Team. PI @MITIBMLab. ❤️ Density Ratios.
6K Followers 696 FollowingRL Scientist @OpenAI. Prev. co-founder @diffeo, acquired by @salesforce // co-authored The Principles of Deep Learning Theory // studied gravity.
374K Followers 711 FollowingCharles W. Eliot University Professor and President Emeritus at Harvard. Secretary of the Treasury for President Clinton and Director of NEC for President Obama
85K Followers 1K Followingi help make https://t.co/jZh799yNH4, the best AI for self-improvement, introspection, and emotional processing. https://t.co/ac0cp4UZ9h
No recent Favorites. New Favorites will appear here.