Superalignment Fast Grants
We’re launching $10M in grants to support technical research towards the alignment and safety of superhuman AI systems, including weak-to-strong generalization, interpretability, scalable oversight, and more.
We’re launching $10M in grants to support technical research towards the alignment and safety of superhuman AI systems, including weak-to-strong generalization, interpretability, scalable oversight, and more.
Artificial intelligence is advancing rapidly, but researchers are facing a significant challenge. AI systems struggle to adapt to diverse environments outside their training data, which is critical in areas like self-driving cars, where failures can have catastrophic consequences. Despite efforts by researchers to tackle this problem with algorithms for domain generalization, no algorithm has yet…
We are launching a new generation of embedding models, new GPT-4 Turbo and moderation models, new API usage management tools, and soon, lower pricing on GPT-3.5 Turbo.
Diffusion models are at the forefront of generative model research. These models, essential in replicating complex data distributions, have shown remarkable success in various applications, notably in generating intricate and realistic images. They establish a stochastic process that progressively adds noise to data, followed by a learned reversal of this process to create new data…
We’re testing the ability for ChatGPT to remember things you discuss to make future chats more helpful. You’re in control of ChatGPT’s memory.
Welcome Interested in sponsorship opportunities? Join the AI conversation and transform your advertising strategy with AI weekly sponsorship aiweekly.co In the News The 15 biggest announcements at Google I/O 2025 Google just wrapped up its big keynote at I/O 2025. As expected, it was full of AI-related announcements, ranging from updates across Google’s image and…
Large language models (LLMs) have revolutionized various AI-infused applications, from chat models to autonomous driving. This evolution has spurred the need for systems that can efficiently deploy and serve these models, especially under the increasing demand for handling long-prompt workloads. The major hurdle in this domain has been balancing high throughput and low latency in…