The NVIDIA Digits is a $3,000 personal AI supercomputer designed to lower the barrier for developers working on large models. It features the GB10 Grace Blackwell Superchip, can handle models with up to 200 billion parameters, and includes 128GB of coherent memory and 4TB of NVMe storage. It offers up to one petaflop of AI performance at FP4, making it a powerful tool for AI development on a local machine.
Meta removed AI character accounts after users criticized them as 'creepy and unnecessary.' The AI characters, part of a test, were managed by people but faced backlash for their perceived lack of authenticity. Meta cited a bug that affected users' ability to block these accounts as the reason for their removal.
NVIDIA is focusing on custom chip manufacturing to compete with companies like Broadcom, which designs custom chips for AI applications. By establishing an R&D center in Taiwan and recruiting Taiwanese engineers, NVIDIA aims to develop ASIC (Application-Specific Integrated Circuit) solutions tailored to specific AI workloads, reducing reliance on off-the-shelf GPUs and improving efficiency for AI developers.
OpenAI is delaying the launch of AI agents due to concerns about prompt injection attacks, where malicious inputs could bypass the model's restrictions. Agents, which can interact with the web and sensitive infrastructure, pose a higher risk if compromised. OpenAI is working to mitigate these risks before releasing the agents to the public.
PRIME is a novel approach to online reinforcement learning that uses process rewards to improve the reasoning abilities of AI models. It involves generating diverse solutions to problems, filtering out incorrect answers, and rewarding the most efficient and correct reasoning traces. This method has shown significant improvements in benchmarks, such as the Math Olympiad, by encouraging models to explore new solutions while maintaining accuracy.
The ICLR paper found that language models shift from pre-trained semantic representations to context-aligned ones when given structured tasks. By using a graph-tracing approach, the study showed that models adapt their internal representations based on the context of the input sequence. This suggests that models can dynamically adjust the meaning of words based on their usage in specific contexts, which has implications for jailbreaks and adversarial attacks.
METAGENE-1 is a foundation model trained on metagenomic sequences, which are short DNA fragments from environmental samples like sewage. The model is designed to detect pathogens and disease indicators cost-effectively. By analyzing these sequences, it can provide early warnings of pandemics and other health threats, making it a valuable tool for public health monitoring.
TransPixar is designed to improve text-to-video generation by adding transparency (alpha channel) to video outputs. This allows for more realistic special effects, such as explosions or overlays, by enabling the model to predict both the RGB and alpha channels simultaneously. The model was trained on a dataset of high-resolution green screen videos and has shown significant improvements in video quality and motion alignment.
The growth in training compute for AI models is driven by three main factors: an increase in hardware quantity (doubling annually since 2018), longer training durations (1.5x per year since 2022), and improvements in hardware performance (more flops per GPU). These factors together have contributed to a 4.2x annual growth in training compute since 2018.
InfAlign is an approach to language model alignment that accounts for inference-time scaling, where models generate multiple outputs and select the best one. Traditional alignment methods, like RLHF, don't account for this process, leading to misalignment. InfAlign uses a positive exponential transformation of rewards to prioritize the best outputs, ensuring that the model's alignment is consistent with its usage during inference.
Our 196th episode with a summary and discussion of last week's* big AI news!
*and sometimes last last week's
Recorded on 01/10/2024
Join our brand new Discord here! https://discord.gg/wDQkratW
Hosted by Andrey Kurenkov and Jeremie Harris.
Feel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai
Read out our text newsletter and comment on the podcast at https://lastweekin.ai/.
Sponsors:
In this episode:
- Nvidia announced a $3,000 personal AI supercomputer called Digits, featuring the GB10 Grace Blackwell Superchip, aiming to lower the barrier for developers working on large models.
- The U.S. Department of Justice finalizes a rule restricting the transmission of specific data types to countries of concern, including China and Russia, under executive order 14117.
- Meta allegedly trained Llama on pirated content from LibGen, with internal concerns about the legality confirmed through court filings.
- Microsoft paused construction on a section of a large data center project in Wisconsin to reassess based on new technological changes.
If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.
Timestamps + Links: