discord

Announcing Discord’s Social SDK, Helping Power Your Game’s Social Experiences (opens in new tab)

Discord has announced the release of its new Social SDK at the annual Game Developer’s Conference, offering a free toolkit for developers to integrate social features directly into their titles. This initiative aims to bridge the gap between in-game activity and the Discord platform, fostering better player coordination and community engagement. By providing these tools at no cost, Discord intends to empower developers of all sizes to enhance the social layer of their multiplayer experiences. ### The Discord Social SDK * The SDK is available immediately for download and implementation at no cost to game developers. * It is designed to be accessible to studios of all sizes, allowing for the direct embedding of Discord’s social infrastructure into various game environments. * The toolkit focuses on reducing friction for players who want to share gameplay experiences or coordinate with their social circles without leaving the game. ### Cross-Platform Communication Features * In-game players can communicate directly with friends and teammates on Discord to coordinate sessions in real-time. * The SDK supports bidirectional communication, allowing Discord users to talk to players currently inside a game. * A key technical highlight is the ability for in-game players to interact with Discord users even if the player does not have a registered Discord account themselves. By implementing the Social SDK, developers can transform their games into more connected environments that leverage Discord's massive user base. This integration simplifies the multiplayer experience by removing traditional barriers to communication and providing a universal social layer across different gaming platforms.

google

Deciphering language processing in the human brain through LLM representations (opens in new tab)

Recent research by Google Research and collaborating universities indicates that Large Language Models (LLMs) process natural language through internal representations that closely mirror neural activity in the human brain. By comparing intracranial recordings from spontaneous conversations with the internal embeddings of the Whisper speech-to-text model, the study found a high degree of linear alignment between artificial and biological language processing. These findings suggest that the statistical structures learned by LLMs via next-word prediction provide a viable computational framework for understanding how humans comprehend and produce speech. ## Mapping LLM Embeddings to Brain Activity * Researchers utilized intracranial electrodes to record neural signals during real-world, free-flowing conversations. * The study compared neural activity against two distinct types of embeddings from the Transformer-based Whisper model: "speech embeddings" from the model’s encoder and "language embeddings" from the decoder. * A linear transformation was used to predict brain signals based on these embeddings, revealing that LLMs and the human brain share similar multidimensional spaces for coding linguistic information. * The alignment suggests that human language processing may rely more on statistical structures and contextual embeddings rather than traditional symbolic rules or syntactic parts of speech. ## Neural Sequences in Speech Comprehension * When a subject listens to speech, the brain follows a specific chronological sequence that aligns with model representations. * Initially, speech embeddings predict cortical activity in the superior temporal gyrus (STG), which is responsible for processing auditory speech sounds. * A few hundred milliseconds later, language embeddings predict activity in Broca’s area (located in the inferior frontal gyrus), marking the transition from sound perception to decoding meaning. ## Reversed Dynamics in Speech Production * During speech production, the neural sequence is reversed, beginning approximately 500 milliseconds before a word is articulated. * Processing starts in Broca’s area, where language embeddings predict activity as the brain plans the semantic content of the utterance. * This is followed by activity in the motor cortex (MC), aligned with speech embeddings, as the brain prepares the physical articulatory movements. * Finally, after articulation, speech embeddings predict activity back in the STG, suggesting the brain is monitoring the sound of the speaker's own voice. This research validates the use of LLMs as powerful predictive tools for neuroscience, offering a new lens through which to study the temporal and spatial dynamics of human communication. By bridging the gap between artificial intelligence and cognitive biology, researchers can better model how the brain integrates sound and meaning in real-time.

google

Load balancing with random job arrivals (opens in new tab)

Research from Google explores the competitive ratio of online load balancing when tasks arrive in a uniformly random order rather than an adversarial one. By analyzing a "tree balancing game" where edges must be oriented to minimize node indegree, the authors demonstrate that random arrival sequences still impose significant mathematical limitations on deterministic algorithms. The study ultimately concludes that no online algorithm can achieve a competitive ratio significantly better than $\sqrt{\log n}$, establishing new theoretical boundaries for efficient cluster management. ### The Online Load Balancing Challenge * Modern cluster management systems, such as Google’s Borg, must distribute hundreds of thousands of jobs across machines to maximize utilization and minimize the maximum load (makespan). * In the online version of this problem, jobs arrive one-by-one, and the system must assign them immediately without knowing what future jobs will look like. * Traditionally, these algorithms are evaluated using "competitive analysis," comparing the performance of an online algorithm against an optimal offline version that has full knowledge of the job sequence. ### The Tree Balancing Game * The problem is modeled as a game where an adversary presents edges of a tree (representing jobs and machines) one at a time. * For every undirected edge $(u, v)$ presented, the algorithm must choose an orientation ($u \to v$ or $v \to u$), with the goal of minimizing the maximum number of edges pointing at any single node. * In a worst-case adversarial arrival order, it has been mathematically proven since the 1990s that no deterministic algorithm can guarantee a maximum indegree of less than $\log n$, where $n$ is the number of nodes. ### Performance Under Random Arrival Orders * The research specifically investigates "random order arrivals," where every possible permutation of the job sequence is equally likely, simulating a more natural distribution than a malicious adversary. * While previous assumptions suggested that a simple "greedy algorithm" (assigning the job to the machine with the currently lower load) performed better in this model, this research proves a new, stricter lower bound. * The authors demonstrate that even with random arrivals, any online algorithm will still incur a maximum load proportional to at least $\sqrt{\log n}$. * For more general load balancing scenarios beyond simple trees, the researchers established a lower bound of $\sqrt{\log \log n}$. ### Practical Implications These findings suggest that while random job arrival provides a slight performance advantage over adversarial scenarios, system designers cannot rely on randomness alone to eliminate load imbalances. Because the maximum load grows predictably according to the $\sqrt{\log n}$ limit, large-scale systems must be architected to handle this inherent logarithmic growth in resource pressure to maintain high utilization and stability.

google

Loss of Pulse Detection on the Google Pixel Watch 3 (opens in new tab)

Google Research has developed a "Loss of Pulse Detection" feature for the Pixel Watch 3 to address the high mortality rates associated with unwitnessed out-of-hospital cardiac arrests (OHCA). By utilizing a multimodal algorithm that combines photoplethysmography (PPG) and accelerometer data, the device can automatically identify the transition to a pulseless state and contact emergency services. This innovation aims to transform unwitnessed medical emergencies into functionally witnessed ones, potentially increasing survival rates by ensuring timely intervention. ### The Impact of Witness Status on Survival * Unwitnessed cardiac arrests currently face a major public health challenge, with survival rates as low as 4% compared to 20% for witnessed events. * The "Chain of Survival" traditionally relies on human bystanders to activate emergency responses, leaving those alone at a significant disadvantage. * Every minute without resuscitation decreases the chance of survival by 7–10%, making rapid detection the most critical factor in prognosis. * Converting an unwitnessed event into a "functionally witnessed" one via a wearable device could equate to a number needed to treat (NNT) of only six people to save one life. ### Multimodal Detection and the Three-Gate Process * The system uses PPG sensors to measure blood pulsatility by detecting photons backscattered by tissue at green and infrared wavelengths. * To prevent false positives and errant emergency calls, the algorithm must pass three sequential "gates" before making a classification. * **Gate 1:** Detects a sudden, significant drop in the alternating current (AC) component of the green PPG signal, which suggests a transition from a pulsatile to a pulseless state, paired with physical stillness. * **Gate 2:** Employs a machine learning algorithm trained on diverse user data to quantify the probability of a true pulseless transition. * **Gate 3:** Conducts additional sensor checks using various LED and photodiode geometries, wavelengths, and gain settings to confirm the absence of even a weak pulse. ### On-Device Processing and User Verification * All data processing occurs entirely on the watch to maintain user privacy, consistent with Google’s established health data policies. * If the algorithm detects a loss of pulse, it initiates two check-in prompts involving haptic, visual, and audio notifications to assess user responsiveness. * The process can be de-escalated immediately if the user moves their arm purposefully, ensuring that emergency services are only contacted during true incapacitation. * When a user remains unresponsive, the watch automatically contacts emergency services to provide the individual's current location and medical situation. By providing a passive, opportunistic monitoring system on a mass-market wearable, this technology offers a critical safety net for individuals at risk of unwitnessed cardiac events. For the broader population, the Pixel Watch 3 serves as a life-saving tool that bridges the gap between a sudden medical emergency and the arrival of professional responders.

google

Generating synthetic data with differentially private LLM inference (opens in new tab)

Researchers at Google have developed an inference-only method for generating differentially private (DP) synthetic data that avoids the high costs and data requirements associated with private fine-tuning. By prompting off-the-shelf large language models (LLMs) with sensitive examples in parallel and aggregating their outputs, the approach can generate thousands of high-quality synthetic data points while maintaining rigorous privacy guarantees. This method allows synthetic data to serve as a secure interface for model development, enabling teams to collaborate without requiring specialized knowledge of differential privacy. ## Differentially Private Prediction and Aggregation The core of this method relies on "private prediction," where privacy is applied to the model's output rather than the model itself. * Sensitive data points are distributed across multiple independent prompts, ensuring that no single individual's record can significantly influence the final output. * The LLM generates next-token predictions for each prompt in parallel, which are then aggregated to mask individual contributions. * The researchers designed a DP token sampling algorithm that treats the standard LLM "softmax" sampling process as a version of the exponential mechanism, a mathematical framework used to select the best option from a set while maintaining privacy. ## Enhancing Efficiency via KV Caching Previous attempts at private prediction were computationally expensive because they required a fresh batch of sensitive examples for every single token generated. * A new privacy analysis allows the system to reuse a fixed batch of sensitive examples across an entire generation sequence. * By maintaining the same context for each generation step, the system becomes compatible with standard inference optimization techniques like KV (Key-Value) caching. * This improvement enables the generation of synthetic data at a scale two to three orders of magnitude larger than prior methods. ## Optimizing Privacy Spend with Public Drafters To preserve the "privacy budget"—the limited amount of information that can be released before privacy is compromised—the method introduces a public drafter model. * The drafter model predicts the next token based solely on previously generated synthetic text, without ever seeing the sensitive data. * Using the sparse vector technique, the system only consumes the privacy budget when the public drafter’s suggestion disagrees with the private aggregate of the sensitive data. * This is particularly useful for structured data, where the drafter can handle formatting and syntax tokens, saving the privacy budget for the actual content. By leveraging off-the-shelf models like Gemma, this approach provides a scalable way to transform sensitive datasets into useful synthetic versions. These synthetic datasets are high-quality enough to replace real data in downstream machine learning tasks, such as in-context learning or fine-tuning models like BERT, without the risk of leaking individual user information.

discord

The Game Developer Playbook: Three Incredible Game-Focused Communities (opens in new tab)

Discord suggests that there is no universal set of "best practices" for managing a community, as every server has unique requirements that demand creative, tailored solutions. By analyzing high-performing communities through Server Discovery, developers can find inspiration to iterate on their own structures and engagement strategies. This approach emphasizes that true success is measured by active engagement rather than raw member counts. ## Leveraging Server Discovery for Community Inspiration * Server Discovery serves as a primary research tool for developers to observe how successful communities are structured and operated. * Success should be evaluated based on engagement levels rather than the total number of members in a community. * Research should extend beyond featured pages to include servers with similar topics or even non-gaming communities to find adaptable ideas. ## Analyzing Best-in-Class Multiplayer Communities * The guide highlights three specific communities—Fortnite, Rocket League, and Deep Rock Galactic—as models for effective server architecture. * Each of these servers demonstrates how to align technical setups with specific marketing goals and community needs. * The analysis goes beyond basic configurations, exploring advanced implementations of categories, channels, and permission systems to enhance user experience. To build a comprehensive community strategy, developers should review these case studies in conjunction with earlier phases of the GameDev Playbook, specifically focusing on the transitions from private playtest environments to early access and pre-launch configurations.

discord

Modern Image Formats at Discord: Supporting WebP and AVIF (opens in new tab)

Discord’s Media Infrastructure team has modernized its image pipeline by integrating support for animated WebP and AVIF formats across its entire platform. This update ensures that attachments, embeds, and animated emojis are delivered using high-efficiency codecs that maintain visual fidelity while minimizing resource consumption. By adopting these modern standards, Discord has optimized the balance between high-quality animation and fast, cross-platform playback performance. **Expansion of Format Support** * Native support for animated WebP and AVIF has been added for all user-generated attachments and embedded media. * The platform has transitioned all animated emojis to serve as animated WebP, ensuring a consistent viewing experience across desktop, web, and mobile clients. * The pipeline update allows for a unified delivery system that handles modern media containers seamlessly across different device architectures. **Performance and Infrastructure Benefits** * The shift to modern codecs has resulted in a drastic decrease in file sizes, which directly translates to faster loading times for users on limited bandwidth. * Enhanced playback performance reduces the computational overhead required to render complex animations. * Higher visual quality is maintained even at lower bitrates, allowing for richer media experiences without increasing data costs. For developers and platforms managing high-volume media assets, transitioning to animated WebP and AVIF represents a necessary evolution to meet modern performance expectations. Implementing these formats provides a scalable way to deliver high-fidelity content while significantly reducing storage and bandwidth requirements.