AI Unfiltered

Chinese AI • Open Source • Security • Incidents. Signal, not noise.

Gemini 3 Flash: frontier intelligence built for speed

Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost.

This Nobel Prize–winning chemist dreams of making water from thin air

Omar Yaghi was a quiet child, diligent, unlikely to roughhouse with his nine siblings. So when he was old enough, his parents tasked him with one of the family’s most vital chores: fetching water. Like most homes in his Palestinian neighborhood in Amman, Jordan, the Yaghis’ had no electricity or...

Leveraging LLMs for Structured Data Extraction from Unstructured Patient Records

arXiv:2512.13700v1 Announce Type: new Abstract: Manual chart review remains an extremely time-consuming and resource-intensive component of clinical research, requiring experts to extract often complex information from unstructured electronic health record (EHR) narratives. We present a secure,...

Blind Radio Mapping via Spatially Regularized Bayesian Trajectory Inference

arXiv:2512.13701v1 Announce Type: new Abstract: Radio maps enable intelligent wireless applications by capturing the spatial distribution of channel characteristics. However, conventional construction methods demand extensive location-labeled data, which are costly and impractical in many...

Adjudicator: Correcting Noisy Labels with a KG-Informed Council of LLM Agents

arXiv:2512.13704v1 Announce Type: new Abstract: The performance of production machine learning systems is fundamentally limited by the quality of their training data. In high-stakes industrial applications, noisy labels can degrade performance and erode user trust. This paper presents Adjudicator,...

LoopBench: Discovering Emergent Symmetry Breaking Strategies with LLM Swarms

arXiv:2512.13713v1 Announce Type: new Abstract: Large Language Models (LLMs) are increasingly being utilized as autonomous agents, yet their ability to coordinate in distributed systems remains poorly understood. We introduce \textbf{LoopBench}, a benchmark to evaluate LLM reasoning in distributed...

AI-Powered Annotation Pipelines for Stabilizing Large Language Models: A Human-AI Synergy Approach

arXiv:2512.13714v1 Announce Type: new Abstract: LLM implementations are failing in highly regulated industries owing to instability issues, inconsistent reasoning, hallucinations and performance variability, especially in workflows. These reliability issues restrict safe use of LLM in areas that...

Physics-Guided Deep Learning for Heat Pump Stress Detection: A Comprehensive Analysis on When2Heat Dataset

arXiv:2512.13696v1 Announce Type: new Abstract: Heat pump systems are critical components in modern energy-efficient buildings, yet their operational stress detection remains challenging due to complex thermodynamic interactions and limited real-world data. This paper presents a novel...

Scaling and Transferability of Annealing Strategies in Large Language Model Training

arXiv:2512.13705v1 Announce Type: new Abstract: Learning rate scheduling is crucial for training large language models, yet understanding the optimal annealing strategies across different model configurations remains challenging. In this work, we investigate the transferability of annealing...

Mitigating Catastrophic Forgetting in Mathematical Reasoning Finetuning through Mixed Training

arXiv:2512.13706v1 Announce Type: new Abstract: When finetuning large language models for specialized tasks such as mathematical reasoning, models exhibit catastrophic forgetting, losing previously learned capabilities. We investigate this by finetuning Flan-T5-Base (250M parameters) on the...

Variational Physics-Informed Ansatz for Reconstructing Hidden Interaction Networks from Steady States

arXiv:2512.13708v1 Announce Type: new Abstract: The interaction structure of a complex dynamical system governs its collective behavior, yet existing reconstruction methods struggle with nonlinear, heterogeneous, and higher-order couplings, especially when only steady states are observable. We...

Predictive Modeling of Flood-Prone Areas Using SAR and Environmental Variables

arXiv:2512.13710v1 Announce Type: new Abstract: Flooding is one of the most destructive natural hazards worldwide, posing serious risks to ecosystems, infrastructure, and human livelihoods. This study combines Synthetic Aperture Radar (SAR) imagery with environmental and hydrological data to model...

FiNERweb: Datasets and Artifacts for Scalable Multilingual Named Entity Recognition

arXiv:2512.13884v1 Announce Type: new Abstract: Recent multilingual named entity recognition (NER) work has shown that large language models (LLMs) can provide effective synthetic supervision, yet such datasets have mostly appeared as by-products of broader experiments rather than as systematic,...

Olmo 3

arXiv:2512.13961v1 Announce Type: new Abstract: We introduce Olmo 3, a family of state-of-the-art, fully-open language models at the 7B and 32B parameter scales. Olmo 3 model construction targets long-context reasoning, function calling, coding, instruction following, general chat, and knowledge...

Structure-Aware Decoding Mechanisms for Complex Entity Extraction with Large-Scale Language Models

arXiv:2512.13980v1 Announce Type: new Abstract: This paper proposes a structure-aware decoding method based on large language models to address the difficulty of traditional approaches in maintaining both semantic integrity and structural consistency in nested and overlapping entity extraction...

What Affects the Effective Depth of Large Language Models?

arXiv:2512.14064v1 Announce Type: new Abstract: The scaling of large language models (LLMs) emphasizes increasing depth, yet performance gains diminish with added layers. Prior work introduces the concept of "effective depth", arguing that deeper models fail to fully utilize their layers for...

Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in Speed

arXiv:2512.14067v1 Announce Type: new Abstract: Diffusion language models (dLMs) have emerged as a promising paradigm that enables parallel, non-autoregressive generation, but their learning efficiency lags behind that of autoregressive (AR) language models when trained from scratch. To this end,...

Import AI 437: Co-improving AI; RL dreams; AI labels might be annoying

Do you believe the singularity is nigh?

Import AI 436: Another 2GW datacenter; why regulation is scary; how to fight a superintelligence

Is AI balkanization measurable?

Import AI 435: 100k training runs; AI systems absorb human power; intelligence per watt

At what point will AI change your daily life?

Import AI 434: Pragmatic AI personhood; SPACE COMPUTERS; and global government or human extinction;

The future is biomechanical computation

Import AI 433: AI auditors; robot dreams; and software for helping an AI run a lab

Would Alan Turing be surprised?