PAPER PLAINE

Fresh research, simply explained. Updates twice daily.

CLAD: A Clustered Label-Agnostic Federated Learning Framework for Joint Anomaly Detection and Attack Classification

Training security systems across IoT devices without sharing raw data

A new framework called CLAD trains security systems across thousands of IoT devices while keeping data private and handling the reality that most collected data comes without labels. It achieves 30% better detection of network attacks than existing methods while using half the communication bandwidth, even when 80% of the data lacks security labels.

As factories, smart homes, and critical infrastructure rely on millions of connected devices, security breaches can cascade rapidly across networks. CLAD makes it practical for these devices to collectively learn threat patterns without exposing sensitive operational data to central servers, while actually improving detection accuracy by making use of unlabeled data that would otherwise be wasted.

On the (In-)Security of the Shuffling Defense in the Transformer Secure Inference

How shuffling AI model outputs doesn't actually hide them from hackers

A security technique meant to protect AI models during remote computation—shuffling the model's internal activations before revealing them—can be broken for about $1 worth of queries. Researchers show how to align these shuffled values back to their original order, then use them to recover the model's actual weights, demonstrating the attack works on real models like GPT-2.

As AI systems move to cloud computing, companies rely on cryptographic defenses to keep model weights secret while still computing results. This attack shows a widely-used shuffling defense provides a false sense of security—meaning companies using it may think their models are protected when they're actually vulnerable to cheap theft. Developers now need better defenses before deploying sensitive models to untrusted servers.

Unsupervised Denoising of Real Clinical Low Dose Liver CT with Perceptual Attention Networks

Cleaning up blurry CT scans without needing perfect reference images

Researchers developed an artificial intelligence system that removes noise from low-dose CT scans without requiring paired clean images for training—a major obstacle in medical imaging. The system was tested on real clinical scans and validated by radiologists, achieving results comparable to supervised methods while solving the practical problem that hospitals rarely have perfectly clean versions of the same scan to learn from.

Low-dose CT reduces radiation risk to patients, but the grainy images can make tumors and other abnormalities harder to spot, potentially leading to missed diagnoses. This technique cleans up those images automatically using only the noisy scans themselves, making it immediately usable in hospitals without requiring expensive paired training data. Radiologists who reviewed the results confirmed it meets clinical standards, meaning patients could get safer imaging without sacrificing diagnostic clarity.

One Single Hub Text Breaks CLIP: Identifying Vulnerabilities in Cross-Modal Encoders via Hubness

How a single confusing text can fool systems that match images to captions

Researchers found a critical weakness in CLIP and similar image-text matching systems: a single generic piece of text can be artificially close to nearly every image in a dataset, tricking the system into giving it high similarity scores even when it's meaningless. This reveals that these widely-used systems rely on flawed geometry in their internal representation space, making them vulnerable to subtle manipulation.

Image-to-text systems power real applications—from photo search to automated caption evaluation—and companies rely on them to be robust. This vulnerability means a single malicious or accidental hub text could poison search results or break evaluation metrics that measure whether AI-generated captions match human standards, undermining trust in systems used for content moderation, accessibility, and quality assurance.

Defending Quantum Classifiers against Adversarial Perturbations through Quantum Autoencoders

Protecting quantum AI classifiers from sneaky adversarial tricks

Quantum machine learning systems that classify images can be fooled by specially crafted noise, just like regular AI systems. Researchers developed a defense using quantum autoencoders to clean up corrupted data before classification, improving accuracy by up to 68% under attack without needing to retrain the system on known threats.

As quantum computers become practical tools for real tasks, securing them against adversarial attacks matters for any high-stakes application—medical imaging, security screening, or autonomous systems. This defense works without the overhead of constantly retraining on new attack types, making it more practical to deploy when attackers keep changing their tactics.

Strait: Perceiving Priority and Interference in ML Inference Serving

Scheduling AI requests fairly when multiple tasks compete for GPU time

Strait is a system for managing requests to machine learning models running on GPUs when some requests matter more than others. It predicts how long each request will take even when multiple requests run simultaneously, then uses those predictions to prioritize urgent requests—cutting missed deadlines for high-priority tasks by up to 11 percentage points without completely starving lower-priority work.

Companies running AI services on their own hardware often need to handle both time-sensitive requests (like fraud detection) and routine ones (like recommendations) on the same machines. Current systems either guess badly at how long things will take under load or simply interrupt low-priority tasks—wasting GPU power. Strait lets businesses meet their critical deadlines while still processing regular work efficiently, making on-premises AI infrastructure more practical.

Mapping the Phase Diagram of the Vicsek Model with Machine Learning

Using AI to map where flocking behavior switches between chaos and order

Researchers used machine learning to chart the complete phase diagram of the Vicsek model—a mathematical model of how animals flock together—across its full parameter space. By training a neural network on simulated data, they achieved 92% accuracy in predicting when the system transitions between disordered, ordered, and mixed states, and revealed a previously unclear boundary region between ordered and chaotic behavior.

Phase diagrams are critical maps in physics and biology that show where systems behave differently. This machine-learning approach turns expensive simulations into comprehensive maps that can predict behavior across untested regions, potentially accelerating research into real collective motion—from bird flocks to autonomous robot swarms—by replacing exhaustive simulations with trained algorithms.

Explainable Load Forecasting with Covariate-Informed Time Series Foundation Models

Making AI power grid forecasts understandable and trustworthy

Researchers found that advanced AI models can predict electricity demand as accurately as traditional ones while remaining interpretable—a crucial requirement for critical infrastructure. By developing a method to explain which factors (weather, time of day, historical patterns) drive each prediction, they showed that these models reliably use the right information to make decisions, matching established expertise about what actually moves power consumption.

Power grid operators need to understand *why* a forecast says demand will spike before they commit expensive resources. Black-box predictions, no matter how accurate, create operational risk and regulatory friction. This work proves that grid forecasting can be both cutting-edge and transparent, removing a major barrier to deploying faster, more efficient AI systems in electricity infrastructure.