ZeroSim: Zero-Shot Analog Circuit Evaluation with Unified Transformer Embeddings

arXiv — cs.LGWednesday, November 12, 2025 at 5:00:00 AM
The introduction of ZeroSim marks a significant advancement in the field of analog circuit design automation, particularly in performance evaluation, which has been a major bottleneck due to the time-consuming nature of traditional SPICE simulations. ZeroSim utilizes a transformer-based framework that allows for robust in-distribution generalization across trained topologies and zero-shot generalization to unseen topologies without the need for fine-tuning. This is achieved through a comprehensive training corpus of 3.6 million instances covering over 60 amplifier topologies, as well as innovative strategies such as unified topology embeddings and topology-conditioned parameter mapping. The framework's ability to deliver accurate predictions across different amplifier topologies positions it as a superior alternative to existing machine learning methods, which often require extensive retraining or manual adjustments. By significantly outperforming baseline models, ZeroSim not only enha…
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Meta-SimGNN: Adaptive and Robust WiFi Localization Across Dynamic Configurations and Diverse Scenarios
PositiveArtificial Intelligence
Meta-SimGNN is a novel WiFi localization system that combines graph neural networks with meta-learning to enhance localization generalization and robustness. It addresses the limitations of existing deep learning-based localization methods, which primarily focus on environmental variations while neglecting the impact of device configuration changes. By introducing a fine-grained channel state information (CSI) graph construction scheme, Meta-SimGNN adapts to variations in the number of access points (APs) and improves usability in diverse scenarios.
DeepBlip: Estimating Conditional Average Treatment Effects Over Time
PositiveArtificial Intelligence
DeepBlip is a novel neural framework designed to estimate conditional average treatment effects over time using structural nested mean models (SNMMs). This approach allows for the decomposition of treatment sequences into localized, time-specific 'blip effects', enhancing interpretability and enabling efficient evaluation of treatment policies. DeepBlip integrates sequential neural networks like LSTMs and transformers, addressing the limitations of existing methods by allowing simultaneous learning of all blip functions.
Bayes optimal learning of attention-indexed models
PositiveArtificial Intelligence
The paper introduces the attention-indexed model (AIM), a framework for analyzing learning in deep attention layers. AIM captures the emergence of token-level outputs from bilinear interactions over high-dimensional embeddings. It allows full-width key and query matrices, aligning with practical transformers. The study derives predictions for Bayes-optimal generalization error and identifies phase transitions based on sample complexity, model width, and sequence length, proposing a message passing algorithm and demonstrating optimal performance via gradient descent.
CLAReSNet: When Convolution Meets Latent Attention for Hyperspectral Image Classification
PositiveArtificial Intelligence
CLAReSNet, a new hybrid architecture for hyperspectral image classification, integrates multi-scale convolutional extraction with transformer-style attention through an adaptive latent bottleneck. This model addresses challenges such as high spectral dimensionality, complex spectral-spatial correlations, and limited training samples with severe class imbalance. By combining convolutional networks and transformers, CLAReSNet aims to enhance classification accuracy and efficiency in hyperspectral imaging applications.
Flow-Attentional Graph Neural Networks
PositiveArtificial Intelligence
Graph Neural Networks (GNNs) are crucial for analyzing graph-structured data, but current models overlook the conservation laws relevant to physical resource flows, such as electrical currents in power grids. To improve performance, a new approach called flow attention is introduced, which aligns with Kirchhoff's first law. Experiments on electronic circuits and power grids demonstrate that this method enhances the effectiveness of attention-based GNNs in classification and regression tasks.
Hypergraph Neural Network with State Space Models for Node Classification
PositiveArtificial Intelligence
Recent advancements in graph neural networks (GNNs) have highlighted their effectiveness in node classification tasks. However, traditional GNNs often neglect role-based characteristics that can enhance node representation learning. To overcome these limitations, a new model called the hypergraph neural network with state space model (HGMN) has been proposed, integrating role-aware representations and employing hypergraph construction techniques to capture complex relationships among nodes.
Multi-View Polymer Representations for the Open Polymer Prediction
PositiveArtificial Intelligence
The article discusses a novel approach to polymer property prediction using a multi-view design that incorporates various representations. The system combines four families of representations: tabular RDKit/Morgan descriptors, graph neural networks, 3D-informed representations, and pretrained SMILES language models. This ensemble method achieved a public mean absolute error (MAE) of 0.057 and a private MAE of 0.082, ranking 9th out of 2241 teams in the Open Polymer Prediction Challenge at NeurIPS 2025.
Multistability of Self-Attention Dynamics in Transformers
NeutralArtificial Intelligence
The paper titled 'Multistability of Self-Attention Dynamics in Transformers' explores a continuous-time multiagent model of self-attention mechanisms in transformers. It establishes a connection between self-attention dynamics and a multiagent version of the Oja flow, which computes the principal eigenvector of a matrix related to the value matrix in transformers. The study classifies the equilibria of the single-head self-attention system into four categories: consensus, bipartite consensus, clustering, and polygonal equilibria, noting that multiple stable equilibria can coexist.