80 research outputs found
DeepSeq: Deep Sequential Circuit Learning
Circuit representation learning is a promising research direction in the
electronic design automation (EDA) field. With sufficient data for
pre-training, the learned general yet effective representation can help to
solve multiple downstream EDA tasks by fine-tuning it on a small set of
task-related data. However, existing solutions only target combinational
circuits, significantly limiting their applications. In this work, we propose
DeepSeq, a novel representation learning framework for sequential netlists.
Specifically, we introduce a dedicated graph neural network (GNN) with a
customized propagation scheme to exploit the temporal correlations between
gates in sequential circuits. To ensure effective learning, we propose to use a
multi-task training objective with two sets of strongly related supervision:
logic probability and transition probability at each node. A novel dual
attention aggregation mechanism is introduced to facilitate learning both tasks
efficiently. Experimental results on various benchmark circuits show that
DeepSeq outperforms other GNN models for sequential circuit learning. We
evaluate the generalization capability of DeepSeq on a downstream power
estimation task. After fine-tuning, DeepSeq can accurately estimate power
across various circuits under different workloads
Addressing Variable Dependency in GNN-based SAT Solving
Boolean satisfiability problem (SAT) is fundamental to many applications.
Existing works have used graph neural networks (GNNs) for (approximate) SAT
solving. Typical GNN-based end-to-end SAT solvers predict SAT solutions
concurrently. We show that for a group of symmetric SAT problems, the
concurrent prediction is guaranteed to produce a wrong answer because it
neglects the dependency among Boolean variables in SAT problems. % We propose
AsymSAT, a GNN-based architecture which integrates recurrent neural networks to
generate dependent predictions for variable assignments. The experiment results
show that dependent variable prediction extends the solving capability of the
GNN-based method as it improves the number of solved SAT instances on large
test sets
EDA-Driven Preprocessing for SAT Solving
Effective formulation of problems into Conjunctive Normal Form (CNF) is
critical in modern Boolean Satisfiability (SAT) solving for optimizing solver
performance. Addressing the limitations of existing methods, our Electronic
Design Automation (EDA)-driven preprocessing framework introduces a novel
methodology for preparing SAT instances, leveraging both circuit and CNF
formats for enhanced flexibility and efficiency. Central to our approach is the
integration of a new logic synthesis technique, guided by a reinforcement
learning agent, and a novel cost-customized LUT mapping strategy, enabling
efficient handling of diverse SAT challenges. By transforming the SAT
competition benchmarks into circuit instances, our framework demonstrates
substantial performance improvements, as evidenced by a 52.42% reduction on
average compared to solving directly. Moreover, our framework achieves a
remarkable 96.14% runtime reduction on average for a set of logic equivalence
checking problems that exhibit inherent circuit structures. These results
highlight the effectiveness and versatility of our approach in handling both
CNF and circuit instances. The code is available at
https://github.com/cure-lab/EDA4SAT
- …