Efficient Stimuli Generation using Reinforcement Learning in Design Verification
CoRR(2024)
Abstract
The increasing design complexity of System-on-Chips (SoCs) has led to
significant verification challenges, particularly in meeting coverage targets
within a timely manner. At present, coverage closure is heavily dependent on
constrained random and coverage driven verification methodologies where the
randomized stimuli are bounded to verify certain scenarios and to reach
coverage goals. This process is said to be exhaustive and to consume a lot of
project time. In this paper, a novel methodology is proposed to generate
efficient stimuli with the help of Reinforcement Learning (RL) to reach the
maximum code coverage of the Design Under Verification (DUV). Additionally, an
automated framework is created using metamodeling to generate a SystemVerilog
testbench and an RL environment for any given design. The proposed approach is
applied to various designs and the produced results proves that the RL agent
provides effective stimuli to achieve code coverage faster in comparison with
baseline random simulations. Furthermore, various RL agents and reward schemes
are analyzed in our work.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined