88 research outputs found

    IC optimisation using parallel processing and response surface methodology

    Get PDF

    Technology Independent Synthesis of CMOS Operational Amplifiers

    Get PDF
    Analog circuit design does not enjoy as much automation as its digital counterpart. Analog sizing is inherently knowledge intensive and requires accurate modeling of the different parametric effects of the devices. Besides, the set of constraints in a typical analog design problem is large, involving complex tradeoffs. For these reasons, the task of modeling an analog design problem in a form viable for automation is much more tedious than the digital design. Consequently, analog blocks are still handcrafted intuitively and often become a bottleneck in the integrated circuit design, thereby increasing the time to market. In this work, we address the problem of automatically solving an analog circuit design problem. Specifically, we propose methods to automate the transistor-level sizing of OpAmps. Given the specifications and the netlist of the OpAmp, our methodology produces a design that has the accuracy of the BSIM models used for simulation and the advantage of a quick design time. The approach is based on generating an initial first-order design and then refining it. In principle, the refining approach is a simulated-annealing scheme that uses (i) localized simulations and (ii) convex optimization scheme (COS). The optimal set of input variables for localized simulations has been selected by using techniques from Design of Experiments (DOE). To formulate the design problem as a COS problem, we have used monomial circuit models that are fitted from simulation data. These models accurately predict the performance of the circuit in the proximity of the initial guess. The models can also be used to gain valuable insight into the behavior of the circuit and understand the interrelations between the different performance constraints. A software framework that implements this methodology has been coded in SKILL language of Cadence. The methodology can be applied to design different OpAmp topologies across different technologies. In other words, the framework is both technology independent and topology independent. In addition, we develop a scheme to empirically model the small signal parameters like \u27gm\u27 and \u27gds\u27 of CMOS transistors. The monomial device models are reusable for a given technology and can be used to formulate the OpAmp design problem as a COS problem. The efficacy of the framework has been demonstrated by automatically designing different OpAmp topologies across different technologies. We designed a two-stage OpAmp and a telescopic OpAmp in TSMC025 and AMI016 technologies. Our results show significant (10–15%) improvement in the performance of both the OpAmps in both the technologies. While the methodology has shown encouraging results in the sub-micrometer regime, the effectiveness of the tool has to be investigated in the deep-sub-micron technologies

    AI/ML Algorithms and Applications in VLSI Design and Technology

    Full text link
    An evident challenge ahead for the integrated circuit (IC) industry in the nanometer regime is the investigation and development of methods that can reduce the design complexity ensuing from growing process variations and curtail the turnaround time of chip manufacturing. Conventional methodologies employed for such tasks are largely manual; thus, time-consuming and resource-intensive. In contrast, the unique learning strategies of artificial intelligence (AI) provide numerous exciting automated approaches for handling complex and data-intensive tasks in very-large-scale integration (VLSI) design and testing. Employing AI and machine learning (ML) algorithms in VLSI design and manufacturing reduces the time and effort for understanding and processing the data within and across different abstraction levels via automated learning algorithms. It, in turn, improves the IC yield and reduces the manufacturing turnaround time. This paper thoroughly reviews the AI/ML automated approaches introduced in the past towards VLSI design and manufacturing. Moreover, we discuss the scope of AI/ML applications in the future at various abstraction levels to revolutionize the field of VLSI design, aiming for high-speed, highly intelligent, and efficient implementations

    MakerFluidics: low cost microfluidics for synthetic biology

    Full text link
    Recent advancements in multilayer, multicellular, genetic logic circuits often rely on manual intervention throughout the computation cycle and orthogonal signals for each chemical “wire”. These constraints can prevent genetic circuits from scaling. Microfluidic devices can be used to mitigate these constraints. However, continuous-flow microfluidics are largely designed through artisanal processes involving hand-drawing features and accomplishing design rule checks visually: processes that are also inextensible. Additionally, continuous-flow microfluidic routing is only a consideration during chip design and, once built, the routing structure becomes “frozen in silicon,” or for many microfluidic chips “frozen in polydimethylsiloxane (PDMS)”; any changes to fluid routing often require an entirely new device and control infrastructure. The cost of fabricating and controlling a new device is high in terms of time and money; attempts to reduce one cost measure are, generally, paid through increases in the other. This work has three main thrusts: to create a microfluidic fabrication framework, called MakerFluidics, that lowers the barrier to entry for designing and fabricating microfluidics in a manner amenable to automation; to prove this methodology can design, fabricate, and control complex and novel microfluidic devices; and to demonstrate the methodology can be used to solve biologically-relevant problems. Utilizing accessible technologies, rapid prototyping, and scalable design practices, the MakerFluidics framework has demonstrated its ability to design, fabricate and control novel, complex and scalable microfludic devices. This was proven through the development of a reconfigurable, continuous-flow routing fabric driven by a modular, scalable primitive called a transposer. In addition to creating complex microfluidic networks, MakerFluidics was deployed in support of cutting-edge, application-focused research at the Charles Stark Draper Laboratory. Informed by a design of experiments approach using the parametric rapid prototyping capabilities made possible by MakerFluidics, a plastic blood--bacteria separation device was optimized, demonstrating that the new device geometry can separate bacteria from blood while operating at 275% greater flow rate as well as reduce the power requirement by 82% for equivalent separation performance when compared to the state of the art. Ultimately, MakerFluidics demonstrated the ability to design, fabricate, and control complex and practical microfluidic devices while lowering the barrier to entry to continuous-flow microfluidics, thus democratizing cutting edge technology beyond a handful of well-resourced and specialized labs

    Decision-maker Trade-offs In Multiple Response Surface Optimization

    Get PDF
    The focus of this dissertation is on improving decision-maker trade-offs and the development of a new constrained methodology for multiple response surface optimization. There are three key components of the research: development of the necessary conditions and assumptions associated with constrained multiple response surface optimization methodologies; development of a new constrained multiple response surface methodology; and demonstration of the new method. The necessary conditions for and assumptions associated with constrained multiple response surface optimization methods were identified and found to be less restrictive than requirements previously described in the literature. The conditions and assumptions required for a constrained method to find the most preferred non-dominated solution are to generate non-dominated solutions and to generate solutions consistent with decision-maker preferences among the response objectives. Additionally, if a Lagrangian constrained method is used, the preservation of convexity is required in order to be able to generate all non-dominated solutions. The conditions required for constrained methods are significantly fewer than those required for combined methods. Most of the existing constrained methodologies do not incorporate any provision for a decision-maker to explicitly determine the relative importance of the multiple objectives. Research into the larger area of multi-criteria decision-making identified the interactive surrogate worth trade-off algorithm as a potential methodology that would provide that capability in multiple response surface optimization problems. The ISWT algorithm uses an Δ-constraint formulation to guarantee a non-dominated solution, and then interacts with the decision-maker after each iteration to determine the preference of the decision-maker in trading-off the value of the primary response for an increase in value of a secondary response. The current research modified the ISWT algorithm to develop a new constrained multiple response surface methodology that explicitly accounts for decision-maker preferences. The new Modified ISWT (MISWT) method maintains the essence of the original method while taking advantage of the specific properties of multiple response surface problems to simplify the application of the method. The MISWT is an accessible computer-based implementation of the ISWT. Five test problems from the multiple response surface optimization literature were used to demonstrate the new methodology. It was shown that this methodology can handle a variety of types and numbers of responses and independent variables. Furthermore, it was demonstrated that the methodology can be successful using a priori information from the decision-maker about bounds or targets or can use the extreme values obtained from the region of operability. In all cases, the methodology explicitly considered decision-maker preferences and provided non-dominated solutions. The contribution of this method is the removal of implicit assumptions and includes the decision-maker in explicit trade-offs among multiple objectives or responses
    • 

    corecore