189 research outputs found

    La caractérisation du transistor par le tuner source & load pull pour l'amplificateur de puissance classe F inverse

    Get PDF
    RÉSUMÉ: Dans ce travail, la méthodologie de conception de premier passage pour des amplificateurs de puissance de classe F inverse est présentée. Pour concevoir l'amplificateur de puissance de classe F inverse à 3.5 GHz avec un signal 1-ton et un signal LTE, le transistor CGH40010 de Cree Inc. est analysé et caractérisé avec un signal I-ton et un signal LTE. La caractérisation et l'analyse sont exploitées par la simulation source & load pull dans le logiciel ADS 2011.10 et par le système source & load pull tuner multi-harmonique passif de Focus microwaves Inc. Les 2ème et 3ème harmoniques à l'entrée et à la sortie du transistor sont pris en compte dans les caractérisations. L'analyse des résultats de la caractérisation par la simulation et par le système de tuner a prouvé que non seulement les 2ème et 3ème harmoniques à la sortie du transistor, mais également les 2ème et 3ème harmoniques à l'entrée du transistor sont importantes pour atteindre un rendement élevé et une puissance de sortie élevée. Le coefficient de réflexion (r) maximal du tuner passif à des fréquences harmoniques est important dans la caractérisation du transistor, car un r élevé peut augmenter le rendement en puissance ajoutée et la puissance de sortie de l'amplificateur de puissance de classe F inverse. Pour augmenter le coefficient de réflexion (r) maximal de tuner, les accessoires utilisés dans le système de tuner, tels que les circuits de polarisation, les coupleurs directifs et l'isolateur, sont analysés. Sur la base de cette analyse, le r maximal du tuner pourrait être augmenté de 0.902 à 0.930 par choisir les accessoires. Sur la base des résultats des caractérisations obtenus par la simulation et le système source & load pull tuner, les amplificateurs de classe F inverse sont conçus et fabriqués. En comparant les résultats mesurés des amplificateurs de puissance qui sont fabriqués sur la base des résultats de caractérisation correspondants obtenus par la simulation et le système de tuner, nous avons constaté que le système de tuner peut prédire les résultats, tel que le rendement en puissance ajoutée et la puissance de sortie, plus précisément que la simulation avec le modèle du transistor de grand signal. Pour l'amplificateur conçu sur la base des résultats de caractérisation obtenus par le système de tuner avec un signal I-ton, lorsque la puissance de sortie est 40.02 dBm, le rendement en puissance ajoutée est 79.76% avec 12.08 dB de gain. Pour l'amplificateur conçu sur la base des résultats de caractérisation obtenus par le système de tuner avec un signal LTE, lorsque la puissance de sortie est de 34.20 dBm et le gain est 16.20 dB, le rendement en puissance ajoutée est 49.56%. Le taux de puissance du canal adjacent 1, qui a un décalage de 10 MHz, est -29.45 dBc, et le taux de puissance du canal adjacent 2, qui a un décalage de 20 MHz, est -50.87 dBc. -- Mots clés: Le circuit de polarisation, la caractérisation de transistor, l'amplificateur de puissance de classe F inverse, tuner source & load pull. -- ABSTRACT: In this work, a first-pass design methodology of designing the inverse c1ass F power amplifier is presented. In order to design the inverse class F power amplifiers for the I-tone signal and the LTE signal, Cree's CGH40010 transistor is analyzed and characterized with a I-tone signal at 3.5 GHz and a LTE signal at 3.5 GHz with 10 MHz bandwidth. The characterization and analysis are operated by the source & load pull simulation in the software ADS 2011.10 with large signal transistor model and by the passive multiharmonic source & load pull tuner system from Focus microwaves Inc. In the characterizations of transistor, 2nd and 3rd harmonic on both input and output side of the transistor are considered. The analysis of the transistor's characterization shows that, not only the 2nd and 3rd harmonics on the output side of the transistor are important to achieve high power added efficiency and output power for an inverse class F power amplifier, but also the 2nd and 3rd harmonics on the input side of the transistor. High reflection coefficient (r) achieved by the passive source and load pull tuner at harmonic frequencies is important in the transistor characterization, since the high r can increase the power added efficiency and the output power of the inverse class F power amplifier. To increase the maximum r of the passive tuner system, the accessories in the passive tuner system, such as bias tee, directional coupler and isolator, are analyzed. Based on the analysis, the maximum r of the passive tuner system could be increased from 0.902 to 0.930 by choosing the accessories. Based on the characterization results obtained by the simulation and the tuner system, bias circuit and impedance matching networks are analyzed and designed for the inverse class F power amplifiers. Bias circuit is designed to maximize the RF isolation to the De input port and the return loss, and minimize the insertion loss at fundamental frequency. The inverse c1ass F power amplifiers designed based on the characterization results are fabricated and measured. By comparing the measured results of the power amplifiers which fabricated based on the corresponding characterization results obtained by simulation and tuner system, we found that the multi-harmonic passive source & load pull tuner system can predict the power added efficiency and the output power more precisely than the simulation with the large signal transistor model. For the fabricated I-tone inverse class F power amplifier designed based on the I-tone characterization result obtained by the tuner system, when the output power is 40.02 dBm, the power added efficiency is 79.76% with a gain of 12.08 dB. For the fabricated LTE inverse class F power amplifier designed based on the LTE characterization result obtained by the tuner system, the measured output power is 34.20 dBm with a power added efficiency of 49.56% and a gain of 16.20 dB. The measured worst adjacent channel power ratio 1 with 10 MHz offset is -29.45 dBc. The measured worst adjacent channel power ratio 2 with 20 MHz offset is -50.87 dBc. -- Keywords : Bias circuit, characterization, inverse c1ass F power amplifier, source & load pull tuner

    EfficientZero V2: Mastering Discrete and Continuous Control with Limited Data

    Full text link
    Sample efficiency remains a crucial challenge in applying Reinforcement Learning (RL) to real-world tasks. While recent algorithms have made significant strides in improving sample efficiency, none have achieved consistently superior performance across diverse domains. In this paper, we introduce EfficientZero V2, a general framework designed for sample-efficient RL algorithms. We have expanded the performance of EfficientZero to multiple domains, encompassing both continuous and discrete actions, as well as visual and low-dimensional inputs. With a series of improvements we propose, EfficientZero V2 outperforms the current state-of-the-art (SOTA) by a significant margin in diverse tasks under the limited data setting. EfficientZero V2 exhibits a notable advancement over the prevailing general algorithm, DreamerV3, achieving superior outcomes in 50 of 66 evaluated tasks across diverse benchmarks, such as Atari 100k, Proprio Control, and Vision Control.Comment: 21 pages,10 figure

    SpaceOctopus: An Octopus-inspired Motion Planning Framework for Multi-arm Space Robot

    Full text link
    Space robots have played a critical role in autonomous maintenance and space junk removal. Multi-arm space robots can efficiently complete the target capture and base reorientation tasks due to their flexibility and the collaborative capabilities between the arms. However, the complex coupling properties arising from both the multiple arms and the free-floating base present challenges to the motion planning problems of multi-arm space robots. We observe that the octopus elegantly achieves similar goals when grabbing prey and escaping from danger. Inspired by the distributed control of octopuses' limbs, we develop a multi-level decentralized motion planning framework to manage the movement of different arms of space robots. This motion planning framework integrates naturally with the multi-agent reinforcement learning (MARL) paradigm. The results indicate that our method outperforms the previous method (centralized training). Leveraging the flexibility of the decentralized framework, we reassemble policies trained for different tasks, enabling the space robot to complete trajectory planning tasks while adjusting the base attitude without further learning. Furthermore, our experiments confirm the superior robustness of our method in the face of external disturbances, changing base masses, and even the failure of one arm.Comment: 8 pages, 9 figure

    CoPa: General Robotic Manipulation through Spatial Constraints of Parts with Foundation Models

    Full text link
    Foundation models pre-trained on web-scale data are shown to encapsulate extensive world knowledge beneficial for robotic manipulation in the form of task planning. However, the actual physical implementation of these plans often relies on task-specific learning methods, which require significant data collection and struggle with generalizability. In this work, we introduce Robotic Manipulation through Spatial Constraints of Parts (CoPa), a novel framework that leverages the common sense knowledge embedded within foundation models to generate a sequence of 6-DoF end-effector poses for open-world robotic manipulation. Specifically, we decompose the manipulation process into two phases: task-oriented grasping and task-aware motion planning. In the task-oriented grasping phase, we employ foundation vision-language models (VLMs) to select the object's grasping part through a novel coarse-to-fine grounding mechanism. During the task-aware motion planning phase, VLMs are utilized again to identify the spatial geometry constraints of task-relevant object parts, which are then used to derive post-grasp poses. We also demonstrate how CoPa can be seamlessly integrated with existing robotic planning algorithms to accomplish complex, long-horizon tasks. Our comprehensive real-world experiments show that CoPa possesses a fine-grained physical understanding of scenes, capable of handling open-set instructions and objects with minimal prompt engineering and without additional training. Project page: https://copa-2024.github.io

    Efficient Exploration Using Extra Safety Budget in Constrained Policy Optimization

    Full text link
    Reinforcement learning (RL) has achieved promising results on most robotic control tasks. Safety of learning-based controllers is an essential notion of ensuring the effectiveness of the controllers. Current methods adopt whole consistency constraints during the training, thus resulting in inefficient exploration in the early stage. In this paper, we propose an algorithm named Constrained Policy Optimization with Extra Safety Budget (ESB-CPO) to strike a balance between the exploration efficiency and the constraints satisfaction. In the early stage, our method loosens the practical constraints of unsafe transitions (adding extra safety budget) with the aid of a new metric we propose. With the training process, the constraints in our optimization problem become tighter. Meanwhile, theoretical analysis and practical experiments demonstrate that our method gradually meets the cost limit's demand in the final training stage. When evaluated on Safety-Gym and Bullet-Safety-Gym benchmarks, our method has shown its advantages over baseline algorithms in terms of safety and optimality. Remarkably, our method gains remarkable performance improvement under the same cost limit compared with baselines.Comment: 7 pages, 8 figure

    DexCatch: Learning to Catch Arbitrary Objects with Dexterous Hands

    Full text link
    Achieving human-like dexterous manipulation remains a crucial area of research in robotics. Current research focuses on improving the success rate of pick-and-place tasks. Compared with pick-and-place, throw-catching behavior has the potential to increase picking speed without transporting objects to their destination. However, dynamic dexterous manipulation poses a major challenge for stable control due to a large number of dynamic contacts. In this paper, we propose a Stability-Constrained Reinforcement Learning (SCRL) algorithm to learn to catch diverse objects with dexterous hands. The SCRL algorithm outperforms baselines by a large margin, and the learned policies show strong zero-shot transfer performance on unseen objects. Remarkably, even though the object in a hand facing sideward is extremely unstable due to the lack of support from the palm, our method can still achieve a high level of success in the most challenging task. Video demonstrations of learned behaviors and the code can be found on the supplementary website

    Foundation Reinforcement Learning: towards Embodied Generalist Agents with Foundation Prior Assistance

    Full text link
    Recently, people have shown that large-scale pre-training from internet-scale data is the key to building generalist models, as witnessed in NLP. To build embodied generalist agents, we and many other researchers hypothesize that such foundation prior is also an indispensable component. However, it is unclear what is the proper concrete form to represent those embodied foundation priors and how they should be used in the downstream task. In this paper, we propose an intuitive and effective set of embodied priors that consist of foundation policy, value, and success reward. The proposed priors are based on the goal-conditioned MDP. To verify their effectiveness, we instantiate an actor-critic method assisted by the priors, called Foundation Actor-Critic (FAC). We name our framework as Foundation Reinforcement Learning (FRL), since it completely relies on embodied foundation priors to explore, learn and reinforce. The benefits of FRL are threefold. (1) Sample efficient. With foundation priors, FAC learns significantly faster than traditional RL. Our evaluation on the Meta-World has proved that FAC can achieve 100% success rates for 7/8 tasks under less than 200k frames, which outperforms the baseline method with careful manual-designed rewards under 1M frames. (2) Robust to noisy priors. Our method tolerates the unavoidable noise in embodied foundation models. We show that FAC works well even under heavy noise or quantization errors. (3) Minimal human intervention: FAC completely learns from the foundation priors, without the need of human-specified dense reward, or providing teleoperated demos. Thus, FAC can be easily scaled up. We believe our FRL framework could enable the future robot to autonomously explore and learn without human intervention in the physical world. In summary, our proposed FRL is a novel and powerful learning paradigm, towards achieving embodied generalist agents
    • …
    corecore