Location of Repository

Let us consider the differential equation $$ \dot{x}=(A+\varepsilon Q(t,\varepsilon))x, \;\;\;\; |\varepsilon|\le\varepsilon_0, $$ where A is an elliptic constant matrix and Q depends on time in a quasi-periodic (and analytic) way. It is also assumed that the eigenvalues of A and the basic frequencies of Q satisfy a diophantine condition. Then it is proved that this system can be reduced to $$ \dot{y}=(A^{*}(\varepsilon)+\varepsilon R^{*}(t,\varepsilon))y, \;\;\;\; |\varepsilon|\le\varepsilon_0, $$ where $R^{*}$ is exponentially small in $\varepsilon$, and the linear change of variables that performs such a reduction is also quasi-periodic with the same basic frequencies as Q. The results are illustrated and discussed in a practical example

Topics:
Anàlisi global (Matemàtica), Global analysis (Mathematics)

Publisher: Society for Industrial and Applied Mathematics

Year: 2016

DOI identifier: 10.1137/S0036141095280967

OAI identifier:
oai:diposit.ub.edu:2445/69315

Provided by:
Diposit Digital de la Universitat de Barcelona

Downloaded from
http://dx.doi.org/10.1137/S0036141095280967

To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.