Anonymous
×
Create a new article
Write your page title here:
We currently have 106 articles on MOR Wiki. Type your article name above or click on one of the titles below and start writing!



Reduced Basis PMOR method: Difference between revisions

Hessm (talk | contribs)
No edit summary
Hessm (talk | contribs)
No edit summary
Line 78: Line 78:
The greedy sampling uses an error estimator <math> \Delta_{N}(\mu) </math> which estimates (even rigorously, in some cases) the approximation error <math> /| u(\mu) - u_N(\mu) /| </math>.
The greedy sampling uses an error estimator <math> \Delta_{N}(\mu) </math> which estimates (even rigorously, in some cases) the approximation error <math> /| u(\mu) - u_N(\mu) /| </math>.


Let <math> \Xi </math> denote a finite sample of <math> \mathcal{D} </math> and set <math> S_1 = \{\mu^1\}$ and $V_1 = span\{ u(\mu^1) \} </math>.
Let <math> \Xi </math> denote a finite sample of <math> \mathcal{D} </math> and set <math> S_1 = \{\mu^1\} \text{ and } V_1 = span\{ u(\mu^1) \} </math>.
For <math> N = 2 , ... , N_{max} <math>, find <math> \mu^N = \text{arg max}_{ \mu \in \Xi } \Delta_{N-1}(\mu) </math>,
For <math> N = 2 , ... , N_{max} <math>, find <math> \mu^N = \text{arg max}_{ \mu \in \Xi } \Delta_{N-1}(\mu) </math>,
and then set <math> S_N = S_{N-1} \cup \mu^N , \quad V_N = V_{N-1} + span\{u(\mu^N)\} </math>.
and then set <math> S_N = S_{N-1} \cup \mu^N , \quad V_N = V_{N-1} + span\{u(\mu^N)\} </math>.

Revision as of 14:42, 19 November 2012


The Reduced Basis Method (RBM) we present here is applicable to static and time-dependent linear PDEs.

Time-Independent PDEs

The typical model problem of the RBM consists of a parametrized PDE stated in weak form with bilinear form a(,;μ) and linear form f(;μ). The parameter μ is considered within a domain 𝒟 and we are interested in an output quantity s(μ) which can be expressed via a linear functional of the field variable l(;μ).

The exact, infinite-dimensional formulation, indicated by the superscript e, is given by


{For μ𝒟P, evaluate se(μ)=l(ue(μ);μ),where ue(μ)Xe(Ω) satisfies a(ue(μ),v;μ)=f(v;μ),vXe.

We assume a large-scale discretization to be given, such that we consider

{For μ𝒟P, evaluate s(μ)=l(u(μ);μ),where u(μ)X(Ω) satisfies a(u(μ),v;μ)=f(v;μ),vX.

The underlying assumption of the RBM is that the parametrically induced manifold ={u(μ)|μ𝒟} can be approximated by a low dimensional space VN.

It also applies the concept of an offline-online decomposition, in that a large pre-processing offline cost is acceptable in view of a very low online cost (of a reduced order model) for each input-output evaluation, when in a many-query or real-time context.

The essential assumption which allows the offline-online decomposition is that there exists an affine parameter dependence

a(w,v;μ)=q=1QΘaq(μ)aq(w,v)

f(v;μ)=q=1QfΘfq(μ)fq(v).

The Lagrange Reduced Basis space is established by iteratively choosing Lagrange parameter samples

SN={μ1,...,μN}

and considering the associated Lagrange RB spaces

VN=span{u𝒩(μn),1nN}

in a greedy sampling. This leads to hierarchical RB spaces: V1V2...VNmax.

We then consider the galerkin projection onto the RB-space VN

{For μ𝒟P, evaluate sN(μ)=f(uN(μ)),where uN(μ)WN satisfies a(uN(μ),v;μ)=f(v),vWN

The greedy sampling uses an error estimator ΔN(μ) which estimates (even rigorously, in some cases) the approximation error /|u(μ)uN(μ)/|.

Let Ξ denote a finite sample of 𝒟 and set S1={μ1} and V1=span{u(μ1)}. For N=2,...,Nmax<math>,find<math>μN=arg maxμΞΔN1(μ), and then set SN=SN1μN,VN=VN1+span{u(μN)}.


Time-Dependent PDEs

References