Improved Variational Approximation for Bayesian PCA

合集下载

关于Lax-Milgram定理和Céa定理的一个注记

关于Lax-Milgram定理和Céa定理的一个注记

关于Lax-Milgram定理和Céa定理的一个注记杜乃杰;杜乃林【期刊名称】《数学杂志》【年(卷),期】2007(027)004【摘要】本文研究抽象变分问题(不必要求具有强制性)的Galerhin方法,利用泛函分析理论证明了:若变分问题的Galerkin逼近问题存在唯一解,那么它本身的解存在唯一且可由Galerhin逼近解无限逼近的充要条件是其Galerkin逼近格式具有某种稳定性.此结果是对Lax-Milgram定理和Céa定理的补充,可以应用于不必具有强制性的变分问题.%This paper investigates Galerkin methods for an abstract variational problem when its bilinear form has no eoereivity,and by the use of funetional analysis theory,proves such a result:the variational problem has a unique solution and the solution can be approximated by their Galerkin approximation solutions if and only if the Galerkin approximation problems sarisfy certain stability,where the Galerkin approximation problems are supposed to have unique solutions.This result may be viewed as replenishment for the Lax-Milgram theorem and the Céa theorem;it is applied to the variational problem with or without coercivity.【总页数】8页(P411-418)【作者】杜乃杰;杜乃林【作者单位】河北工程大学资源学院,河北邯郸,056038;武汉大学数学与统计学院,湖北武汉,430072【正文语种】中文【中图分类】O177.1【相关文献】1.关于完备随机内积模中的Lax-Milgram定理的注记 [J], 郭铁信;周国军x-Milgram定理的一个非线性变体在椭圆型p-Laplace问题上的应用 [J], 朱云峰;吴鹏鹏;张凯强;丛肖汉3.关于环的一个定理的注记—华罗庚定理的简易证明 [J], 何兴旺x-Milgram定理的一个推广及其应用 [J], 张馨予;吕颖5.关于闭图像定理的一个注记 [J], 沈永红;孙小科因版权原因,仅展示原文概要,查看原文内容请购买。

响应面法优化Fenton预处理精细化工废水

响应面法优化Fenton预处理精细化工废水

类、不饱和醛类和芳香类化合物. 研究显示ꎬFenton 预处理能有效降解废水中难降解有机物ꎬ但出水仍未达到后续生化处理要求ꎬ
还需进一步优化或与其他预处理工艺组合.
关键词: Fenton 氧化法ꎻ 精细化工废水ꎻ 响应面ꎻ 反应动力学ꎻ 光谱分析
中图分类号: X703 文章编号: 1001 ̄6929(2019)08 ̄1419 ̄08
Abstract: The wastewater from a fine chemistry enterprise in Gansu Province is characterized by complex ingredientsꎬ high organic matter contentꎬ and poor biodegradability. In order to meet the requirements of subsequent biochemical processesꎬ it is essential to carry out appropriate pretreatment technology research. The present study demonstrated a trial of optimizing Fenton oxidation process for pretreatment of the fine chemistry wastewater via the response surface methodology ( RSM) . Based on the results of the single factor experimentꎬ the initial pHꎬ H2 O2 dosageꎬ n( H2 O2 ) ∶n( Fe2+ ) and reaction time were selected as the investigation factors and the CODCr removal efficiency as the response value. The individual effect of the four independent factors and the synergistic effects of their interactions on CODCr removal were studied and analyzed. Moreoverꎬ the apparent reaction kineticsꎬ structural changes of wastewater organic matterꎬ and the degradation mechanisms were studied by ultraviolet spectroscopy and Fourier transform infrared spectroscopy. The results showed that: ( 1) The optimum process conditions were initial pH of 4ꎬ H2 O2 dosage of 8 mL∕Lꎬ n( H2 O2 ) ∶n( Fe2+ ) of 12ꎬ reaction time of 88 minꎬ and CODCr removal efficiency of 30������ 15%ꎬ which fitted well with the experimental results. (2) The degradation pathways of the organic matter in the fine chemistry wastewater by Fenton oxidation were complicatedꎬ and it′s difficult to fit through a single substrate model. ( 3) Fenton oxidation could effectively degrade unsaturated organic matter in the wastewater. Howeverꎬ the effluent still contained amidesꎬ unsaturated aldehydesꎬ and aromatic compounds. In conclusionꎬ Fenton oxidation pretreatment can effectively degrade the refractory organics in wastewater. Howeverꎬ the effluent still does not meet requirements of the subsequent biochemical treatment. Thusꎬ further optimization or combination with other pretreatment processes is required. Keywords: Fenton oxidation processꎻ fine chemical wastewaterꎻ response surface modelꎻ reaction kineticsꎻ spectral analysis

变分推断-解决复杂的后验概率计算方法

变分推断-解决复杂的后验概率计算方法

如何简单易懂地理解变分推断(variational inference)?简单易懂的理解变分其实就是一句话:用简单的分布q去近似复杂的分布p。

首先,为什么要选择用变分推断?因为,大多数情况下后验分布很难求啊。

如果后验概率好求解的话我们直接EM 就搞出来了。

当后验分布难于求解的时候我们就希望选择一些简单的分布来近似这些复杂的后验分布,至于这种简单的分布怎么选,有很多方法比如:Bethe自由能,平均场定理。

而应用最广泛的要数平均场定理。

为什么?因为它假设各个变量之间相互独立砍断了所有变量之间的依赖关系。

这又有什么好处呢?我们拿一个不太恰当的例子来形象的说明一下:用古代十字军东征来作为例子说明一下mean field。

十字军组成以骑兵为主步兵为辅,开战之前骑兵手持重标枪首先冲击敌阵步兵手持刀斧跟随,一旦接战就成了单对单的决斗。

那么在每个人的战斗力基本相似的情况下某个人的战斗力可以由其他人的均值代替这是平均场的思想。

这样在整个军队没有什么战术配合的情况下军队的战斗力可以由这些单兵的战斗力来近似这是变分的思想。

当求解Inference问题的时候相当于积分掉无关变量求边际分布,如果变量维度过高,积分就会变得非常困难,而且你积分的分布p又可能非常复杂因此就彻底将这条路堵死了。

采用平均场就是将这种复杂的多元积分变成简单的多个一元积分,而且我们选择的q是指数族内的分布,更易于积分求解。

如果变量间的依赖关系很强怎么办?那就是structured mean field解决的问题了。

说到这里我们就知道了为什么要用变分,那么怎么用?过程很简单,推导很复杂。

整个过程只需要:1、根据图模型写出联合分布2、写出mean filed 的形式(给出变分参数及其生成隐变量的分布)3、写出ELBO(为什么是ELBO?优化它跟优化KL divergence等价,KL divergence 因为含有后验分布不好优化)4、求偏导进行变分参数学习这样就搞定了!摘自 Wikipedia: Variational Bayesian methodsVariational Bayesian methods are primarily used for two purposes:1.To provide an analytical approximation to the posteriorprobability of the unobserved variables, in order to dostatistical inference over these variables.2.To derive a lower bound for the marginal likelihood(sometimes called the "evidence") of the observed data (i.e.the marginal probability of the data given the model, withmarginalization performed over unobserved variables). Thisis typically used for performing model selection, the generalidea being that a higher marginal likelihood for a given modelindicates a better fit of the data by that model and hencea greater probability that the model in question was the onethat generated the data. (See also the Bayes factor article.) 前面两位答主说的主要是第1点,不过在深度学习中第2点更常见。

降低衰老γh2ax表达

降低衰老γh2ax表达

降低衰老γh2ax表达英文回答:To reduce γH2AX expression and slow down aging, there are several approaches that can be taken. One way is to focus on lifestyle factors that can impact DNA damage and repair. For example, maintaining a healthy diet rich in antioxidants can help reduce oxidative stress and DNA damage. Regular exercise has also been shown to improve DNA repair mechanisms and reduce DNA damage. Additionally, getting enough sleep and managing stress levels can also play a role in reducing DNA damage and promoting healthy aging.Another approach is to explore potential interventions that can directly target γH2AX expression. One potential avenue is the use of certain compounds or drugs that can modulate DNA damage response pathways. For instance,poly(ADP-ribose) polymerase (PARP) inhibitors have been shown to reduce γH2AX levels and improve DNA repairefficiency. These inhibitors are currently being investigated for their potential anti-aging effects.Furthermore, it's important to note that maintaining a healthy l ifestyle and reducing γH2AX expression is not a one-size-fits-all solution. Different individuals may have different genetic predispositions and lifestyle factors that contribute to DNA damage and aging. Therefore, personalized approaches that consider an individual's unique genetic makeup and lifestyle choices may be more effective in reducing γH2AX expression and promoting healthy aging.中文回答:要降低γH2AX的表达并延缓衰老,有几种方法可以采取。

Simulation of ArbitraGrowth in Composite Structures Using the Virtual Crack Extension Method

Simulation of ArbitraGrowth in Composite Structures Using the Virtual Crack Extension Method

COVER SHEETTitle: Simulation of Arbitrary Delamination Growth in Composite Structures Using the Virtual Crack Extension Method Authors: Brett R. DavisPaul A. WawrzynekAnthony R. IngraffeaABSTRACTA finite-element-based toolset has been developed to simulate arbitrary evolution of 3-D, geometrically explicit, interlaminar delaminations in composite structures. A new energy-based growth formulation uses the virtual crack extension (VCE) method to predict point-by-point growth along the delamination front. The VCE method offers an accurate and computationally efficient means to extract both energy release rate, , and rate of change of energy release rate, , from a single finite element analysis. A new VCE implementation is used to decouple and compute 3-D, mixed-mode, fracture parameters, permitting the use of mixed-mode growth criteria. The parameters create an influence matrix that relates an extension at one point to the energy release rate elsewhere along the delamination front. The use of the matrix, in conjunction with an iterative approach that continually updates the delamination configuration by re-meshing, enables the prediction of arbitrary delamination evolution. The numerical techniques and formulations implemented allow a delamination to grow by the rules of mechanics and physics, while reducing computational artifacts, e.g. mesh bias. The evolution of an initial embedded elliptical delamination under central point-loads into a circular configuration is simulated as a proof-of-concept for the new growth formulation.INTRODUCTIONLaminated, fiber-reinforced, composite materials are used in a variety of applications, including the aerospace and marine industries. Although laminated composite structures have been in use for decades, fully understanding and accurately predicting their failure mechanisms remains a significant challenge. The current work seeks to address this challenge for one of the most common modes of failure: delamination. A variety of finite-element-based approaches has been Cornell University, School of Civil and Environmental Engineering638 Frank H.T. Rhodes Hall, Ithaca, NY 14853, U.S.A.developed to handle the problem of progressive delamination growth. These include damage mechanics [1], cohesive crack models [2], nodal release methods via the virtual crack closure technique (VCCT) [3,4] and the extended finite element method (XFEM) [5]. This paper describes the development of a new approach, one that reduces computational limitations, freeing the mechanics and physics to dictate the evolution of the delamination.Perhaps the most widely utilized of the aforementioned are cohesive zone models and nodal release methods. However, these methods limit the characterization of delamination growth by constraining the evolution to the geometry of the predetermined finite element mesh. Cohesive zone elements use traction-displacement curves to govern behavior at the delamination front. When the traction-displacement relationship reaches its limit, i.e. maximum displacement, and zero traction, the cohesive element is eliminated, thus extending the delamination. Alternatively, fracture mechanics methods are used to calculate stress intensity factors or energy release rates at the front, and identify nodes to be released to extend the front. In either case, the delamination path will only follow existing element boundaries, restricting the direction and distance of the predicted advance of the front. These growth prediction approaches have the potential to produce saw-tooth, jagged delamination fronts that do not match physical configurations, as seen in Figure 1.To minimize these limitations, one can use a heavily refined mesh around the front, greatly inflating the computational cost, or one can produce meshes constructed with knowledge of the expected delamination growth pattern. However, this latter technique contradicts the notion of arbitrary evolution by linking a physical feature, delamination shape, with a computational artifact, the mesh. For example, with an initial circular flaw, one might design a mesh that contains an organization of elements in a concentric circular pattern around the initial delamination, Figure 2. However, of interest here is what happens when theFigure 1. Jagged saw-tooth delamination growth via the nodal release method - resulting inmesh biased constrained growth. From [3].initial delamination geometry is complex, and the growth pattern is unknown. Meshes should not provide a predetermined undue bias that obfuscates the physical results of a simulation.This work proposes a new, general approach for the simulation of geometrically explicit delamination evolution. The geometry of the delamination front is continually updated and the finite element model is appropriately re-meshed around the updated front. This procedure of updating the delamination configuration and then re-meshing ensures the arbitrary nature of the evolution. The advance of the delamination and creation of new surface area are governed by a new energy-based formulation.An incremental-iterative procedure utilizes the growth formulation to calculate point-wise advances along the delamination front. The geometry of the delamination drives the finite element simulation; the growth is not dictated by the finite element mesh, but by the mechanics and physics embedded within the energy-based growth formulation.To use effectively the new energy-based growth formulation, fracture mechanics parameters must be accurately and efficiently calculated. A novel contribution of the new energy-based growth formulation is the use of the first order derivative of the energy release rate with respect to the delamination length, the rate of energy release rate, . The parameter, in the 3-D sense, serves as an influence matrix that relates how an extension, , at one point affects the energy release rate, , elsewhere along the front. The virtual crack extension (VCE)Figure 2. Pre-designed finite element mesh that adheres to expectedgrowth pattern – resulting in mesh biased constrained growth.method has been determined to be the most appropriate fracture mechanics tool, since it can extract both energy release rates and rates of energy release rate with a single finite element analysis. Other means of calculating the parameter require a costly and time consuming finite difference approach. Another novelty presented is a VCE implementation that permits the decomposition of the energy fracture parameters, thus allowing the future use of mixed-mode fracture criteria.The simulation technique incorporates three main components: 1) explicitly representing delaminations with re-meshing capabilities, 2) fracture calculations using the VCE method, and 3) prediction of growth using the new energy-based formulation. It should be noted that the techniques developed are independent. The energy-based growth formulation can be implemented using the VCCT (which has been done by the authors), or even be applied within the XFEM environment. The simulation methodology developed aims to achieve arbitrary delamination growth through representing the delamination as a geometric feature while reducing finite element bias.The following sections will introduce the VCE method, providing background and a general framework. The implementation of the new 3-D, mixed-mode decoupling of energy release rates will be discussed. Verification results will be presented showing its effectiveness. The energy-based growth formulation will be introduced, describing the foundation and implementation of the method. Finally, the simulation of an embedded elliptical delamination subject to equal and opposite delamination face point loads is offered as a proof-of-concept for the new growth formulation.THE VIRTUAL CRACK EXTENSION METHODThe virtual crack extension (VCE) method, also known as the stiffness derivative method, is an energy approach first introduced by Dixon and Pook [6] and Watwood [7], and further developed by Hellen [8] and Parks [9]. Early VCE calculations utilized explicit perturbations of the finite element meshes to approximate the method’s required stiffness derivatives. Th e finite difference approach of calculating derivatives often introduces geometric approximation and numerical truncation errors. Haber and Koh [10] developed a variational approach that eliminated the need for perturbations for stiffness derivative calculations. Lin and Abel [11] derived a similar method, rooted in variational principle theory, but generalized and simplified the required integration of [10]. Additionally, and importantly to the energy-based growth formulation herein, the approach of [11] permits the derivation of expressions for higher order derivatives of energy release rates. Hwang utilized the formulation of [11], generalizing the direct-integration approach for 2-D [12], multiply cracked bodies [13], and planar 3-D cracks [14]. The salient features of the VCE method are the accurate calculation of energy release rates and their derivatives, with a single finite element analysis.Virtual Crack Extension FormulationThis section outlines the explicit expressions derived using the variational approach for the 3-D energy release rates and their derivatives. The followingformulation demonstrates the mathematical development of the VCE method. For a more complete mathematical derivation and discussion see [11] and [14].The potential energy, , of a finite element system is given by(1)where , , and are the nodal displacement vector, the global stiffness matrix and the applied nodal force vector, respectively.The energy release rate, , at front position is defined as the negative derivative of the potential energy with respect to an incremental front extension, , at that position.(2)For simplicity, it is assumed that nodal forces are not influenced by the incremental virtual extensions. Therefore, the variational force term, , goes to zero. The necessary parameters for a local energy release rate calculation at position require nodal displacements and the variation in stiffness due to an incremental crack extension. The simplification reduces equation (2) to:(3) It should be noted, however, if crack-face pressures, thermal, and/or body force loadings are considered, the variational force term must be included throughout the expressions and derivations.The expression for the first order derivative of the energy release rate follows from the previous equations, by taking the variation of in equation (3) with respect to another incremental crack extension, , at position .(4) The variation of displacements extends directly from the variation of the global finite element equilibrium equation, , with respect to an incremental crack extension, .(5)Rearranging equation (5) and applying the simplifying assumption of , yields the expression for the variation of nodal displacements:( )(6)The remaining derivations of the expressions for the stiffness derivatives can be found in [14]. These require a ‘strain-like’ matrix created by the virtual extensions applied to th e front elements. The ‘strains’ are created by geometry changes of the finite elements in parametric space. Through the Jacobian and basis functions, variations in the strain-displacement matrices can be formulated, providing the components necessary for the stiffness derivative integrals.Three-Dimensional Mixed-Mode Decomposition of Energy Release Rates Several methods have been proposed to decompose energy release rates using the VCE method. [15] attempted to draw concepts used from the decomposition of the -integral into and . However, inaccuracies were discovered in the term when mode-II was dominant. [10] proposed a general 2-D method using Betti’s reciprocal theorem and Yau’s mutual energy representation. [11] used a similar approach that marries actual displacement fields with analytic solutions for pure mode-I and pure mode-II behavior. This is comparable to methods used in the interaction -integral [16,17]. Ishikawa [18] developed a 2-D approach that first decomposed the displacements into mode-I and mode-II components using symmetry and skew-symmetry about the delamination front. The decomposed displacements were then used in standard VCE equations to calculate the mixed-mode energy release rates. This method was extended to 3-D by Nikishkov and Atluri [19] for the -integral, and identified by [14] as a feasible 3-D approach for mode decomposition within the VCE method. One limitation to the symmetric/skew-symmetric technique is that a symmetric mesh is required around the front. This limitation is overcome through use of a front template that will be discussed later in the implementation section.In the decomposition technique, displacement fields around a straight front can be separated into mode-I, II and III components. For an arbitrary front, the use of a local front coordinate system is required, Figure 3. A point-by-point coordinate transformation scheme is employed to essentially ‘straighten’ the front locally so the symmetric decomposition method can be used.Once the front is in a local orientation, the mode-I, mode-II, and mode-III deformations can be decomposed with respect to the delamination plane along the local axis. Taking advantage of the symmetry of the finite element meshFigure 3. Local delamination front coordinate system.surrounding the front, displacements at point ()and at (), are used to decompose mode-I, mode-II and mode-III displacements at .(7){}(8){}(9){}(10)With the displacements decomposed, and a transformed local stiffness derivative, denoted by the subscript , available, the following equation yields the mixed-mode energy release rates:, where (11) Virtual Crack Extension ImplementationThis section discusses the implementation of the VCE method, and the various numerical schemes employed to improve performance. The core of the VCE method has been implemented as a MATLAB post process. The MATLAB code reads in finite element information that identifies front nodes, elements, etc., and nodal displacements. The front is surrounded by a template comprising three rings of elements whose geometry can be carefully controlled. This satisfies the symmetric requirements of the mode decomposition method discussed previously and facilitates accurate fracture parameter calculations. The elements directly surrounding the front are either 20-noded brick or 15-noded wedge serendipity quarter-point elements,Figure 4.The VCE procedure is applied across the front elements. Non-zero contributions to the stiffness derivatives occur only over elements involved with the virtual extensions. These elements are isolated for stiffness derivative calculations that are summed appropriately for a given virtual extension. Certain considerations must be addressed when dealing with a 3-D delamination. Along a 3-D front, virtual extensions between adjacent positions have an interaction component that must be accounted for in second order stiffness derivative calculations (). Also, the 3-D virtual extensions have an area associated with them depending on the element size that must be used to normalize the energy release rates and their derivatives. Since the formulation utilizes the variational approach, the virtualextension applied can be of unit length; no optimized distance must be determined, thereby simplifying the implementation.To ease the effort of computations, the use of an intermediate global force derivative parameter is employed. For each element involved with a particular front node , first order stiffness derivatives are calculated then multiplied by the element nodal displacements.(12)The force derivative parameter is mapped into a global environment and summed over the elements for the current front position. The global force derivative for each front node is stored in a matrix. This matrix is of size: number of degrees of freedom by number of front nodes, and is utilized in the calculation of the variation in displacements. Significant computational cost lies in the execution of equation(6). Unlike the stiffness derivative, the variation in displacements is a global calculation. The virtual extensions have an influence on all points within the finite element model. Equation (6) requires -number of back solves with the symmetric global stiffness matrix and global force derivative. To accelerate the calculation and alleviate memory issues associated with the back solves, a standalone, parallelizable, sparse direct solver, MUMPS (MU ltifrontal M assively P arallel sparse direct S olver), is utilized.VCE calculations can be conducted at corner or midside nodes along the front. The selection of corner or midside nodes alters the profile of the virtual extension: the former is linear, the latter quadratic. It was determined through a previous numerical investigation that calculations at corner nodes outperformed midside nodes. This observation agrees with results published in the literature for other methods [20].Also incorporated in the implementation is the ability to carry out the numerical integration over one or two rings of elements, Figure 5. The single ring of integration comprises only quarter-point elements. The second ring utilizes the quarter-point elements and the next layer of surrounding brick elements. The theory behind adding the second ring is to improve results by effectivelyshifting Figure 4. 20-noded brick and 15-noded wedge quarter point elements used to surround thedelamination front.X2X1Figure 5. (a) 1-ring and (b) 2-ring elements involved in numerical integration for the VCE method.Top figures show the initial configuration. Bottom figures show the applied virtual extensions. the area of high virtual strain caused by the virtual extensions away from the singular elements, to elements where field gradients are not as severe. The singular elements around the delamination front experience negligible virtual strain, i.e. shape and volume change in the element, and the brunt is moved to the second ring, Figure 6. In the same numerical investigation alluded to earlier, the second ring of integration was shown to provide more accurate results for both energy release rates and rates of energy release rate.Herein, results will be reported only for the simplest, albeit still effective, form of the mixed-mode VCE method that uses quarter-point brick elements and a single ring of numerical integration at corner nodes along the front.Verification Results for 3-D, Mixed-Mode Virtual Crack Extension Method This section presents preliminary verification for the newly implemented 3-D, mixed-mode VCE method. To check the formulation and implementation, analytical displacements [22] from prescribed energy release rates were imposed on a straight front. Four scenarios were analyzed: 1) pure mode-I, 2) pure mode-II, 3) pure mode-III, and 4) mixed-mode I/II/III. In each instance an energy release rate of unity was prescribed.For brevity, the results from the mixed-mode I/II/III scenario are shown in Figure 7. The pure mode-I, II, and III situations yielded similar levels of accuracyfor their respective energy release rates. For each analysis, the decomposed energy release rates were within a fraction of one percent of the prescribed value. The energy release rates also agreed well with the total energy release rate calculations, verifying the known linear relationship.ENERGY-BASED PREDICTION OF DELAMINATION GROWTHThe new energy-based growth formulation and implementation draw inspiration from experiences in plasticity. Both plasticity and delamination growth are characterized by behavioral transitions. These are denoted by a critical limit, for plasticity, the yield stress, and for delamination growth, a critical energy release rate. With this general connection between delamination growth and plasticity, a new formulation is developed for discretized front evolution.Energy-Based Growth FormulationThe formulation extends directly from an expansion of the energy release rate.(13)The current energy release rate, , is expanded into three components: the previous energy release rate prior to the load increment , , a portion due to the change in loading, , and a portion due to extension, . HereFigure 6. 2-ring virtual extension showing area of high virtual strain along the delamination front.Figure 7. Mixed-mode energy release rate results using the 3-D VCE decomposition method.characterizes the change in energy release rate with respect to the loading, is the aforementioned influence matrix, and is the extension increment. The expanded energy release rate forms a general stability equation that can be manipulated to calculate growth increments for an arbitrary front for a given load change.A local front failure criterion must be selected. For simplicity, a local critical energy release rate, , is set as the failure criterion. However, the formulation is not limited to this form of criterion. Effective energy release rates comprised of decomposed modes subject to power laws, etc. can easily be used with the formulation.Two primary assumptions constrain the growth formulation to make results physically meaningful. The first asserts that the front cannot retreat, equation (14). The second restricts the current energy release rate from exceeding the critical criterion value, equation (15). Physically, the current front cannot exist at energy levels above the material’s critical value, thus indicating a necessary change – i.e. the shape of the front.(14)(15)Substituting the local failure criterion, , into the general stability equation (13), yields the general growth condition.(16)This stability equation can be incorporated into an incremental loading schemewhere resulting front extensions are calculated, determining a stable shape for each load increment.Energy-Based Growth ImplementationThe energy-based growth formulation is imbedded within an incremental-iterative procedure. This procedure requires an initially stable configuration. Thedelamination is then incrementally loaded. For each load increment, a growthcondition is checked. If growth is detected, an iterative approach is employed usingthe energy-based growth formulation to achieve a stable configuration for thatgiven load increment. If growth is not detected, the algorithm is allowed to proceedto the next load increment and continue on in the simulation. Figure 8 depicts theiterative scheme of the simulation technique that incorporates finite element modelgeneration, analysis, fracture mechanics calculations, and the growth formulation.As introduced previously, the mixed-mode virtual crack extension (VCE)method is used to calculate energy release rates along the front. Energy release rates need to be extracted for both the stable configuration, { }, and after the load increment, {}. Here {}denotes a vector of quantities, i.e. the energy release rate for each point along a discretized front.At each load level, the VCE results are compared to critical values to determineif a growth condition is reached. The initial load level should be stable, meaning for all points along the front { }{}. After the load increment is applied, the energy release rates are checked again. Positions along the front are separated into mobilized, {}{ }, and stationary points, {}{ }. As the notation implies, the mobilized points are those that are expected to advance, where the stationary points, remain at their current locations. The mobilized points exceed the critical criterion for the given load increment, which is not physically attainable, indicating that an update in the front geometry is required.The stability equation is rearranged and employed to determine the portion of the load increment, , that results in the mobilized positions going from stableFigure 8. Simulation flow chart outlining use of finite element analysis, the VCE method, and the energy-based growth formulation in an iterative scheme.{ } levels to the critical values, {}. To determine , it is assumed that the front geometry is unchanged, i.e. . Herein, the mobilized notation is dropped for clarity.{}{}{ }(17)The {} term is obtained through a finite difference calculation between {} and { }. The value for each of the mobilized positions along the front is subtracted from the total load increment, , to attain the portion of the load increment that contributes energy to the system resulting in extension.{ }{}{}(18) In the next stage of the calculation, it is assumed that operations are centered at the local failure criterion. This sets {}{}. The added energy { } term is inserted back into the reorganized stability equation to calculate the delamination extensions for the mobilized positions, {}.{}{}{}{ }[]{}(19){}{}(20){}[]{}{ }(21) The [] notation signifies a matrix. Equation (21), with the use of the [] influence matrix, makes this method unique and capable of capturing arbitrary delamination growth.Each mobilized position along the front is advanced according to {}in an outward normal direction. The front geometry is updated, and re-meshed. With an updated finite element model, the iterative process is continued. The updated front geometry is loaded at the initial stable level prior to . A new { }is calculated. The load increment is applied to the updated configuration. A new {}is calculated. A new set of mobilized and stationary nodes areidentified. The previously described procedure is repeated with the new mobilized nodes, calculating the next iteration of extensions. The iterations are continued until a stable configuration is reached for the given load increment. The stability of the front geometry is achieved when, for all positions along the front, the energy release rates are below the failure criterion.SIMULATION AND RESULTSTo test the formulation and implementation of the energy-based growth formulation, a proof-of-concept simulation was designed. The finite element re-meshing, front advances, and model generations were carried out by in-house software. The models generated were imported into the ABAQUS finite element software. ABAQUS served as both the finite element environment and solver. The virtual crack extension (VCE) method used the ABAQUS results to calculate the necessary fracture mechanics parameters. The energy-based growth formulation was then employed to calculate point-by-point extensions for a given load increment.An initial, embedded elliptical delamination, aspect ratio 2:1, in an isotropic material subjected to centrally applied point loads to the surfaces was simulated, Figure 9. Simple supports were used to prevent rigid body motion. For this example, a constant local critical energy release rate was chosen as the failure criterion. This configuration was selected because of its inherently stable growth. The elliptical geometry offers an opportunity to view non-uniform growth, rather than simulating a circular delamination that grows concentrically.The load was separated into 25 increments. A stable configuration was reached for each load increment. On average, each increment required four iterations. The initially elliptical delamination evolved along the minor axis, eventually bowing into a circular configuration. Figures 10-15 show a sample of the stable configurations for selected load increments.To assess the quantitative accuracy of the energy-based growth formulation, the simulation was continued to observe the circular growth pattern (Figures 14-15). Five concentric circular growth increments were simulated. An average percent difference of 0.1% was found for the simulated radii when compared to an analytical expression [21].Figure 9. Geometry and boundary conditions for proof-of-concept simulation.。

bjnu

bjnu

北京师范大学模糊系统与人工智能方向简介(讨论稿)北京师范大学模糊数学与人工智能方向是国内最早从事模糊数学及其应用研究的单位之一,可以说是国内模糊数学研究的重要基地。

早在1979年北师大数学科学学院开始就开始招收模糊数学研究方向的硕士研究生,是我国最早从事模糊数学研究的硕士学科点。

1986年,汪培庄先生牵头,以模糊数学为主申请下来应用数学博士点,这也是我国最早从事模糊数学研究的博士学科点。

迄今为止,北师大数学科学学院已培养几十名硕士和博士研究生,并且在各种工作岗位已成为骨干力量。

北京师范大学模糊系统与模糊信息研究中心暨复杂系统实时智能控制实验室创建于2000年。

现任中心主任为国家级有突出贡献中青年专家李洪兴教授。

目前,实验室拥有博导教授2人,副教授3人,博士后2人,在读博士生15人(其中具有教授职称者2人,副教授4人),硕士研究生19人。

该研究中心现有一个应用数学的博士学位授权点,应用数学和控制理论与控制工程两个硕士学位授权点。

1982年至今,北京师范大学模糊数学与人工智能研究群体先后提出并研究了因素空间、真值流推理、随机集落影、模糊计算机、模糊摄动理论、幂结构提升理论、基于变权综合的智能信息处理、模糊系统的插值表示、变论域智能计算、复杂系统建模以及知识表示的数学理论模糊计算机等一些先进的理论方法。

近期的主要研究成果包括:1)给出因素空间理论,建立知识表示的数学框架,并系统研究概念的内涵与外延表示问题,为专家经验、领域知识在软件系统中的表示与计算提供了理论基础;2)揭示了模糊逻辑系统的数学本质,给出常用模糊逻辑系统地插值表示,并系统研究了模糊逻辑系统的构造、分析以及泛逼近性等理论问题;3)提出变论域自适应智能信息处理理论,设计了基于变论域思想的一类高精度模糊控制器,在世界上第一个实现了四级倒立摆控制实物系统,经教育部组织专家鉴定,确认这是一项原创性的具有国际领先水平的重大科研成果;4)引入变权的概念,并给出基于自适应变权理论的智能信息处理方法;5)提出模糊计算机的概念,并研究了模糊计算机设计的若干理论问题;6)给出数学神经网络理论,从数学上揭示了模糊逻辑系统与人工神经网络之间的关系,首次定义了“输出返回”的模糊逻辑系统并证明了它与反馈式神经网络等价;7)提出一种基于数据集成、规则提取和模糊推理的复杂系统的建模方法,即基于模糊推理的建模方法,由此可突破障碍模糊控制理论发展的一些瓶颈问题,诸如稳定性、能控性、能观测性等的判据问题。

Inverse Problems for Partial Differential Equations


Mathematics Subject Classification (1991): 35R30, 86A22, 80A23, 65M30, 35JXX, 35L With four illustrations.
Library of Congress Cataloging-in-Publication Data Isakov, Victor, 1947— Inverse problems for partial differential equations / Victor Isakov. p. cm. — (Applied mathematical sciences ; 127) Includes bibliographical references (p. - ) and index. ISBN 978-1-4899-0032-6 1. Inverse problems (Differential equations) 2. Differential equations, Partial. I. Title. II. Series: Applied mathematical sciences (Springer-Verlag New York Inc.) ; v. 127. QA1.A647 vol.127 [QA374] 510s-^dc21 97-16663 [515'.353] Printed on acid-free paper. © 1998 Springer Science+Business Media New York Originally published by Springer-Verlag New York, Inc. in 1998 Softcover reprint of the hardcover 1st edition 1998 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher, except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, similar or dissimilar methodology now known or hereafter developed is forbidden. The use of general descriptive names, trade names, trademarks, etc., in this publication, even if the former are not especially identified, is not to be taken as a sign that such names, as understood by the Trade Marks and Merchandise Marks Act, may accordingly be used freely by anyone. Production managed by Timothy Taylor; manufacturing supervised by Joe Quatela. Typeset by the Bartlett Press, Inc., Marietta, GA. SPIN 10628737 9 8 7 6 5 4 3 2 1 ISBN 978-1-4899-0032-6 ISBN 978-1-4899-0030-2 (eBook) DOI 10.1007/978-1-4899-0030-2

, R.J. Needs


In this paper we report an investigation of wavefunction optimization using the method of minimization of the variance of the energy 7]. We have studied three systems: the homogeneous electron gas (HEG), diamond-structure germanium with local pseudopotentials representing the ion cores, and the ground state of the corresponding germanium pseudoatom. Homogeneous systems are important in condensed matter theory because they are the simplest extended systems and they serve as a good testing ground for more realistic systems. We have developed a correlated wavefunction for the HEG which is accurate and allows for e cient optimization and rapid evaluation within a QMC calculation. We have used the electron correlation function developed for the HEG in calculations on diamondstructure germanium, adding a one-body term to introduce additional variational freedom in the charge density. Finally, we have studied the germanium pseudo-atom. The issue of most interest here is whether employing wavefunctions of similar exibility in the pseudoatom and pseudo-solid results in a large enough cancellation of errors to give a good value of the cohesive energy. For each of the three systems studied, the accuracy of the optimized wavefunctions is gauged by comparing with DMC results obtained using the optimized wavefunctions as guiding functions.

药物替代疗法英语作文

药物替代疗法英语作文下载温馨提示:该文档是我店铺精心编制而成,希望大家下载以后,能够帮助大家解决实际的问题。

文档下载后可定制随意修改,请根据实际需要进行相应的调整和使用,谢谢!并且,本店铺为大家提供各种各样类型的实用资料,如教育随笔、日记赏析、句子摘抄、古诗大全、经典美文、话题作文、工作总结、词语解析、文案摘录、其他资料等等,如想了解不同资料格式和写法,敬请关注!Download tips: This document is carefully compiled by theeditor. I hope that after you download them,they can help yousolve practical problems. The document can be customized andmodified after downloading,please adjust and use it according toactual needs, thank you!In addition, our shop provides you with various types ofpractical materials,such as educational essays, diaryappreciation,sentence excerpts,ancient poems,classic articles,topic composition,work summary,word parsing,copyexcerpts,other materials and so on,want to know different data formats andwriting methods,please pay attention!Alternative therapy, also known as complementary or holistic medicine, is gaining popularity as a substitutefor traditional medication. This approach focuses on treating the whole person, rather than just the symptoms, and often uses natural remedies and techniques. Many people turn to alternative therapy for various reasons, such as dissatisfaction with conventional medicine, desire for a more natural approach, or the belief that it can provide better results.One popular form of alternative therapy is acupuncture. This ancient Chinese practice involves inserting thin needles into specific points on the body to stimulate energy flow and promote healing. Acupuncture is believed to be effective in treating a wide range of conditions, including chronic pain, digestive disorders, and even mental health issues like anxiety and depression. Many people find relief from their symptoms through regular acupuncture sessions.Another alternative therapy gaining attention is herbal medicine. This practice utilizes plants and plant extracts to treat various ailments. Herbal remedies can be taken in the form of teas, tinctures, or supplements. For example, ginger is often used to alleviate nausea, while chamomile is known for its calming properties. Herbal medicine is seen as a more natural and gentle approach to healing, with fewer side effects compared to pharmaceutical drugs.Massage therapy is also widely used as an alternative treatment. This hands-on technique involves manipulating the body's soft tissues to improve circulation, relieve muscle tension, and promote relaxation. Different types of massage, such as Swedish, deep tissue, or aromatherapy, offer various benefits and can target specific issues. Many people find massage therapy not only physically beneficial but also emotionally and mentally soothing.In addition to these specific therapies, there are also alternative practices that focus on overall well-being. Yoga, for example, combines physical postures, breathingexercises, and meditation to promote balance and harmony in the body and mind. Tai chi, a Chinese martial art, emphasizes slow, flowing movements and deep breathing to improve flexibility, balance, and mental clarity. These practices are not only beneficial for physical health but also for reducing stress and promoting mental well-being.While alternative therapy may not be suitable for every condition or individual, it offers a different approach to healing and wellness. It is important to consult with a qualified practitioner and discuss any concerns orpotential interactions with conventional medications. Ultimately, the choice between traditional medicine and alternative therapy is a personal one, and it is essential to find the approach that works best for each individual's needs and beliefs.。

一类非光滑优化问题的邻近交替方向法

一类非光滑优化问题的邻近交替方向法钱伟懿;杨岩【摘要】非光滑优化问题在现实生活中有着广泛应用.针对一类带有结构特征为两个连续凸函数与具有Lipschitz梯度的二次可微函数的和的无约束非光滑非凸优化问题,给出了一种邻近交替方向法,称之为二次上界逼近算法.该算法结合交替方向法与邻近点算法的思想,将上述优化问题转化为平行的子问题.在求解子问题的过程中,对目标函数中的光滑部分线性化,此时子问题被转化为凸优化问题.然后分别对两个凸优化子问题交替利用邻近点算法求解.基于以上思想,首先我们给出算法的伪代码,然后建立了算法收敛性的充分条件,最后证明在该条件下,算法产生迭代序列的每个极限点是原问题的临界点.【期刊名称】《渤海大学学报(自然科学版)》【年(卷),期】2018(039)002【总页数】5页(P134-138)【关键词】非光滑优化;交替方向法;邻近点算法;收敛性分析;临界点【作者】钱伟懿;杨岩【作者单位】渤海大学数理学院, 辽宁锦州121013;渤海大学数理学院, 辽宁锦州121013【正文语种】中文【中图分类】O2210 引言考虑下列非凸非光滑的极小化问题(P) min{φ(x,y):=f(x)+g(y)+h(x,y)|x∈Rn,y∈Rm}其中函数φ有下界→(-∞,+∞),g: Rm→(-∞,+∞)都是正常的连续凸函数,h:Rn×Rm→R是具有Lipschitz梯度的二次可微函数,即存在常数L∈(0,∞),使得‖▽h(x,y)-▽h(x′,y′)‖≤L‖(x,y)-(x′,y′)‖Attouch等人〔1〕最先对问题(P)进行研究,将常规的Gauss-Seidel迭代引入算法中,给定初始点(x0,y0),由下列公式产生迭代序列{(xk,yk)}k∈N该方法被称为交替极小化方法. Gauss-Seidel方法的收敛性分析在很多文献中可见〔2-4〕,证明其收敛性的必要假设条件之一是每步迭代得到唯一最小解〔5〕,否则算法可能无限循环没有收敛性〔6〕 . 当一个变量固定,假设连续可微函数φ关于另一个变量是严格凸的,按照以上的方法迭代产生的迭代序列{(xk,yk)}k∈N 的极限点极小化目标函数φ〔3〕. 对凸光滑约束最小化问题,Beck和Tetruashvili〔7〕提出了块坐标梯度投影算法,并讨论了其全局收敛速率.去掉严格凸的假设,考虑邻近正则化Gauss-Seidel迭代其中αk,βk是正实数.该方法最先是Auslender〔8〕提出的,并进一步研究了含有非二次邻近项的线性约束凸问题〔9〕. 以上结果只是得到子序列的收敛性.当φ非凸非光滑情况下,收敛性分析是一个值得研究的课题.当φ是非凸非光滑的条件下,Attouch等人〔1,10〕证明了由邻近Gauss-Seidel 迭代〔10〕产生的序列是收敛的. 在文献〔10〕中,Attouch用熟知的proximal-forward-backward (PFB)算法求解非凸非光滑最小化问题,也得到了相似的结论. Bolte〔11〕和Daniilidis〔12〕等人在假设目标函数φ满足Kurdyka-Lojasiewicz(KL)性质的条件下,研究了一类非光滑最优化问题.交替方向法(Alternating direction method,简称ADM)最初是由Gabay和Mercier〔13〕提出. 该方法与Douglas-Rachford算子分裂算法紧密相关〔14-16〕. Eckstein〔17〕将邻近点算法(Proximal point algorithm,简称PPA)与ADM方法相结合得到了邻近交替方向法(PADM). 基于ADM方法,Beck〔18〕对凸最小化问题提出了次线性收敛速度的迭代再加权最小二乘法. Bolte和Sabach 等人〔19〕在强Kurdyka-ojasiewicz性质下对非光滑非凸优化问题提出了邻近交替线性化算法,该方法是对优化问题中光滑部分向前一个梯度步,非光滑部分向后一个邻近步,非精确求解每个线性化的子问题,迭代产生序列收敛到一个临界点. Fazel等人〔20〕提出了带半正定邻近项的交替方法,是在一定的条件下将邻近项中的正定算子扩展到半正定算子.1 预备知识本节,我们陈述一些基本概念和性质〔21〕,方便之后的证明.定义1.1 设S⊂Rn,如果对∀x1,x2∈S,0≤λ≤1,有λx1+(1-λ)x2∈S,则称S为凸集.定义1.2 设S为Rn上的凸集,如果对任意x,y ∈S,0≤λ≤1,有f(λx+(1-λ)y)≤λf(x)+(1-λ)f(y),∀x,y∈S,λ∈[0,1]则称f(x)为S上的凸函数.定理1.1 对于定义在一个开凸集O⊂Rn上的可微函数f,下面条件等价:(a)f在O上是凸函数;(b)<x1-x0,▽f(x1)-▽f(x0)>≥0对于任意的x0和x1在O上成立;(c)f(y)≥f(x)+<▽f(x),y-x>对于任意的x和y在O上成立;(d)▽2f(x)是半正定对任意的x在O上.定义1.3 函数f的次微分∂f:Rn→Rn,定义为∂f(x)={w∈Rn:f(x)-f(v)≤<w,x-v>,∀v∈Rn}若那么点称为函数f:Rn→R的临界点.定义1.4 设S⊆Rn为非空闭凸集,若f:S→R可微,其满足对任意的x,y∈S,μ>0总有f(y)≥f(x)+<▽则称f在非空闭凸集C上是μ强凸的.2 算法及收敛性分析设(1)(2)其中s∈Rn,x∈Rn,y∈Rm,t∈Rm,x∈Rn,y∈Rm. 式子(1)和(2)是将问题(P)用交替方向法产生的逼近子问题,因而称为二次上界逼近算法(Quadratic Upper-bound Approximation algorithm,简称QUA算法.QUA算法的伪代码:1. 给定初始点x0∈Rn,y0∈Rm,正实数选择正实数αk↘↘令k=0,2. k=k+13. xk+1∈arg min{ux(x,xk,yk,αk):x∈Rn}(3)4. yk+1∈arg min{uy(y,xk+1,yk,βk):y∈Rm}(4)回到第二步,直到满足某一终止条件.引理2.1 设(xk,yk)是由QUA算法迭代产生的序列,那么(5)且无穷级数和是可和的,从而有‖xk+1-xk‖→0和‖yk+1-yk‖→0.证明由二阶梯度的定义得‖▽2h(x,y)‖≤L, ∀x,y∈Rn函数h分别对x和y泰勒展开,可得下列不等式(6)(7)由αK>L ,βk>L 得f(xk+1)+g(yk)+h(xk+1,yk)≤ux(xk+1,xk,yk,αk)(8)f(xk+1)+g(yk+1)+h(xk+1,yk+1)≤uy(yk+1,xk+1,yk,βk)(9)因为在xk+1和yk+1取得极小,所以有ux(xk+1,xk,yk,αk)≤ux(xk,xk,yk,αk)=f(xk)+g(yk)+h(xk,yk)(10)uy(yk+1,xk+1,yk,βk)≤uy(yk,xk+1,yk,βk)=f(xk+1)+g(yk)+h(xk+1,yk) (11)▽xh(xk,yk)>-f(xk+1)▽yh(xk+1,yk)>-g(yk+1)应用不等式(6)和(7)得(12)(13)将不等式(12)和(13)相加得不等式(5).进一步,由不等式(5)得因此,无穷级数是可和的. 证毕.定理2.1 QUA算法迭代序列(xk,yk)的每个极限点(x*,y*)是问题(P)的临界点. 证明对迭代序列(xk,yk)的每个极限点(x*,y*)总是存在一个子序列,使得(xkj,ykj)→(x*,y*). 因为xkj+1∈arg min{ux(x,xkj,ykj,αkj):x∈Rn}(14)ykj+1∈arg min{uy(y,xkj+1,ykj,βkj):y∈Rm}(15)可得ux(xkj+1,xkj,ykj,αkj)≤ux(x,xkj,ykj,αkj), ∀x∈Rn(16)uy(ykj+1,xkj+1,ykj,βkj)≤uy(y,xkj+1,ykj,βkj), ∀y∈Rm(17)由引理2.1知‖xkj+1-xkj‖→0,‖ykj+1-ykj‖→0,从而(xkj+1,ykj+1)→(x*,y*).令j→∞得∀x∈Rn(18)∀y∈Rm(19)由最优性条件得-▽xh(x*,y*)∈∂f(x*)-▽yh(x*,y*)∈∂g(y*)极限点(x*,y*)是问题(P)的临界点.证毕.参考文献:【相关文献】〔1〕ATTOUCH H, BOLTE J, REDONT P, et al. Proximal alternating minimization and projection methods for nonconvex problems: an approach based on the Kurdyka-ojasiewicz inequality〔J〕. Mathematics of Operations Research, 2010, 35(2): 438-457. 〔2〕AUSLENDER A. Méthodes numériques pour la décomposition et la minimisation de functions non différentiables〔J〕. Numerische Mathematik, 1971, 18: 213-223.〔3〕BERTSEKAS D P, TSITSIKLIS J N. Parallel and distributed computation: numerical methods, prentice-hall〔M〕. New Jersey, 1989.〔4〕TSENG P. Convergence of block coordinate descent method for nondifferentiable minimization〔J〕. Journal of Optimization Theory and Applications, 2001, 109(3): 475-494.〔5〕ZANGWILL W L. Nonlinear programming: a unified approach〔M〕. Prentice Hall, 1971.〔6〕POWELL M. On search directions for minimization algorithms〔J〕. Mathematical Programming, 1973, 4: 193-201.〔7〕BECK A, TETRUASHVILI L. On the convergence of block coordinate descent type methods〔M〕. Preprint, 2011.〔8〕AUSLENDER A. Asymptotic properties of the fenchel dual functional and applications to decomposition problems〔J〕. Journal of Optimization Theory & Applications, 1992,73(3): 427-449.〔9〕AUSLENDER A, TEBOULLE M, BEN-TIBA S. Coupling the logarithmic-quadratic proximal method and the block nonlinear Gauss-Seidel algorithm for linearly constrained convex minimization〔J〕. 1999, 477: 35-47.〔10〕ATTOUCH H, BOLTE J, SVAITER B F. Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward-backward splitting, and regularized Gauss-Seidel methods〔J〕. Mathematical Programming, 2013, 137(1-2): 91-129.〔11〕BOLTE J, DANIILIDIS A, LEWIS A A. The ojasiewicz inequality for nonsmooth subanalytic functions with applications to subgradient dynamical systems〔J〕. Siam Journal on Optimization, 2007, 17(4): 1205-1223.〔12〕DANIILIDIS A, LEWIS A A, SHIOTAH M. Clarke subgradients of stratifiable functions 〔J〕. Siam Journal on Optimization, 2007, 18(2): 556-572.〔13〕GABAY D, B MERCIER. A dual algorithm for the solution of nonlinear variational problems via finite element approximations〔J〕. Computers & Mathematics with Applications, 1976, 2(1): 17-40.〔14〕DOUGLAS J, RACHFORD H. On the numerical solution of heat conduction problems in two and three space variables〔M〕. Transactions of the American Mathematical Society, 1956, 82(2): 421-439.〔15〕SVAITER B. On weak convergence of the Douglas-Rachford method〔M〕. Society for Industrial and Applied Mathematics, 2011.〔16〕BOT R, HENDRICH C. A Douglas-Rachford type primal-dual method for solving inclusions with mixtures of composite and parallel-sum type monotone operators〔J〕. Siam Journal on Optimization, 2012, 23(4): 2541-2565.〔17〕ECKSTEIN J. Some saddle-function splitting methods for convex programming〔J〕. Optimization Methods & Software, 1994, 4(1): 75-83.〔18〕BECK A. On the convergence of alternating minimization for convex programming with applications to Iteratively reweighted least squares and decomposition schemes〔J〕. Siam Journal on Optimization, 2013, 25(1): 185-209.〔19〕BOLTE J, SABACH S, TEBOULLE M. Proximal alternating linearized minimization for nonconvex and nonsmooth problems〔J〕. Mathematical Programming, 2014, 146(1-2): 459-494.〔20〕FAZEL M, PONG T K, SUN D, et al. Hankel matrix rank minimization with applications to system identification and realization〔J〕. Siam Journal on Matrix Analysis & Applications, 2012, 34(3): 946-977.〔21〕ROCKAFELLAR R T. Convex analysis〔M〕. Princeton University Press, 1970.。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

,
(21) (22)
d
(11) (12) (13) (14)
q (W|τ ) q ∗ (τ ) q ∗ (α)


=
k=1
k) −1 ˜ k |m( N (w ), w , (τ Λw )
= G (τ |aτ , bτ ),
d−1
=
i=1 N
G (αi |cα , dα ),
n) N (xn |m( x , Σx ), n=1
p(D, θ ) dθ q (θ ) p(D, θ ) ≥ q (θ ) ln dθ q (θ ) = L(q ),
(9)
where we have applied Jensen’s inequality. It is easily shown that maximizing the lower bound L(q ) is equivalent to minimizing the Kullback-Leibler divergence between q (θ ) and the true posterior p(θ |D).
1
Introduction
The probabilistic PCA model is defined by1 p(t|x, µ, W, τ ) = N (t|Wx + µ, τ −1 Id ), (1)
where t ∈ Rd is the observed variable, x ∈ Rq (q < d) is a latent variable with prior distribution p(x) = N (x|0, Iq ), (2)
d−1
(4)
(5) (6) (7)
=
i=1
N (wi |0, (αi τ )−1 Id ),
p(τ ) = G (τ |a0 , b0 ), where wi refers to the ith column of W. The prior over the hyper-parameters α is defined as
As shown in Appendix A, the components of the resulting variational distribution q ∗ (µ, W, τ, α, X ) have the following forms4 : q ∗ (µ, W, τ ) = q ∗ (µ|W, τ )q ∗ (W|τ )q ∗ (τ ), q ∗ (µ|W, τ ) = N (µ|Wsµ + mµ , (βµ τ )−1 Id ),
2
Prior
We define a hierarchical prior p(µ, W, τ |α) over the model parameters µ ∈ Rd , W ∈ Rd×(d−1) and τ ∈ R, governed by a vector of hyper-parameters α ∈ Rd−1 , as follows: p(µ, W, τ |α) = p(µ|W, τ )p(W|τ, α)p(τ ), with2 p(µ|W, τ ) p(W|τ, α) = N (µ|Ws0 + m0 , (β0 τ )−1 Id ),
1 For
z ∈ Rd , N (z|m, Σ) =
1 |Σ|−1/2 (2π )d/2
1 exp{− 2 (z − m)T Σ−1 (z − m)}.
1
a more accurate variational approximation was recently made in [3]; however this contains some technical errors and falls short of achieving this goal. Here we formulate a more general hierarchical prior which, while simulating a sparseness-promoting prior, has some desirable conjugacy properties. As we show, this allows for a more accurate variational approximation to the true posterior.
Improved Variational Approximation for Bayesian PCA
Shivani Agarwal and Christopher M. Bishop August 7, 2003
Abstract As with most non-trivial models, an exact Bayesian treatment of the probabilistic PCA model (under a meaningful prior) is analytically intractable. Various approximations have therefore been proposed in the literature; these include approximations based on type-II maximum likelihood as well as variational approximations. In this document, we describe an improved variational approximation for Bayesian PCA. This is achieved by defining a more general prior over the model parameters that has stronger conjugacy properties, thereby allowing for a more accurate variational approximation to the true posterior.
( i)
(15)
q (X )
=
(16)
˜ k denotes a column vector corresponding to the k th row of W, and the various parameters are given where w by βµ sµ mµ Λw
k) m( w
= β0 + N,
N −1 = βµ
(17) xn
n=1 N
β0 s0 −
,
(18)
−1 = βµ
β0 m0 +
n=1
tn
,
N T
(19) xn xT n ,
n=1
=
diag α +
N
β0 s0 sT 0
− βµ sµ sµ +
(20)
1 = Λ− w n=1
tnk xn − β0 m0k s0 + βµ mµk sµ Nd , 2
2 For
τ ∈ R, G (τ |a, b) =
1 ba τ a−1 Γ(a)
exp{−bτ }.
2
Here θ = (µ, W, τ, α, X ), where X = {xn }N n=1 is the set of latent variables corresponding to the observed data D. We consider maximizing the lower bound L(q ) subject to the following factorization constraint3 : q (µ, W, τ, α, X ) = q (µ, W, τ )q (α)q (X ). (10)
and µ ∈ Rd , W ∈ Rd×q and τ ∈ R constitute the model parameters.
d Given a set of observed data points D = {tn }N n=1 ⊂ R , the Bayesian approach consists in defining a suitable prior p(µ, W, τ ) over the model parameters and finding the posterior p(µ, W, τ |D) under this prior. The predictive density is then obtained by averaging over the model parameters according to the posterior:
d−1
p(α) =
i=1
G (αi |c0 , d0 ).
(8)
In the above, s0 ∈ Rd−1 , m0 ∈ Rd and β0 , a0 , b0 , c0 , d0 ∈ R are (fixed) hyper-parameters that can be chosen to give suitable priors.
3
Variational Posterior
We find a variational approximation to the true posterior p(µ, W, τ |D) under the above prior that maximizes a rigorous lower bound on the log likelihoodall the parameters and latent variables in the model as θ , we have for all distributions q (θ ), ln p(D) = ln = ln p(D, θ )dθ q (θ )
相关文档
最新文档