A Lower Bound for the Size of Syntactically Multilinear Arithmetic Circuits
LTE_3GPP_36.213-860(中文版)

3GPP
Release 8
3
3GPP TS 36.213 V8.6.0 (2009-03)
Contents
Foreword ...................................................................................................................................................... 5 1 2 3
Internet
Copyright Notification No part may be reproduced except as authorized by written permission. The copyright and the foregoing restriction extend to reproduction in all media.
© 2009, 3GPP Organizational Partners (ARIB, ATIS, CCSA, ETSI, TTA, TTC). All rights reserved. UMTS™ is a Trade Mark of ETSI registered for the benefit of its members 3GPP™ is a Trade Mark of ETSI registered for the benefit of its Members and of the 3GPP Organizational Partners LTE™ is a Trade Mark of ETSI currently being registered for the benefit of i ts Members and of the 3GPP Organizational Partners GSM® and the GSM logo are registered and owned by the GSM Association
链条的强度取决于它的薄弱环节英语作文

链条的强度取决于它的薄弱环节英语作文The strength of a chain is determined by its weakest link, a concept that extends beyond the literal to the metaphorical realms of life.Just as a single fragile link can compromise the entire chain, so too can a single vulnerability undermine the robustness of a system, a team, or an individual"s resolve.链条的强度取决于其最薄弱的一环,这个概念不仅仅局限于字面意义,还延伸至生活的比喻层面。
正如一个单独的脆弱环节可能危及整条链,一个系统、一个团队,或是个体决心的任何一处弱点,也可能削弱其整体的坚固性。
In any collaborative endeavor, the performance of the group is often limited by the least capable member, highlighting the importance of identifying and reinforcing these weak links if success is to be achieved.在任何一个协作努力中,团队的表现往往受限于最不擅长的成员,这凸显了在追求成功的过程中,识别并加强这些薄弱环节的重要性。
This principle is particularly relevant in fields that require precision and reliability, such as engineering, where a tiny flaw can lead to catastrophic consequences.Ignoring the weak link is a recipe for disaster, as it only takes one failure to bring down the entire structure.这一原则在需要精确性和可靠性的领域尤为重要,比如工程领域,一个小小的缺陷可能导致灾难性的后果。
siesta 使用说明

U s e r’s G u i d eS I E S T A 2.0.2December12,2008Emilio Artacho University of CambridgeJulian D.Gale Curtin University of Technology,PerthAlberto Garc´ıa Institut de Ci`e ncia de Materials,CSIC,Barcelona Javier Junquera Universidad de Cantabria,SantanderRichard M.Martin University of Illinois at Urbana-ChampaignPablo Ordej´o n Centre de Investigaci´o en Nanoci`e nciai Nanotecnologia,(CSIC-ICN),BarcelonaDaniel S´a nchez-Portal Unidad de F´ısica de Materiales,Centro Mixto CSIC-UPV/EHU,San Sebasti´a n Jos´e M.Soler Universidad Aut´o noma de Madridhttp://www.uam.es/siestasiesta@uam.esCopyright c Fundaci´o n General Universidad Aut´o noma de Madrid:E.Artacho,J.D.Gale,A.Garc´ıa,J.Junquera,P.Ordej´o n,D.S´a nchez-Portal and J.M.Soler,1996-2008Contents1INTRODUCTION4 2VERSION UPDATE63QUICK START63.1Compilation (6)3.2Running the program (6)4PSEUDOPOTENTIAL HANDLING85ATOMIC-ORBITAL BASES IMPLEMENTED IN SIESTA95.1Size:number of orbitals per atom (10)5.2Range:cutoffradii of orbitals (11)5.3Shape (11)6COMPILING THE PROGRAM127INPUT DATA FILE127.1The Flexible Data Format(FDF) (12)7.2General system descriptors (14)7.3Basis definition (16)7.4Lattice,coordinates,k-sampling (21)7.5DFT,Grid,SCF (34)7.6Eigenvalue problem:order-N or diagonalization (46)7.7Molecular dynamics and relaxations (50)7.8Parallel options (56)7.9Efficiency options (57)7.10Output options (57)7.11Options for saving/reading information (61)7.12User-provided basis orbitals (65)7.13Pseudopotentials (65)8OUTPUT FILES658.1Standard output (65)8.2Used parameters (66)18.3Array sizes (66)8.4Basis (66)8.5Pseudopotentials (67)8.6Hamiltonian and overlap matrices (67)8.7Forces on the atoms (67)8.8Sampling k points (67)8.9Charge densities and potentials (67)8.10Energy bands (67)8.11Wavefunction coefficients (68)8.12Eigenvalues (68)8.13Coordinates in specific formats (68)8.14Dynamics historyfiles (68)8.15Force Constant Matrixfile (69)8.16PHONON forcesfile (69)8.17Intermediate and restartfiles (69)9SPECIALIZED OPTIONS7010PROBLEM HANDLING7010.1Error and warning messages (70)10.2Known but unsolved problems and bugs (71)11PROJECTED CHANGES AND ADDITIONS71 12REPORTING BUGS72 13ACKNOWLEDGMENTS72 14APPENDIX:Physical unit names recognized by FDF74 15APPENDIX:NetCDF76 16APPENDIX:Parallel Siesta78 17APPENDIX:XML Output81 18APPENDIX:Selection of precision for storage832Index8331INTRODUCTIONSiesta(Spanish Initiative for Electronic Simulations with Thousands of Atoms)is both a method and its computer program implementation,to perform electronic structure calculations and ab initio molecular dynamics simulations of molecules and solids.Its main characteristics are:•It uses the standard Kohn-Sham selfconsistent density functional method in the local density(LDA-LSD)or generalized gradient(GGA)approximations.•It uses norm-conserving pseudopotentials in its fully nonlocal(Kleinman-Bylander)form.•It uses atomic orbitals as basis set,allowing unlimited multiple-zeta and angular momenta, polarization and off-site orbitals.The radial shape of every orbital is numerical and any shape can be used and provided by the user,with the only condition that it has to be of finite support,i.e.,it has to be strictly zero beyond a user-provided distance from the cor-responding nucleus.Finite-support basis sets are the key for calculating the Hamiltonian and overlap matrices in O(N)operations.•Projects the electron wavefunctions and density onto a real-space grid in order to calculate the Hartree and exchange-correlation potentials and their matrix elements.•Besides the standard Rayleigh-Ritz eigenstate method,it allows the use of localized linear combinations of the occupied orbitals(valence-bond or Wannier-like functions),making the computer time and memory scale linearly with the number of atoms.Simulations with several hundred atoms are feasible with modest workstations.•It is written in Fortran90and memory is allocated dynamically.•It may be compiled for serial or parallel execution(under MPI).(Note:This feature might not be available in all distributions.)It routinely provides:•Total and partial energies.•Atomic forces.•Stress tensor.•Electric dipole moment.•Atomic,orbital and bond populations(Mulliken).•Electron density.And also(though not all options are compatible):•Geometry relaxation,fixed or variable cell.4•Constant-temperature molecular dynamics(Nose thermostat).•Variable cell dynamics(Parrinello-Rahman).•Spin polarized calculations(collinear or not).•k-sampling of the Brillouin zone.•Local and orbital-projected density of states.•Band structure.References:•“Unconstrained minimization approach for electronic computations that scales linearly with system size”P.Ordej´o n,D.A.Drabold,M.P.Grumbach and R.M.Martin,Phys.Rev.B48,14646(1993);“Linear system-size methods for electronic-structure calcula-tions”Phys.Rev.B511456(1995),and references therein.Description of the order-N eigensolvers implemented in this code.•“Self-consistent order-N density-functional calculations for very large systems”P.Ordej´o n,E.Artacho and J.M.Soler,Phys.Rev.B53,10441,(1996).Description of a previous version of this methodology.•“Density functional method for very large systems with LCAO basis sets”D.S´a nchez-Portal,P.Ordej´o n,E.Artacho and J.M.Soler,Int.J.Quantum Chem.,65,453(1997).Description of the present method and code.•“Linear-scaling ab-initio calculations for large and complex systems” E.Artacho, D.S´a nchez-Portal,P.Ordej´o n,A.Garc´ıa and J.M.Soler,Phys.Stat.Sol.(b)215,809 (1999).Description of the numerical atomic orbitals(NAOs)most commonly used in the code, and brief review of applications as of March1999.•“Numerical atomic orbitals for linear-scaling calculations”J.Junquera,O.Paz, D.S´a nchez-Portal,and E.Artacho,Phys.Rev.B64,235111,(2001).Improved,soft-confined NAOs.•“The Siesta method for ab initio order-N materials simulation”J.M.Soler,E.Artacho, J.D.Gale,A.Garc´ıa,J.Junquera,P.Ordej´o n,and D.S´a nchez-Portal,J.Phys.:Condens.Matter14,2745-2779(2002)Extensive description of the Siesta method.•“Computing the properties of materials fromfirst principles with Siesta”,D.S´a nchez-Portal,P.Ordej´o n,and E.Canadell,Structure and Bonding113,103-170(2004).Extensive review of applications as of summer2003.For more information you can visit the web page http://www.uam.es/siesta.The following is a short description of the compilation procedures and of the datafile format for the Siesta code.52VERSION UPDATEIf you have a previous version of Siesta,the update is simply replacing the old siesta directory tree with the new one,saving the arch.makefile that you built to compile Siesta for your architecture(the format of thisfile has changed slightly,but you should be able to translate the immportanfields,such as library locations and compiler switches,to the new version). You also have the option of using the new configure script(see below)to see whether the automatically generated arch.makefile provides anything new or interesting for your setup.If you have workingfiles within the old Siesta tree,including pseudopotential etc.,you will have tofish them out.That is why we recommend working directories outside the package.3QUICK START3.1CompilationUnpack the Siesta distribution.Go to the Src directory,where the source code resides together with the Makefile.You will need afile called arch.make to suit your particular computer setup. The command./configure will start an automatic scan of your system and try to build an arch.make for you.Please note that the configure script might need some help in order tofind your Fortran compiler,and that the created arch.make may not be optimal,mostly in regard to compiler switches,but the process should provide a reasonable workingfile.Type./configure --help to see theflags understood by the script,and take a look at the Src/Confs subdirectory for some examples of their explicit use.You can also create your own arch.make by looking at the examples in the Src/Sys subdirectory.If you intend to create a parallel version of Siesta, make sure you have all the extra support libraries(MPI,scalapack,blacs...).Type make. The executable should work for any job(This is not exactly true,since some of the parameters in the atomic routines are still hardwired(see Src/atmparams.f),but those would seldom need to be changed.)3.2Running the programA fast way to test your installation of Siesta and get a feeling for the workings of the program is implemented in directory Tests.In it you canfind several subdirectories with pre-packaged FDFfiles and pseudopotential references.Everything is automated:after compiling Siesta you can just go into any subdirectory and type make.The program does its work in subdirectory work,and there you canfind all the resultingfiles.For convenience,the outputfile is copied to the parent directory.A collection of reference outputfiles can be found in Tests/Reference. Please note that small numerical and formatting differences are to be expected,depending on the compiler.Other examples are provided in the Examples directory.This directory contains basically the .fdffiles and the pseudopotential generation inputfiles.Since at some point you will have to generate your own pseudopotentials and run your own jobs,we describe here the whole process by means of the simple example of the water-molecule.It is advisable to create independent directories for each job,so that everything is clean and neat,and out of the siesta directory,6so that one can easily update version by replacing the whole siesta tree.Go to your favorite working directory and:$mkdir h2o$cd h2o$cp∼/siesta/Examples/H20/h2o.fdf.We need to generate the required pseudopotentials(We are going to streamline this process for this time,but you must realize that this is a tricky business that you must master before using Siesta responsibly.Every pseudopotential must be thoroughly checked before use.Please refer to the ATOM program manual in∼/siesta/Pseudo/atom/Docs for details regarding what follows.)$cd∼/siesta/Pseudo/atom$makeNow the pseudopotential-generation program,called atm,should be compiled(you might want to change the definition of the compiler in the makefile).$cd Tutorial/O$cat O.tm2.inpThis is the inputfile,for the oxygen pseudopotential,that we have prepared for you.It is in a standard(but obscure)format that you will need to understand in the future:------------------------------------------------------------pg Oxygentm2 2.0n=O c=ca0.00.00.00.00.00.01420 2.000.0021 4.000.00320.000.00430.000.001.15 1.15 1.15 1.15------------------------------------------------------------To generate the pseudopotential do the following;$sh../pg.sh O.tm2.inpNow there should be a new subdirectory called O.tm2(O for oxygen)and O.tm2.vps(unfor-matted)and O.tm2.psf(ASCII)files.$cp O.tm2.psf∼/whateveryourworkingdir/h2o/O.psfcopies the generated pseudopotentialfile to your working directory.(The unformatted and ASCIIfiles are functionally equivalent,but the latter is more transportable and easier to look at,if you so desire.)The same could be repeated for the pseudopotential for H,but you may as well copy H.psf from siesta/Examples/Vps/to your h2o working directory.7Now you are ready to run the program:siesta<h2o.fdf|tee h2o.out(If you are running the parallel version you should use some other invocation,such as mpirun -np2siesta...,but we cannot go into that here.)After a successful run of the program,you should have manyfiles in your directory including the following:•out.fdf(contains all the data used,explicit or default-ed)•O.ion and H.ion(complete information about the basis and KB projectors)•h2o.XV(contains thefinal positions and velocities)•h2o.STRUCT OUT(contains thefinal cell vectors and positions in“crystallographic”format)•h2o.DM(contains the density matrix to allow a restart)•h2o.ANI(contains the coordinates of every MD step,in this case only one)•h2o.FA(contains the forces on the atoms)•h2o.EIG(contains the eigenvalues of the Kohn-Sham Hamiltonian)•h2o.out(standard output)•h2o.xml(XML marked-up output)The Systemlabel.out is the standard output of the program,that you have already seen passing on the screen.Have a look at it and refer to the output-explanation section if necessary.You may also want to look at the out.fdffile to see all the default values that siesta has chosen for you,before studying the input-explanation section and start changing them.Now look at the other datafiles in Examples(all with an.fdf suffix)choose one and repeat the process for it.4PSEUDOPOTENTIAL HANDLINGThe atomic pseudopotentials are stored either in binaryfiles(with extension.vps)or in ASCII files(with extension.psf),and are read at the beginning of the execution,for each species defined in the inputfile.The datafiles must be named*.vps(or*.psf),where*is the label of the chemical species(see the ChemicalSpeciesLabel descriptor below).Thesefiles are generated by the ATOM program(read siesta/Pseudo/atom/README for more complete authorship and copyright acknowledgements).It is included(with permission)in siesta/Pseudo/atom.Remember that all pseudopotentials should be thoroughly tested before using them.We refer you to the standard literature on pseudopotentials and to the ATOM manual siesta/Pseudo/atom/atom.tex.85ATOMIC-ORBITAL BASES IMPLEMENTED IN SIESTA The main advantage of atomic orbitals is their efficiency(fewer orbitals needed per electron for similar precision)and their main disadvantage is the lack of systematics for optimal convergence, an issue that quantum chemists have been working on for many years.They have also clearly shown that there is no limitation on precision intrinsic to LCAO.This section provides some information about how basis sets can be generated for Siesta.It is important to stress at this point that neither the Siesta method nor the program are bound to the use of any particular kind of atomic orbitals.The user can feed into Siesta the atomic basis set he/she choses by means of radial tables(see User.Basis below),the only limitations being:(i)the functions have to be atomic-like(radial functions mutiplied by spherical harmonics),and(ii)they have to be offinite support,i.e.,each orbital becomes strictly zero beyond some cutoffradius chosen by the user.Most users,however,do not have their own basis sets.For these users we have devised some schemes to generate reasonable basis sets within the program.These bases depend on several parameters per atomic species that are for the user to choose,and can be important for both quality and efficiency.A description of these bases and some performance tests can be found in“Numerical atomic orbitals for linear-scaling calculations”J.Junquera,O.Paz,D.S´a nchez-Portal,and E.Artacho,Phys.Rev.B64235111,(2001)An important point here is that the basis set selection is a variational problem and,therefore, minimizing the energy with respect to any parameters defining the basis is an“ab initio”way to define them.We have also devised a quite simple and systematic way of generating basis sets based on specifying only one main parameter(the energy shift)besides the basis size.It does not offer the best NAO results one can get for a given basis size but it has the important advantages mentioned above.More about it in:“Linear-scaling ab-initio calculations for large and complex systems”E.Artacho,D.S´a nchez-Portal,P.Ordej´o n,A.Garc´ıa and J.M.Soler,Phys.Stat.Sol.(b)215,809(1999).In addition to Siesta we provide the program Gen-basis,which reads Siesta’s input and generates basisfiles for later use.Gen-basis is compiled automatically at the same time as Siesta.It should be run from the Tutorials/Bases directory,using the gen-basis.sh script. It is limited to a single species.In the following we give some clues on the basics of the basis sets that Siesta generates.The starting point is always the solution of Kohn-Sham’s Hamiltonian for the isolated pseudo-atoms, solved in a radial grid,with the same approximations as for the solid or molecule(the same exchange-correlation functional and pseudopotential),plus some way of confinement(see below). We describe in the following three main features of a basis set of atomic orbitals:size,range, and radial shape.95.1Size:number of orbitals per atomFollowing the nomenclature of Quantum Chemistry,we establish a hierarchy of basis sets,from single-ζto multiple-ζwith polarization and diffuse orbitals,covering from quick calculations of low quality to high precision,as high as thefinest obtained in Quantum Chemistry.A single-ζ(also called minimal)basis set(SZ in the following)has one single radial function per angular momentum channel,and only for those angular momenta with substantial electronic population in the valence of the free atom.It offers quick calculations and some insight on qualitative trends in the chemical bonding and other properties.It remains too rigid,however,for more quantitative calculations requiring both radial and angularflexibilization.Starting by the radialflexibilization of SZ,a better basis is obtained by adding a second function per channel:double-ζ(DZ).In Quantum Chemistry,the split valence scheme is widely used: starting from the expansion in Gaussians of one atomic orbital,the most contracted gaussians are used to define thefirst orbital of the double-ζand the most extended ones for the second.For strictly localized functions there was afirst proposal of using the excited states of the confined atoms,but it would work only for tight confinement(see PAO.BasisType nodes below).This construction was proposed and tested in D.S´a nchez-Portal et al.,J.Phys.:Condens.Matter8, 3859-3880(1996).We found that the basis set convergence is slow,requiring high levels of multiple-ζto achieve what other schemes do at the double-ζlevel.This scheme is related with the basis sets used in the OpenMX project[see T.Ozaki,Phys.Rev.B67,155108(2003);T.Ozaki and H.Kino, Phys.Rev.B69,195113(2004)].We then proposed an extension of the split valence idea of Quantum Chemistry to strictly localized NAO which has become the standard and has been used quite succesfully in many systems(see PAO.BasisType split below).It is based on the idea of suplementing thefirst ζwith,instead of a gaussian,a numerical orbital that reproduces the tail of the original PAO outside a matching radius r m,and continues smoothly towards the origin as r l(a−br2),with a and b ensuring continuity and differenciability at r m.Within exactly the same Hilbert space, the second orbital can be chosen to be the difference between the smooth one and the original PAO,which gives a basis orbital strictly confined within the matching radius r m(smaller than the original PAO!)continuously differenciable throughout.Extra parameters have thus appeared:one r m per orbital to be doubled.The user can again introduce them by hand(see PAO.Basis below).Alternatively,all the r m’s can be defined at once by specifying the value of the tail of the original PAO beyond r m,the so-called split norm.Variational optimization of this split norm performed on different systems shows a very general and stable performance for values around15%(except for the∼50%for hydrogen).It generalizes to multiple-ζtrivially by adding an additional matching radius per new zeta. Angularflexibility is obtained by adding shells of higher angular momentum.Ways to generate these so-called polarization orbitals have been described in the literature for Gaussians.For NAOs there are two ways for Siesta and Genbasis to generate them:(i)Use atomic PAO’s of higher angular momentum with suitable confinement,and(ii)solve the pseudoatom in the presence of an electricfield and obtain the l+1orbitals from the perturbation of the l orbitals by thefield.Finally,the method allows the inclusion of offsite orbitals(not centered around any specific10atom).The orbitals again can be of any shape,including atomic orbitals as if an atom would be there(useful for calculating the counterpoise correction for basis-set superposition errors). Bessel functions for any radius and any excitation level can also be added anywhere to the basis set.5.2Range:cutoffradii of orbitalsStrictly localized orbitals(zero beyond a cutoffradius)are used in order to obtain sparse Hamil-tonian and overlap matrices for linear scaling.One cutoffradius per angular momentum channel has to be given for each species.A balanced and systematic starting point for defining all the different radii is achieved by giving one single parameter,the energy shift,i.e.,the energy raise suffered by the orbital when confined.Allowing for system and physical-quantity variablity,as a rule of thumb∆E PAO≈100meV gives typical precisions within the accuracy of current GGA functionals.The user can,nevertheless,change the cutoffradii at will.5.3ShapeWithin the pseudopotential framework it is important to keep the consistency between the pseudopotential and the form of the pseudoatomic orbitals in the core region.The shape of the orbitals at larger radii depends on the cutoffradius(see above)and on the way the localization is enforced.Thefirst proposal(and quite a standard among Siesta users)uses an infinite square-well poten-tial.It was oroginally proposed and has been widely and succesfully used by Otto Sankey and collaborators,for minimal bases within the ab initio tight-binding scheme,using the Fireball program,but also for moreflexible bases using the methodology of Siesta.This scheme has the disadavantage,however,of generating orbitals with a discontinuous derivative at r c.This discontinuity is more pronounced for smaller r c’s and tends to disappear for long enough values of this cutoff.It does remain,however,appreciable for sensible values of r c for those orbitals that would be very wide in the free atom.It is surprising how small an effect such kink produces in the total energy of condensed systems.It is,on the other hand,a problem for forces and stresses,especially if they are calculated using a(coarse)finite three-dimensional grid. Another problem of this scheme is related to its defining the basis considering the free atoms. Free atoms can present extremely extended orbitals,their extension being,besides problematic, of no practical use for the calculation in condensed systems:the electrons far away from the atom can be described by the basis functions of other atoms.A traditional scheme to deal with this is the one based on the radial scaling of the orbitals by suitable scale factors.In addition to very basic bonding arguments,it is soundly based on restoring virial’s theorem forfinite bases,in the case of coulombic potentials(all-electron calculations).The use of pseudopotentials limits its applicability,allowing only for extremely small deviations from unity(∼1%)in the scale factors obtained variationally(with the exception of hydrogen that can contract up to25%).This possiblity is available to the user.Another way of dealing with that problem and that of the kink at the same time is adding a soft confinement potential to the atomic Hamiltonian used to generate the basis orbitals: it smoothens the kink and contracts the orbital as suited.Two additional parameters are11introduced for the purpose,which can be defined again variationally.The confining potential is flat(zero)in the core region,starts offat some internal radius r i with all derivatives continuous and diverges at r c ensuring the strict localization there.It isV(r)=V o e−r c−r ir−r ir c−r(1)and both r i and V o can be given to Siesta together with r c in the input(see PAO.Basis below). Finally,the shape of an orbital is also changed by the ionic character of the atom.Orbitals in cations tend to shrink,and they swell in anions.Introducing aδQ in the basis-generating free-atom calculations gives orbitals better adapted to ionic situations in the condensed systems. More information about basis sets can be found in the proposed literature.The directory Tutorials/Bases in the main Siesta distribution contains some tutorial ma-terial for the generation of basis sets and KB projectors.6COMPILING THE PROGRAMThe compilation of the program is done using a Makefile that is provided with the code.This Makefile will generate the executable for any of several architectures,with a minimum of tuning required from the user in a separatefile called arch.make to reside in the Src/directory.The instructions are in directory siesta/Src/Sys,where there are also a number of.makefiles already prepared for several architectures and operating sistems.If none of thesefit your needs, you will have to prepare one on your own.The command$./configurewill start an automatic scan of your system and try to build an arch.make for you.Please note that the configure script might need some help in order tofind your Fortran compiler,and that the created arch.make may not be optimal,mostly in regard to compiler switches,but the process should provide a reasonable workingfile.Type./configure--help to see theflags understood by the script,and take a look at the Src/Confs subdirectory for some examples of their explicit use.You canfine tune arch.make by looking at the examples in the Src/Sys subdirectory.If you intend to create a parallel version of Siesta,make sure you have all the extra support libraries(MPI,scalapack,blacs...).After arch.make is ready,ype make.The executable should work for any job(This is not exactly true,since some of the parameters in the atomic routines are still hardwired(see Src/atmparams.f),but those would seldom need to be changed.)7INPUT DATA FILE7.1The Flexible Data Format(FDF)The main inputfile,which is read as the standard input(unit5),contains all the physical data of the system and the parameters of the simulation to be performed.Thisfile is written in a12special format called FDF,developed by Alberto Garc´ıa and Jos´e M.Soler.This format allows data to be given in any order,or to be omitted in favor of default values.Refer to documentation in∼/siesta/Src/fdf for details.Here we offer a glimpse of it through the following rules:•The FDF syntax is a’data label’followed by its value.Values that are not specified in the datafile are assigned a default value.•FDF labels are case insensitive,and characters-.in a data label are ignored.Thus, LatticeConstant and lattice constant represent the same label.•All text following the#character is taken as comment.•Logical values can be specified as T,true,.true.,yes,F,false,.false.,no.Blank is also equivalent to true.•Character strings should not be in apostrophes.•Real values which represent a physical magnitude must be followed by its units.Look at function fdf convfac infile∼/siesta/Src/fdf/fdf.f for the units that are currently supported.It is important to include a decimal point in a real number to distinguish it from an integer, in order to prevent ambiguities when mixing the types on the same input line.•Complex data structures are called blocks and are placed between‘%block label’and a ‘%endblock label’(without the quotes).•You may‘include’other FDFfiles and redirect the search for a particular data label to anotherfile.If a data label appears more than once,itsfirst appearance is used.These are some examples:SystemName Water molecule#This is a commentSystemLabel h2oSpinPolarized yesSaveRhoNumberOfAtoms64LatticeConstant 5.42Ang%block LatticeVectors1.0000.0000.0000.000 1.0000.0000.0000.000 1.000%endblock LatticeVectorsKgridCutoff<BZ_sampling.fdf#Reading the coordinates from a file%block AtomicCoordinatesAndAtomicSpecies<coordinates.data#Even reading more FDF information from somewhere else%include mydefaults.fdf13。
open fpt(1)

Report from Dagstuhl Seminar12241Data Reduction and Problem KernelsEdited byMichael R.Fellows1,Jiong Guo2,Dániel Marx3,andSaket Saurabh41Charles Darwin University,AU,michael.fellows@.au2Universität des Saarlandes,DE,j.guo@mmci.uni-saarland.de3Tel A viv University,IL,dmarx@cs.bme.hu4The Institute of Mathematical Sciences–Chennai,IN,saket@imsc.res.induction and Problem Kernels”.During the seminar,several participants presented their current research,and ongoing work and open problems were discussed.Abstracts of the presentations given during the seminar as well as abstracts of seminar results and ideas are put together in this paper.Thefirst section describes the seminar topics and goals in general.Links to extended abstracts or full papers are provided,if available.Seminar10.–15.June,2012–www.dagstuhl.de/122411998ACM Subject Classification F.2Analysis of Algorithms and Problem Complexity,G.22 Graph Theory,F.1.3Complexity Measures and ClassesKeywords and phrases Preprocessing,Fixed-parameter tractability,Parameterized algorithmics Digital Object Identifier10.4230/DagRep.2.6.26Edited in cooperation with Neeldhara Misra(Institute of Mathematical Sciences,Chennai,In-dia,mail@)1Executive SummaryMichael R.FellowsJiong GuoDániel MarxSaket SaurabhLicense Creative Commons BY-NC-ND3.0Unported license©Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket SaurabhPreprocessing(data reduction or kernelization)is used universally in almost every practical computer implementation that aims to deal with an NP-hard problem.The history of preprocessing,such as applying reduction rules to simplify truth functions,can be traced back to the origins of Computer Science—the1950’s work of Quine,and much more.A modern example showing the striking power of efficient preprocessing is the commercial integer linear program solver CPLEX.The goal of a preprocessing subroutine is to solve efficiently the“easy parts”of a problem instance and reduce it(shrinking it)to its computationally difficult“core”structure(the problem kernel of the instance).How can we measure the efficiency of such a kernelization subroutine?For a long time, the mathematical analysis of polynomial time preprocessing algorithms was neglected.The basic reason for this anomalous development of theoretical computer science,was that if we seek to start with an instance I of an NP-hard problem and try tofind an efficient(P-time) Except where otherwise noted,content of this report is licensedunder a Creative Commons BY-NC-ND3.0Unported licenseDagstuhl Reports,Vol.2,Issue6,pp.26–50Editors:Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket SaurabhDagstuhl ReportsSchloss Dagstuhl–Leibniz-Zentrum für Informatik,Dagstuhl Publishing,GermanyMichael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh27 subroutine to replace I with an equivalent instance I with|I |<|I|then success would implyP=NP—discouraging efforts in this research direction,from a mathematically-poweredpoint of view.The situation in regards the systematic,mathematically sophisticated investigation of preprocessing subroutines has changed drastically with advent of parameterized complexity,where the issues are naturally framed.More specifically,we ask for upper bounds on thereduced instance sizes as a function of a parameter of the input,assuming a polynomial time reduction/preprocessing algorithm.A typical example is the famous Nemhauser-Trotter kernel for the Vertex Cover problem,showing that a“kernel"of at most2k vertices can be obtained,with k the requested maximumsize of a solution.A large number of results have been obtained in the past years,and the research in this area shows a rapid growth,not only in terms of number of papers appearingin top Theoretical Computer Science and Algorithms conferences and journals,but also interms of techniques.Importantly,very recent developments were the introduction of newlower bound techniques,showing(under complexity theoretic assumptions)that certain problems must have kernels of at least certain sizes,meta-results that show that large classesof problems all have small(e.g.,linear)kernels—these include a large collection of problemson planar graphs and matroid based techniques to obtain randomized kernels.Kernelization is a vibrant and rapidly developing area.This meeting on kernelization consolidated the results achieved in the recent years,discussed future research directions,and exploreed further the applications potential of kernelization algorithms,and gave excellent opportunities for the participants to engage in joint research and discussions on open problemsand future directions.This workshop was also special as we celebrated the60th birthday ofone of the founder of parameterized complexity,Prof.Michael R.Fellows.We organised aspecial day in which we remembered his contributions to parameterized complexity,sciencein general and mathematics for children.The main highlights of the workshop were talks on the solution to two main open problemsin the area of kernelization.We give a brief overview of these new developments below.The AND ConjectureThe OR-SAT problem asks if,given m formulas each of size n,at least one of them issatisfiable.In2008,Fortnow and Santhanam showed that if there is a reduction fromOR-SAT to any language L with the property that the reduction reduces to instances ofsize polynomial in n(independent of m)then the polynomial-time hierarchy collapses.Sucha reduction is called an OR-distillation,and this work motivated the notion of an OR-composition,which produces a boolean OR of parameterized instances of a given problem,without any restriction on the size.It was then established that an OR-composition anda polynomial kernel cannot co-exist,because these ingredients can be combined to lead toan OR-distillation.Thus,an OR-composition counts as evidence against the existence of a polynomial kernel,and it has turned into a very successful framework for establishing kernellower bounds.The question of whether there is similar evidence against the existence of an AND-distillation(defined analogously)has since been open.Such a result would imply that problems that have AND-compositions are also unlikely to admit polynomial kernels,andwould therefore be a significant addition to the kernel lower bound toolkit.The question hasbeen a central open problem for the kernelization community and was settled by Drucker inhis work on classical and quantum instance compression.The route to the result is quite involved,and forges new connections between classical and parameterized complexity.122412812241–Data Reduction and Problem KernelsTools from Matroid and Odd Cycle TraversalThe Odd Cycle Traversal problem asks if,given a graph G,there is a subset S of size atmost k whose removal makes the graph bipartite.Equivalently,the question is if there is asubset S of size at most k that intersects every odd cycle in G.The problem wasfirst shownto be FPT by Reed,Smith,and Vetta in2004,and this was also thefirst illustration of thetechnique of iterative compression.However,the question of whether the problem admits apolynomial kernel was among the main open questions in the study of kernelization.A breakthrough was recently made in work by Kratsch and Wahlström,providing thefirst(randomized)polynomial kernelization for the problem.It is a novel approach based onmatroid theory,where all relevant information about a problem instance is encoded into amatroid with a representation of size polynomial in k.Organization of the seminar and activitiesThe seminar consisted of twenty two talks,a session on open questions,and informaldiscussions among the participants.The organizers selected the talks in order to havecomprehensive lectures giving overview of main topics and communications of new researchresults.Each day consisted of talks and free time for informal gatherings among participants.On the fourth day of the seminar we celebrated the60th birthday of Mike Fellows,one ofthe founder of parameterized complexity.On this day we had several talks on the origin,history and the current developments in thefield of parameterized complexity.Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh292Table of ContentsExecutive SummaryMichael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh (26)Overview of TalksGraph decompositions for algorithms and graph structureBruno Courcelle (32)(Non)constructive advancesHans L.Bodlaender (32)Tight Compression Bounds for Problems in Graphs with Small DegeneracyMarek Cygan (32)New Evidence for the AND-and OR-ConjecturesAndrew Drucker (33)Train marshaling isfixed parameter tractableRudolf Fleischer (33)Parameterized Complexity of the Workflow Satisfiability ProblemGregory Z.Gutin (33)Faster than Courcelle’s Theorem on ShrubsPetr Hlineny (34)Preprocessing Subgraph and Minor Problems:When Does a Small Vertex CoverHelp?Bart Jansen (34)Max-Cut Parameterized Above the Edwards-Erdos BoundMark Jones (35)Data Reduction for Finding Diameter-Two SubgraphsChristian Komusiewicz (35)Kernel lower bounds using co-nondeterminism:Finding induced hereditary sub-graphsStefan Kratsch (36)Planar F-Deletion:Kernelization,Approximation and FPT Algorithms(I)Daniel Lokshtanov (36)FPT suspects and tough customers:Open problems of Downey and FellowsDániel Marx (37)Planar-F Deletion:Approximation,Kernelization and Optimal FPT Algorithms(II)Neeldhara Misra (37)Planar-F deletion in parameterized single exponential timeChristophe Paul (37)Graph separation:New incompressibility resultsMarcin Pilipczuk (38)Tight bounds for Edge Clique CoverMichal Pilipczuk (38)122413012241–Data Reduction and Problem KernelsLinear Kernels on Graphs Excluding a Topological MinorSomnath Sikdar (39)A Polynomial kernel for Proper Interval Vertex DeletionYngve Villanger (39)Uses of Matroids in KernelizationMagnus Wahlström (39)Different parameterizations of the Test Cover problemAnders Yeo (39)Open ProblemsAbove Guarantee Independent Set on Planar GraphsVenkatesh Raman (40)BicliqueMike Fellows (40)Chromatic Number of P5-free graphsFedor Fomin (41)Clique for Line SegmentsDániel Marx (41)CliquewidthDaniel Lokshtanov (41)Contraction Decomposition Beyond H-minor Free GraphsMohammadTaghi Hajiaghayi (41)Directed Feedback Vertex SetSaket Saurabh,Stefan Kratsch,and Magnus Wahlström (42)Disjoint PathsSaket Saurabh (42)Even SetMike Fellows (42)Group Feedback Edge/Vertex SetStefan Kratsch and Magnus Wahlström (43)Knapsack Parameterized by ItemsStefan Kratsch and Magnus Wahlström (43)Lower Bounds for Turing KernelsMike Fellows (43)Multiway CutStefan Kratsch and Magnus Wahlström (44)MulticutStefan Kratsch and Magnus Wahlström (44)Multiway Cut and Multicut in directed graphsStefan Kratsch and Magnus Wahlström (44)Parameterized Approximation for Dominating SetMike Fellows (45)Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh31 Polynomial Kernels for F-deletionDaniel Lokshtanov (45)Polynomial Kernel for ImbalanceSaket Saurabh (45)Quadratic Integer ProgrammingDaniel Lokshtanov (46)TreewidthHans L.Bodlaender (46)Participants (49)122413212241–Data Reduction and Problem Kernels3Overview of Talks3.1Graph decompositions for algorithms and graph structureBruno Courcelle(UniversitéBordeaux,FR)License Creative Commons BY-NC-ND3.0Unported license©Bruno CourcelleSeveral graph decompositions are important for algorithmic purposes,and not only tree-decompositions,rank-decompositions and those for clique-width.Many of them lead to"multi-kernelization"as they reduce a problem to several related problems for"prime"or"indecomposable"subgraphs.I will review the algorithmic properties and uses of several known*canonical*decomposi-tions:Tutte decomposition in3-connected components,modular decomposition and splitdecomposition.I will introduce a new one for strongly connected graphs,linked to Tutte decompositionthat I call the**atomic decomposition**.The initial motivation is the study of Gauss words(curves in the plane)but there are other applications in view.It is related but different to anoncanonical decomposition of the same graphs by Knuth(1974)3.2(Non)constructive advancesHans L.Bodlaender(Utrecht University,NL)License Creative Commons BY-NC-ND3.0Unported license©Hans L.BodlaenderThe talk surveys early results of Fellows and Langston and memorates Mike Fellows contri-butions to thefield.3.3Tight Compression Bounds for Problems in Graphs with SmallDegeneracyMarek Cygan(University of Warsaw,PL)License Creative Commons BY-NC-ND3.0Unported license©Marek CyganWe study kernelization in d-degenerate graphs.It is known that a few problems admit k O(d)kernels in d-degenerate graphs,including Induced Matching,Independent Dominating Set,Capacitated Vertex Cover,Connected Vertex Cover.Moreover a k O(d2)kernel is known forDominating Set.Simple reductions show that for Capacitated Vertex Cover and ConnectedVertex Cover kΩ(d)lower bounds exist.We show kΩ(d)lower bounds for Induced Matchingand Independent Dominating Set.Furthermore,most interestingly,we also prove kΩ(d2)lower bound for Dominating Set, which matches the known upper bound by Philip et al.[TALG]for this problem as well.Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh333.4New Evidence for the AND-and OR-ConjecturesAndrew Drucker(MIT–Cambridge,US)License Creative Commons BY-NC-ND3.0Unported license©Andrew DruckerIn the OR(SAT)problem,one is given a collection of Boolean formulas,each of length atmost k,and wants to know whether at least one is satisfiable.Similarly,in the AND(SAT) problem,one wants to know whether all the formulas are individually satisfiable.These problems are not known to have polynomial kernels.Work beginning with[Harnikand Naor’06;Bodlaender,Downey,Fellows,and Hermelin’08]has established that,ifOR(SAT)is not polynomially kernelizable,then many other natural problems fail to have polynomial kernels.Bodlaender et al.also showed that the"kernelization-hardness"ofAND(SAT)would imply a number of other hardness results.Thus,these two hypotheses, the"OR-"and"AND-conjectures,"have a great deal of explanatory power.But should webelieve them?In support of the OR-conjecture,[Fortnow and Santhanam’08]showed thatOR(SAT)does not have polynomial kernels unless NP is in coNP/poly.In this work we provide equally strong evidence for the AND-conjecture:if AND(SAT)has poly kernels then NP is in coNP/poly,and even in SZK/poly.We also extend the hardness evidence for OR(SAT)in several ways;for instance,we give thefirst strong evidenceagainst probabilistic kernelizations for OR(SAT)with two-sided bounded error.To proveour results,we exploit the information bottleneck of a kernelization reduction,using a new,general method to"disguise"information being fed into a compressive mapping.3.5Train marshaling isfixed parameter tractableRudolf Fleischer(German University of Technology–Oman,OM)License Creative Commons BY-NC-ND3.0Unported licenseFleischerThe train marshalling problem is about reordering the cars of a train using as few auxiliaryrails as possible.The problem is known to be NP-complete.We show that it isfixed parameter tractable(FPT)with the number of auxiliary rails as parameter.3.6Parameterized Complexity of the Workflow Satisfiability ProblemGregory Z.Gutin(RHUL–London,GB)Joint work of Jason Crampton,Gregory Z.Gutin and Anders Yeo.License Creative Commons BY-NC-ND3.0Unported license©Gregory Z.GutinThe Workflow Satisfiability Problem(WSP)defined below arises in Access Control in Information Security.In WSP,we are given a set S of steps and a set U of users and asked to decide whetherthere is a functionπ:S→U that satisfies some constraints.Firstly,each step can be assigned(mapped to)some subset of U.Secondly,there are some relationsρon U(ie.,ρ⊆U×U)such that all constraints of the type(ρ,S ,S”),where S ,S are subsets of S,122413412241–Data Reduction and Problem Kernelsmust be satisfied meaning that there exist s ∈S and s ∈S such that(π(s ),π(s ))∈ρ.Examples ofρinclude=and=.Wang and Li(ACM Trans.Inf.Syst.Secur.,2010)proved that WSP is NP-hard.They also observed that k=|S|is relatively small(with respect to n=|U|)and proved thatk-WSP is W[1]-hard.They obtained afixed-parameter algorithm for special cases of k-WSPwhen only relations=and=are allowed.Using a result of Bjorklund,Husfeldt and Koivisto(SIAM put.,2009)we obtain a newfixed-parameter algorithm that significantly improves the runtime of Wang and Li andwiden the special case for which k-WSP is fpt(including there organizations with hierarchicalstructures).In particular,we improve a result of Fellows,Friedrich,Hermelin,Narodytska,and Rosamond(IJCAI2011).We also investigate the existence of polynomial-size kernels andobtain both positive and negative results using,in particular,a result of Dom,Lokshtanovand Saurabh(ICALP2009).3.7Faster than Courcelle’s Theorem on ShrubsPetr Hlineny(Masaryk University,CZ)License Creative Commons BY-NC-ND3.0Unported license©Petr HlinenyURL /abs/1204.5194Famous Courcelle’s theorem claims FPT solvability of any MSO2-definable property in linearFPT time on the graphs of bounded tree-width(alternatively,of MSO1on clique-widthby Courcelle-Makowsky-Rotics).A drawback of this powerful algorithmic metatheorem isthat its runtime has a nonelementary dependence on the quantifier alternation depth of thedefining formula.This is indeed unavoidable in full generality(even on trees)as shown byFrick and Grohe.We show a new kernelization approach to this problem,giving an MSO model checking algorithm on trees of bounded height in FPT with elementary dependence on the formula;actually,we“trade”a nonelementary runtime dependence on the formula for a nonelementarydependence of our kernel on the tree height.This implies a faster(than Courcelle’s)newalgorithm for all MSO2-definable properties on the graphs of bounded tree-depth,andsimilarly a faster algorithm for all MSO1-definable properties on the classes of boundedshrub-depth.3.8Preprocessing Subgraph and Minor Problems:When Does a SmallVertex Cover Help?Bart Jansen(Utrecht University,NL)License Creative Commons BY-NC-ND3.0Unported license©Bart JansenWe prove a number of results around kernelization of problems parameterized by the vertexcover of a graph.We provide two simple general conditions characterizing problems admittingkernels of polynomial size.Our characterizations not only give generic explanations forthe existence of many known polynomial kernels for problems like Odd Cycle Transversal,Michael R.Fellows,Jiong Guo,Dániel Marx,and Saket Saurabh35 Chordal Deletion,Planarization,η-Transversal,Long Path,Long Cycle,or H-packing,theyalso imply new polynomial kernels for problems like F-Minor-Free Deletion,which is todelete at most k vertices to obtain a graph with no minor from afixedfinite set F.While our characterization captures many interesting problems,the kernelization com-plexity landscape of problems parameterized by vertex cover is much more involved.We demonstrate this by several results about induced subgraph and minor containment,whichwefind surprising.While it was known that testing for an induced complete subgraph hasno polynomial kernel unless NP is in coNP/poly,we show that the problem of testing if agraph contains a given complete graph on t vertices as a minor admits a polynomial kernel.On the other hand,it was known that testing for a path on t vertices as a minor admits a polynomial kernel,but we show that testing for containment of an induced path on t verticesis unlikely to admit a polynomial kernel.3.9Max-Cut Parameterized Above the Edwards-Erdos BoundMark Jones(RHUL–London,GB)License Creative Commons BY-NC-ND3.0Unported license©Mark JonesWe study the problem Max Cut:Given a graph,find a bipartite subgraph with the mostedges.The Edwards-Erdos bound states that for any connected graph with n vertices,medges,there is a bipartite subgraph with at least m/2+(n−1)/4edges.We study Max Cut parameterized above this bound:Given a connected graph with n vertices,m edges,decide whether there is a bipartite subgraph with at least m/2+(n−1)/4+kedges.We show that the problem isfixed-parameter tractable with running time2(3k)n O(1),and has a kernel of size O(k5).3.10Data Reduction for Finding Diameter-Two SubgraphsChristian Komusiewicz(TU Berlin,DE)License Creative Commons BY-NC-ND3.0Unported license©Christian KomusiewiczGiven an undirected graph G=(V,E)and an integer l>1,the NP-hard2-club problemasks for a vertex set S⊆V of size at least l such that G[S]has diameter at most2.We study the2-club problem with respect to many-to-one-and Turing-kernelizability fora variety of parameters such as bandwidth of G,vertex cover size of G,the dual parameter|V|−l,and the feedback edge set number of G.122413.11Kernel lower bounds using co-nondeterminism:Finding inducedhereditary subgraphsStefan Kratsch(Utrecht University,NL)License Creative Commons BY-NC-ND3.0Unported license©Stefan KratschThis work further explores the applications of co-nondeterminism for showing kerneliza-tion lower bounds.The only known example excludes polynomial kernelizations for the RAMSEY(k)problem offinding an independent set or a clique of at least k vertices in a given graph(Kratsch,SODA2012).We study the more general problem offinding in-duced subgraphs on k vertices fulfilling some hereditary propertyΠ,calledΠ-INDUCED SUBGRAPH(k).The problem is NP-hard for all non-trivial choices ofΠby a classic result of Lewis and Yannakakis(JCSS1980).The parameterized complexity of this problem was classified by Khot and Raman(TCS2002)depending on the choice ofΠ.The interesting cases for kernelization are forΠcontaining all independent sets and all cliques,since the problem is trivial or W[1]-hard otherwise.Our results are twofold.RegardingΠ-INDUCED SUBGRAPH(k),we show that for a large choice of natural graph propertiesΠ,including chordal,perfect,cluster,and cograph, there is no polynomial kernel with respect to k.This is established by two theorems:one using a co-nondeterministic variant of cross-composition and one by a polynomial parameter transformation from RAMSEY(k).Additionally,we show how to use improvement versions of NP-hard problems as source problems for lower bounds,without requiring their NP-hardness. E.g.,forΠ-INDUCED SUBGRAPH(k)our compositions may assume existing solutions of size k−1.We believe this to be useful for further lower bound proofs,since improvement versions simplify the construction of a disjunction(OR)of instances required in compositions.This adds a second way of using co-nondeterminism for lower bounds3.12Planar F-Deletion:Kernelization,Approximation and FPTAlgorithms(I)Daniel Lokshtanov(University of California–San Diego,US)License Creative Commons BY-NC-ND3.0Unported license©Daniel LokshtanovIn the F-Deletion problem you are given a graph G and integer k and asked whether there is a set S on at most k vertices such that G does not contain any minors from F,where F is afinite list of graphs.We show that if F contains at least one planar graph,then the F-Deletion problem admits polynomial kernels,constant factor approximation algorithms.If additionally all graphs in F are connected the F-Deletion problem admits c k·n time FPT algorithms.On the way we develop some new and interesting tools.Our results are stringed together by a common theme of polynomial time preprocessing3.13FPT suspects and tough customers:Open problems of Downeyand FellowsDániel Marx(MTA–Budapest,HU)License Creative Commons BY-NC-ND3.0Unported license©Dániel MarxWe give an update on the status of open problems from the book“Parameterized Complexity"by Downey and Fellows.3.14Planar-F Deletion:Approximation,Kernelization and OptimalFPT Algorithms(II)Neeldhara Misra(The Institute of Mathematical Sciences–Chennai,IN)License Creative Commons BY-NC-ND3.0Unported license©Neeldhara MisraThe notion of protrusions–constant treewidth subgraphs that can be separated from the instance by constant-sized separators–has been very useful in the context of kernelization algorithms on sparse graphs.When the optimization problem in question has certain properties,protrusions lend themselves to vastly general reduction rules,leading to a numberof interesting meta theorems on sparse graphs.Unfortunately,however,the technique is noteasily amenable to work the same way on general graphs.In particular,for the Planar F-deletion problem on general graphs,it turns out that evenfor apparently simply cases,non-trivial degree reduction rules crafted"by hand"have tocome into play before protrusion-based reductions can be applied.It is not clear that this approach is amenable to generalization for more complex cases.We therefore revisit the notion of a protrusion and introduce a moreflexible variant,namely a rmally,a near-protrusion is a subgraph which can becomea protrusion in the future,after removing some vertices of some optimal solution.The usefulness of near-protrusions is that they allow us tofind an irrelevant edge,i.e.,an edgewhich removal does not change the problem.We give a brief overview of the ideas involved in making protrusion-based reductionswork in more general situations.3.15Planar-F deletion in parameterized single exponential timeChristophe Paul(CNRS,UniversitéMontpellier II,FR)License Creative Commons BY-NC-ND3.0Unported licensePaulLet F be afinite family of graphs containing at least one planar graph.In the parameterized PLANAR-F DELETION problem,we are given an n-vertex graph G and a non-negativeinteger k(the parameter),and the question is whether G has a set X of vertices of sizeat most k such that G−X is H-minor-free for every H in F.This problem encompassesa number of well-studied parameterized problems such as Vertex Cover,Feedback VertexSet,or Treewidth-t Vertex Deletion for every value of t≥0.We present a algorithm12241for the parameterized PLANAR-F DELETION problem running in parameterized single-exponential time.Our approach significantly deviates from previous work as we do not use any reduction rule,but instead we apply a series of branching steps.This allows us to deal, in particular,with the case where the graphs in F are not necessarily connected,which was not known to admit a single-exponential algorithm3.16Graph separation:New incompressibility resultsMarcin Pilipczuk(University of Warsaw,PL)License Creative Commons BY-NC-ND3.0Unported license©Marcin PilipczukIn the talk we plan to present the recent developments on the kernelization hardness of graph separation problems.We show that,unless NP is contained in coNP/poly,the following parameterized problems do not admit a polynomial kernel:Directed Edge/Vertex Multiway Cut,parameterized by the size of the cutset,even in the case of two terminals,Edge/Vertex Multicut,parameterized by the size of the cutset,and k-Way Cut,parameterized by the size of the cutset.Our results complement very recent developments in designing parameterized algorithms for cut problems by Marx and Razgon[STOC’11],Bousquet et al.[STOC’11],Kawarabayashi and Thorup[FOCS’11]and Chitnis et al.[SODA’12].The presented results are included in the ICALP’12paper"Clique cover and graph separation:New incompressibility results"(joint work with Marek Cygan,Stefan Kratsch, Michal Pilipczuk and Magnus Wahlstrom).3.17Tight bounds for Edge Clique CoverMichal Pilipczuk(University of Bergen,NO)License Creative Commons BY-NC-ND3.0Unported license©Michal PilipczukIn the EDGE CLIQUE COVER problem,given a graph G and an integer k,we ask whether the edges of G can be covered with k complete subgraphs of G or,equivalently,whether G admits an intersection model on k-element universe.Gramm et al.[JEA2008]have shown a set of simple rules that reduce the number of vertices of G to2k,and no algorithm is known with significantly better running time bound than a brute-force search on this reduced instance.In this work we show that the approach of Gramm et al.is essentially optimal:we present a polynomial time algorithm that reduces an arbitrary3-CNF-SAT formula with n variables and m clauses to an equivalent EDGE CLIQUE COVER instance(G,k)with k=O(log n)and|V(G)|=O(n+m).This implies that EDGE CLIQUE COVER does not admit an FPT algorithm that has better than doubly-exponential running time dependency on k,unless ETH fails.Moreover,we exclude subexponential kernels for the problem under ETH and under NP not contained in coNP/poly.This refines previous work together with Stefan Kratsch and Magnus Wahlstroem[ICALP2012],in which we proved that polynomial kernelization would contradict the second complexity assumption.。
Scalar costa scheme for information embedding

Scalar Costa Scheme for InformationEmbeddingJoachim J.Eggers,Robert B¨a uml,Roman Tzschoppe,Bernd GirodAbstract1Research on information embedding and particularly information hiding techniques has received considerable attention within the last years due to its potential application in multimedia security.Digital watermarking,which is an information hiding technique where the embedded information is robust against malicious or accidental at-tacks,might offer new possibilities to enforce the copyrights of multimedia data.In this article,the specific case of information embedding into independent identically distributed(IID)data and attacks by additive white Gaussian noise(AWGN)is considered.The original data is not available to the decoder.For Gaussian data,Costa pro-posed already in1983a scheme that theoretically achieves the capacity of this communication scenario.However, Costa’s scheme is not practical.Thus,several research groups have proposed suboptimal practical communication schemes based on Costa’s idea.The goal of this artical is to give a complete performance analysis of the scalar Costa scheme(SCS)which is a suboptimal technique using scalar embedding and reception rmation theoretic bounds and simulation results with state-of-the-art coding techniques are compared.Further,reception after amplitude scaling attacks and the invertibility of SCS embedding are investigated.KeywordsInformation embedding,communication with side-information,blind digital watermarking,scalar Costa schemeI.I NTRODUCTIONESEARCH on information embedding has gained substantial attention during the last years.Thisis mainly due to the increased interest in digital watermarking technology which potentially can solve copyright infringements and data integrity disputes.Digital watermarking is considered as the im-perceptible,robust,secure communication of information by embedding it in and retrieving it from other digital data.The basic idea is that the embedded information–the watermark message–travels with the multimedia data wherever the watermarked data goes.Over the last years,many different watermarking schemes for a large variety of data types have been developed.Most of the work considers still image submitted to:IEEE Transactions on Signal Processing,Special Issue on:Signal Processing for Data Hiding in Digital Media &Secure Content Deliverydata,but watermarking of audio and video data is popular as well.Theoretical limits of digital watermark-ing have been investigated since about1999[1],[2],[3].In general,watermark embedding techniques and attacks against watermarks have to be designed specifically for certain host data types.A particularly interesting case is that of i dentically d dditive w aussian n=+Encoder DecoderFig.1.Blind watermark communication facing an AWGN attackWatermark communication as shown in Fig.1can be considered as communication with side-information at the encoder.This has beenfirst realized in1999by Chen and Wornell[7]and Cox,Miller and McK-ellips[8].Chen and Wornell introduced an important but almost forgotten paper by Costa into the wa-termarking community.Costa[9]showed theoretically that the channel capacity for the communication scenario depicted in Fig.1with an IID Gaussian host signal is(1) independent of.The suffix“ICS”stands for i osta satermark-to-n atio.The result(1)is surprising since it shows that the original data need not be considered as interference at the decoder although the decoder does not know.Costa presents a theoretic scheme which involves a random codebook which is(2)where and are realizations of two-dimensional independent random processes and with Gaussian p ensity fcalar C cheme(SCS).The accurate and complete performance analysis of SCS is the main topic of this paper.Before discussing SCS,we give a brief review of related research on the implementation of Costa’s scheme.Chen and Wornell developed in1998q ndex mistortion cdimensional embedding techniques where the dimensionality tends to infinity.This approach enables the analytical derivation of performance bounds.However,little is said about the performance of currently implementable schemes.Further,simulation results using state-of-the-art channel coding techniques are not provided.Chen and Wornell also discuss a simplification of DC-QIM where the indexed quantizers are derived via dithered prototype quantizes.This technique is investigated particularly for the case of uniform scalar prototype quantizers,which is denoted as d ompensated d odulaton(DC-DM).Chen and Wornell present a coarse performance analysis of DC-DM that is based on minimum-distance arguments and the variances of the watermark and the attack noise.However,the specific shape of the involved PDFs of the transmitted and received signals are not modelled accurately so that tight performance limits cannot be computed.Ramkumar and Akansu[12],[13],[14],[15],[16]propose a blind watermarking technique based on periodic embedding and reception functions for self-noise suppression(host signal interference reduc-tion).In particular low dimensional versions of this approach,e.g.,with scalar embedding and reception functions,are closely related to suboptimal implementations of Costa’s scheme.Ramkumar and Akansu consider during their analysis a proper modelling of the PDFs of the transmitted signals.However,their analysis of the receiver performance involves approximations that are only valid if adjacent codebook entries for identical messages are far from each other.This assumption is not valid for a large range of practically relevant s.Further,Ramkumar and Akansu present a capacity analysis based on an equivalent noise variance derived from the PDFs of the transmitted signal.This analysis is a good ap-proximation only for low s.For high s,evaluation of the presented capacity formula results in values above the Shannon limit.Nevertheless,it should be emphasized that certain versions of the technique proposed by Ramkumar and Akansu show good performance particularly for very low s. SCS outperforms their approach in the range of typical s only slightly[17].Chou et al.[18]exploit the duality of communication with side-information at the encoder to source coding with side-information at the decoder to derive a watermarking scheme based on trellis-coded quantization.This work can be considered as an extension of the research on practical implementations of Costa’s scheme in the direction of high dimensional embedding and reception rules.However,research in this direction is difficult and little progress has been made within the last years.Up to now,performance results that are better than the theoretical capacity limit of ST-SCS propose(see Sec.V)have not been test results by Chou et al.[19]show at least a slight improvement of turbo coded trellis-based constructions over simple SCS communication using coded modulation techniques.Note also thatSCS communication might still remain attractive due to its simplicity even if superior performance of high dimensional embedding techniques can be shown in future.Note also that principles of Costa’s work on communication with side information have recently gained some attention within multiuser communications[20],[21],[22].The goal of this paper is to summarize theoretical and experimental results on the performance of the practical SCS embedding and reception technique.An accurate performance analysis is based on properly derived PDFs of the transmitted and received data.Further,a comparison of the performance of state-of-the art coding techniques with theoretical performance limits is given.In Sec.II,SCS is derived formally and the encoding and decoding process is outlined.Theoretical performance limits of SCS are derived in Sec.III.Experimental results for SCS communication at high rates(low noise power)are given in Sec.IV.Sec.V discusses SCS communication at low rates,which is particularly important for robust digital watermarking.Sec.VI discusses the important extension of the AWGN attack to an attack with additional amplitude scaling.An efficient algorithm for the estimation of such amplitude scaling attacks is presented.Finally,the invertibility of SCS watermark embedding is investigated in Sec.VII,which is of interest if the distortion introduced by watermark embedding should be reduced or even removed by the legal user of a watermarked document.II.S CALAR C OSTA S CHEMEFor a practical implementation of Costa’s scheme,the usage of a suboptimal,structured codebook is proposed,while leaving the main concept of Costa’s scheme unchanged.Besides being practical,the developed scheme is independent from the data distribution.This property can be achieved for a properly chosen embedding key sequence[5],[6].When no key is used,a reasonably smooth PDF and must be assumed.To obtain a codebook with a simple structure,is chosen to be a product codebook of dithered uniform scalar quantizers,which is equivalent to an-dimensional cubic lattice[23].A.SCS EncoderFirst,the watermark message,where is a binary representation of,is encoded into a sequence of watermark letters of length.The elements belong to a-ary alphabet .-ary signaling denotes SCS watermarking with an alphabet of size. In many practical cases,binary SCS watermarking()will be used.Second,the-dimensional codebook of Costa’s scheme is structured as a product codebookof one-dimensional component codebooks,where all component codebooks are identical.For-ary signaling,the component codebook must be separated into distinct parts so that(4) The codebook is chosen to be equivalent to the representatives of a scalar uniform quantizer with step size,which is formally denoted as(6)so that each sub-codebook is equivalent to the representatives of a scalar uniform quantizer with step size .A simple and efficient encryption method for the SCS codebook is the derivation of a cryptographically secure pseudo-random sequence from the watermark key,with,and the modification of each component codebook so thatwhere denotes scalar uniform quantization with step size.Finally,the transmitted watermark sequence is given by(9) and the watermarked data is(10)A block diagram of the presented watermark embedding scheme is depicted in Fig.2.Fig.3shows an example input-output characteristic for.The embedding of can be expressed as subtractive dithered quantization,where is the dither sequence and is the step size of the uniform scalar quantizer.Note that the quantization error and thus also,is almost orthogonal to the quantizer input for an almost uniform original data PDF in the range of one quantization bin.For the given codebook encryption by a uniformly distributed key sequence,it can even be shown[24],[25] that and are statistically independent from,as it is in Costa’s ideal scheme.Further,the power of the quantization error is always E for the given distribution of the key sequence.SCS embedding depends on two parameters:the quantizer step size and the scale factor.For a given watermark power,these parameters are related byPSfrag replacementsIEEE TRANSACTIONS ON SIGNAL PROCESSING,VOL.XX,NO.Y,MONTH20028PSfragFig.3.Input-output characteristic for SCS embedding().In general,the decoding reliability can be improved by decoding an entire watermark letter sequence,where the known encoding of into can be exploited to estimate the most likely,or equivalently,toestimate the most likely watermark message.The simple codebook structure of SCS can be exploitedto efficiently estimate.First,data is extracted from the received data.This extraction processoperates sample-wise,where the extraction rule for the th element is(12)For binary SCS,,where should be close to zero if was transmitted,and close to for.Second,depending on the type of error correction encoding of,soft-input decoding algorithms,e.g.a Viterbi decoder for convolutional codes,can be used to decode from the most likely transmitted watermark message.C.Quantization Index Modulation and Dither ModulationQ ndex mither millustrations.SCS and DM are based on uniform scalar quantization with step size.It is assumed that the host data is almost uniformly distributed over the range of several quantizer bins.This assumption is reasonable in most watermarking applications,where the host-data power is much stronger than the watermark power ().Note that the introduced assumptions may no longer be valid in case of SCS embedding for strong attacks since might become quite large.For the following analysis it is not necessary to accurately model the PDF for all possible values of.It is sufficient to have an accurate model for in the range of several quantizer bins.Thus,for mathematical convenience,is considered periodic with period.With the introduced assumptions,the shape of one period of,denoted by, can be easily derived from the embedding rules for SCS and DM:DM:(13)SCS:(14)denotes the Dirac impulse and the rectangular signal is for andfor.The PDFs are almost identical to except for a shift by ,that is(17)where‘’denotes convolution.(17)is valid for.Since it is assumed thatis periodic with period,is also periodic with period.One such period ofPSfrag replacementsPSfragto the encoder and decoder.(18)shows that the capacity of communication with side information at the encoder is given by the difference of information that the codebook gives about the received data and about the side information.In general,maximization over all possible codebooks and over all corresponding embedding functions is required.Here,the capacity of the suboptimum schemes DM and SCS is considered.SCS is constrained to a codebook based on scalar quantizers,which are parameterized by and as shown in(5).and are related forfixed embedding distortion by(11).Thus,there is only one free codebook parameter forfixed embedding distortion so that the capacity of SCS is given by(19)DM is a special case of SCS with so that the capacity of DM is directly given by(20)The watermark message is encoded such that for each data element an alphabetof watermark letters is used,where each letter is equiprobable.Then the mutual information is given by[30](22) This leads with(11)topread-sPSfragocument-to-w atio.This shows that blind watermark reception suffers sig-nificantly from original signal interference.The depicted capacity of blind SS watermarking is for .For weak to moderately strong attacks(i.e.,s greater than about)SCS watermarking outperforms SS watermarking by far due to the data independent nature of SCS water-marking.However,Fig.8also reveals that for very strong attacks(),blind SS is more appropriate than SCS watermarking since here the attack distortion dominates possible interference from the original signal.Note that ICS outperforms blind SS watermarking for all s.Fig.8shows also that the binary SCS capacity is limited for high s due to the binary alphabetof watermark letters.Increasing the size of the signaling alphabet enables higher capacities for high s as shown in Fig.9.It can be observed that for very large signaling alphabets,the capacityPSfragPSfragodulation withc odes(CC-TCM)as proposed by Ungerboeck[35].-ary signaling is used in combina-tion with a new t oded m erial cPSfragA.Repetition Coding and Spread TransformThe simplest approach for the redundant embedding of the information bits into the original data is the repeated embedding of each bit.An alternative approach for redundant embedding of the information bits into the original data is the spread-transform(ST)technique as proposed in[7].We found that repetition coding with SCS performs worse than ST with SCS(ST-SCS)which is not obvious at thefirst glance.Here,we illustrate the reason for this result.Let denote the repetition factor for SCS with repetition coding,e.g.,one information bit is embedded into consecutive data elements.However,instead of deciding for each extracted value what trans-mitted watermark letter is most likely,the decoder can directly estimate the most likely transmitted watermark information bit from consecutive extracted values[4],[5].Spread transform watermarking has been proposed by Chen and Wornell[7].A detailed description of this technique can be found in[7],[4],[5].Here,we focus on the general principle.In ST watermarking, the watermark is not directly embedded into the original signal,but into the projection of onto a random sequence.Note that the term“transform”,as introduced by Chen and Wornell,is somewhat misleading since ST watermarking is mainly a pseudo-random selection of a signal component to be watermarked.All signal components orthogonal to the spreading vector remain unmodified.Let denote the spreading factor,meaning the number of consecutive original data elements belonging to one element.For watermark detection,the received data is projected onto,too.The basic idea behind ST water-marking is that any component of the channel noise that is orthogonal to the spreading vector does not impair watermark detection.Thus,an attacker,not knowing the exact spreading direction,has to introduce much larger distortions to impair a ST watermark as strong as a watermark embedded directly into.For an AWGN attack,the effective after ST with spreading factor is given by(25)Thus,doubling the spreading length gives an additional power advantage of3for the watermark in the ST domain.However,note that repetition coding and ST with and,respectively,achieve more robustness against attack noise at the cost of a reduced watermark rate.For a fair comparison of SCS with repetition coding and ST-SCS the watermark rate of both schemes should be equal,i.e.the repetition factor and the spreading factor should be equal().The s for SCS with repetition coding and ST-SCS after an AWGN attack have been measured fordifferent s.Fig.11shows simulation results for and.It has been observedthat ST-SCS yields significantly lower s than SCS with repetition coding at the same watermarkingrate.The predicted gain of3for the same decoding reliability by doubling can be observed.However,the gain for SCS with repetition coding is less than3when.The observedeffect can be explained by examining the specific structure of the codebook in SCS.PSfragdoes not affect the decision whether the transmitted bit was0or1.Further,each circle is surroundedonly by two crosses.Thus,the probability that AWGN pushes watermarked data into the area where adecoding error occurs is lower for ST-SCS than for SCS with repetition coding.15105−5−10Sfrag−15−15−10−5051015B.Capacity of ST-Watermarking and Optimal Spreading FactorST-SCS watermarking should be considered a different suboptimal approach to implement a transmis-sion scheme with side information at the encoder.Thus,the achievable rate of ST-SCS might be larger than that of SCS.Note that ST-SCS can never perform worse than SCS since SCS is a special case of ST-SCS with.The optimum choice of the spreading factor for attacks of differing noise powers is investigated.Let denote the capacity of a specific watermarking scheme combined with a spread transform with spreading factor for an AWGN attack with given.is the capacity of the respective scheme without ST.The performance of ST watermarking can be computed from that of the respective scheme without ST byurbo-cPSfragPSfragAttack−30−20−10010203010.80.60.40.2−30−20−10010203010.80.60.40.2PSfragsignal power.An exact characterization of is not necessary for our purpose.A sufficiently accurate model is given by(30)where is an appropriate constant with.The model parameters and are directly related to the unknown parameters and.determines the distance between two local maxima,anddetermines their absolute position.The exact relationship is given by(31)Fig.18depicts an example for the given model.The local maxima of the conditional PDFswith a relative distance of are clearly visible.PSfragAll spectra can be combined in an elegant way due to the systematically different phase at and.The spectra are multiplied by(33)Thus,for the model given in(30),has only one peak,which is located exactly at the frequency .Further,is superior to a multiplication by. In the latter case,the spectrum would have another peak at which increases the required sampling interval for the numerical computation of the conditional PDFs.The exact PDFs of the received signal do notfit exactly to the model given in(30).Further,in practice, the PDFs and can be only estimated from the pilot samples.This estimation is obtained from histograms with bins that cover the total range of all received samples.Based on these histograms,is computed at discrete frequencies via a length-DFT. Here,a single peak in the spectrum cannot be exptected due to estimation errors and the inaccuracy of the model(30).Nevertheless,for sufficiently large,a dominating peak should occur at. Details of the outlined implementation are described in[5],[37].B.Estimation performance for differentThe outlined algorithm for the estimation of and is dependent on the following set of param-eters::length of pilot sequence:number of histogram bins used for the pilot sample value:number of histogram bins used for the key value:DFT lengthThe estimation accuracy also depends on the,and on the.In this paper,the estimation performance for different pilot length is discussed for.This range for covers the most interesting range of attack strengths for that SCS watermarking might be useful.The has beenfixed to and the remaining parameters are, ,and.Experimental results that support this choice of parameters are given in [5].The influence of the number of received pilot elements is studied experimentally.For simplicity, and no offset has been considered so that the estimator should ideallyfind and .For the evaluation of the estimation performance,three differentfigures of merit have been used:relative error of(34)relative error of(35)Erelative increase of bit-error probabilityIEEE TRANSACTIONS ON SIGNAL PROCESSING,VOL.XX,NO.Y ,MONTH 200230PSfrag replacementsPSfrag replacements[](b)(d)Fig.19.Estimation performance for different pilot lengths (,,).For ,this effect occurs for .For ,a minimum of about -5is required.Fig.19.(b)depicts which follows in general the behavior of .The resulting relative increase of the uncoded error rateis shown with linear and logarithmic axes in Fig.19.(c)and Fig.19.(d),respectively.increases monotonically with decreasing pilot length .Further,it can beobserved again that for some low the estimation algorithm starts to fail completely.Nevertheless,it is quite promising that even for ,is lower than 2%for all.C.Estimation Based on SS Pilot SequencesSo far,an estimation of the SCS receiver parameter based on a known SCS watermark has been proposed.However,it is also possible to estimate the scale factor ,and thus ,with help of an additive spread-spectrum (SS)pilot watermark.Here,we present an analysis of the estimation accuracy,as defined in Sec.VI-B,when using SS pilot watermarks and compare the result with those for SCS pilot watermarks.We consider again the attack channel defined in (27).However,now,we assume that is a pseudo-noise sequence of length with zero mean ()and power,respectively.is known to the watermark receiver so that can be estimated from based on the correlation between and,that isEE E(40)(42)Var E EVarEE E VarSfragand.The effect of possible remaining bit errors after error correction decoding,and thus imperfect knowledge of,is not investigated.However,it is obvious that for low bit-error rates the influence of the incorrect inverse mapping applied to those samples with incorrectly received dither samples on the overall quality improvement by inverse SCS is negligible.A.Inverse SCS in the Noiseless CaseIn the noiseless case,the watermark decoder receives the signal.In this case,the deterministic embedding procedure can be inverted perfectly.With from(12)the host signal can be reconstructed with the next valid SCS codebook entry(46)bydeviation from the next valid SCS codebook entry is given by(48)For AWGN attacks,the most likely corresponding quantized original signal sample is.Thus,the MMSE estimate isEE(49)where is no longer considered within the minimization,and has to be chosen such that the MSE E is minimized.Straightforward analysis shows that has to be computed byE(50)Thus,to solve the estimation problem,the conditional PDF must be known.It is assumed that is independent from,which is approximately valid for AWGN attacks and an almostflat PDF in the range of one quantization interval,e.g.,fine quantization,so that .Thus,the random variable with support in is introduced and the PDFis considered in the following.First,Bayes’rule is applied which yieldsSfragstoring the data.This quantization can be approximated by low-power noise.In such a scenario,the in-verse scaling derived for the noiseless case might be not appropriate,but the MMSE estimation removes a good deal of the distortion introduced by the SCS watermark,as demonstrated in Fig.22.VIII.C ONCLUSIONS AND O UTLOOKInformation embedding into IID original data and an attack by AWGN has been investigated.The decoder has no access to the original data.This scenario can be considered communication with side-information at the encoder for that a theoretical communication scheme has been derived by Costa in 1983.In this paper,a suboptimal practical version of Costa’s scheme has been studied.The new scheme is named“scalar Costa scheme”(SCS)due to the involved scalar quantization during encoding and de-coding.A performance comparison of different blind watermarking schemes shows that SCS outperforms the related DM techniques for low s and performs significantly better than state-of-the-art blind SS watermarking for the relevant range of s.The latter result is mainly due to the independence of SCS from the characteristics of the original signal.SCS combined with coded modulation achieves a rate of1bit/element at,which is within1.6of the SCS capacity.For, SCS communication with rate1/3turbo coding achieves.For lower s,SCS should be combined with the spread-transform(ST)technique so that SCS operates effectively at a. Two further topics that are relevant for the usage of SCS in practical information hiding systems are investigated.These are the robustness to amplitude scaling on the watermark channel and the removal of watermark embedding distortion by authorized parties.Robustness against amplitude scaling can be achieved via robust estimation of the proper SCS quantizer step size at the receiver as described in Sec.VI. In Sec.VII,it is shown that the reduction of watermark embedding distortion is possible for low attack noise.The performance gap between SCS and ICS has to be bridged by constructing more complicated code-books and by extending the embedding and detection rule to non-scalar operations.Research in this direction has been started,e.g.,by Chou et al.[18].However,SCS might still remain an attractive technique for many information embedding applications due to its simple structure and host signal inde-pendent design.A CKNOWLEDGMENTSThe authors would like to thank Jonathan Su for the contributions to this work during his stay at the University of Erlangen-Nuremberg.Further,we would like to thank Ton Kalker for fruitfull discussionsconcerning the SCS step size estimation which helped to improve the explanation of our alorithm signifi-cantly.R EFERENCES[1] B.Chen and G.W.Wornell,“Provably robust digital watermarking,”in Proceedings of SPIE:Multimedia Systems andApplications II(part of Photonics East’99),Boston,MA,USA,September1999,vol.3845,pp.43–54.[2]P.Moulin and J.A.O’Sullivan,“Information-theoretic analysis of information hiding,”IEEE Transaction on InformationTheory,January2001.[3]J.K.Su and B.Girod,“Power-spectrum condition for energy-efficient watermarking,”in Proceedings of the IEEE Intl.Conference on Image Processing1999(ICIP’99),Kobe,Japan,October1999.[4]J.J.Eggers,J.K.Su,and B.Girod,“Performance of a practical blind watermarking scheme,”in Proc.of SPIE Vol.4314:Security and Watermarking of Multimedia Contents III,San Jose,Ca,USA,January2001.[5]Joachim J.Eggers,Information Embedding and Digital Watermarking as Communication with Side Information,Ph.D.thesis,Lehrstuhl f¨u r Nachrichtentechnik I,Universit¨a t Erlangen-N¨u rnberg,Erlangen,Germany,November2001.[6]J.Eggers and B.Girod,Informed Watermarking,Kluwer Academic Publishers,Boston,Dordrecht,London,2002.[7] B.Chen and G.W.Wornell,“Achievable performance of digital watermarking systems,”in Proceedings of the IEEE Intl.Conference on Multimedia Computing and Systems(ICMCS’99),Florence,Italy,June1999,vol.1,pp.13–18.[8]I.J.Cox,ler,and A.L.McKellips,“Watermarking as communications with side information,”Proceedings ofthe IEEE,Special Issue on Identification and Protection of Multimedia Information,vol.87,no.7,pp.1127–1141,July 1999.[9]M.H.M.Costa,“Writing on dirty paper,”IEEE Transactions on Information Theory,vol.29,no.3,pp.439–441,May1983.[10] B.Chen and G.W.Wornell,“Digital watermarking and information embedding using dither modulation,”in Proc.ofIEEE Workshop on Multimedia Signal Processing(MMSP-98),Redondo Beach,CA,USA,Dec.1998,pp.273–278. [11] B.Chen and G.Wornell,“Preprocessed and postprocessed quantization index modulation methods for digital watermark-ing,”in Proc.of SPIE Vol.3971:Security and Watermarking of Multimedia Contents II,San Jose,Ca,USA,January2000, pp.48–59.[12]M.Ramkumar and A.N.Akansu,“Self-noise suppression schemes in blind image steganography,”in Proceedings ofSPIE:Multimedia Systems and Applications II(part of Photonics East’99),Boston,MA,USA,September1999,vol.3845,pp.55–65.[13]M.Ramkumar,Data Hiding in Multimedia:Theory and Applications,Ph.D.thesis,Dep.of Electrical and ComputerEngineering,New Jersey Institute of Technology,Kearny,NJ,USA,November1999.[14]M.Ramkumar and A.N.Akansu,“FFT based signaling for multimedia steganography,”in Proceedings of the IEEE Intl.Conference on Speech and Signal Processing2000(ICASSP2000),Istanbul,Turkey,June2000.[15]M.Ramkumar and A.N.Akansu,“Floating signal constellations for multimedia steganography,”in Proceedings of theIEEE International Conference on Communications,ICC2000,New Orleans,LA,USA,June2000,vol.1,pp.249–253.[16]L.Gang,A.N.Akansu,and M.Ramkumar,“Periodic signaling scheme in oblivious data hiding,”in Proceedings of34thAsilomar Conf.on Signals,Systems,and Computers,Asilomar,CA,USA,October2000.。
链条的强度取决于它的薄弱环节英语作文

The Strength of a Chain Lies in Its WeakestLinkIn the realm of mechanical engineering, a common adage goes, "The strength of a chain lies in its weakest link." This proverb, while originally referring to the integrity of physical chains, can be aptly applied to various aspects of life, from the smallest personal endeavors to thelargest societal undertakings. The crux of the matter is that the weakest element, regardless of its size or significance, has the potential to determine the overall strength and durability of the entire system.In the context of personal development, one's weakest link could be a lack of self-discipline, a fear of failure, or a limiting belief about one's capabilities. These weak points, often hidden in the shadows of our subconscious minds, can hold us back from achieving our full potential. It is crucial to identify and address these weaknesses if we want to move forward in our personal growth and development.In the workplace, the weakest link could be a poor communication system, a lack of team cohesion, or a failureto adapt to changing market conditions. If these weaknesses are not addressed, they can lead to breakdowns in productivity, decreased morale, and ultimately, failure to meet organizational goals.On a societal level, the weakest link could be afailure to provide equal educational opportunities, a lack of infrastructure development, or a failure to address social injustices. These issues, if left unresolved, can lead to societal breakdowns, increased social tension, and a lack of progress towards collective goals.The importance of addressing the weakest link isfurther emphasized in the realm of technology. In computer systems, for example, a single faulty component can cause the entire system to crash. Similarly, in engineering projects, a minor flaw in the design or construction of a single component can lead to catastrophic failure of the entire structure.In conclusion, the strength of a chain lies in its weakest link. Whether we are considering personal growth, organizational success, or societal progress, it is imperative that we identify and address the weak pointsthat threaten the integrity of the system. By doing so, we can ensure that we are building stronger, more resilient chains that are capable of withstanding the tests of time and adversity.**链条的强度取决于它的薄弱环节**在机械工程领域,有一句谚语说道:“链条的强度取决于它的薄弱环节。
最新操作系统第九版部分课后作业习题答案分析解析

最新操作系统第九版部分课后作业习题答案分析解析CHAPTER 9 Virtual Memory Practice Exercises9.1 Under what circumstances do page faults occur? Describe the actions taken by the operating system when a page fault occurs.Answer:A page fault occurs when an access to a page that has not beenbrought into main memory takes place. The operating system veri?esthe memory access, aborting the program if it is invalid. If it is valid, a free frame is located and I/O is requested to read the needed page into the free frame. Upon completion of I/O, the process table and page table are updated and the instruction is restarted.9.2 Assume that you have a page-reference string for a process with m frames (initially all empty). The page-reference string has length p;n distinct page numbers occur in it. Answer these questions for any page-replacement algorithms:a. What is a lower bound on the number of page faults?b. What is an upper bound on the number of page faults?Answer:a. nb. p9.3 Consider the page table shown in Figure 9.30 for a system with 12-bit virtual and physical addresses and with 256-byte pages. The list of freepage frames is D, E, F (that is, D is at the head of the list, E issecond, and F is last).Convert the following virtual addresses to their equivalent physical addresses in hexadecimal. All numbers are given in hexadecimal. (A dash for a page frame indicates that the page is not in memory.)9EF1112930 Chapter 9 Virtual Memory7000FFAnswer:9E F - 0E F111 - 211700 - D000F F - EFF9.4 Consider the following page-replacement algorithms. Rank these algorithms on a ?ve-point scale from “bad” to “perfect” according to their page-fault rate. Separate those algorithms that suffer from Belady’s an omaly from those that do not.a. LRU replacementb. FIFO replacementc. Optimal replacementd. Second-chance replacementAnswer:Rank Algorithm Suffer from Belady’s anomaly1 Optimal no2 LRU no3 Second-chance yes4 FIFO yes9.5 Discuss the hardware support required to support demand paging. Answer:For every memory-access operation, the page table needs to be consulted to check whether the corresponding page is resident or not and whether the program has read or write privileges for accessing the page. These checks have to be performed in hardware. A TLB could serve as a cache and improve the performance of the lookup operation.9.6 An operating system supports a paged virtual memory, using a central processor with a cycle time of 1 microsecond. It costs an additional 1 microsecond to access a page other than the current one. Pages have 1000 words, and the paging device is a drum that rotates at 3000 revolutions per minute and transfers 1 million words per second. The following statistical measurements were obtained from the system:1 percent of all instructions executed accessed a page other than the current page.Of the instructions that accessed another page, 80 percent accesseda page already in memory.Practice Exercises 31When a new page was required, the replaced page was modi?ed 50 percent of the time.Calculate the effective instruction time on this system, assuming that the system is running one process only and that the processor is idle during drum transfers.Answer:effective access tim e = 0.99 × (1 sec + 0.008 × (2 sec)+ 0.002 × (10,000 sec + 1,000 sec)+ 0.001 × (10,000 sec + 1,000 sec)= (0.99 + 0.016 + 22.0 + 11.0) sec= 34.0 sec9.7 Consider the two-dimensional array A:int A[][] = new int[100][100];where A[0][0] is at location 200 in a paged memory system with pages of size 200. A small process that manipulates the matrix resides in page 0 (locations 0 to 199). Thus, every instruction fetch will be from page 0. For three page frames, how many page faults are generated bythe following array-initialization loops, using LRU replacement andassuming that page frame 1 contains the process and the other twoare initially empty?a. for (int j = 0; j < 100; j++)for (int i = 0; i < 100; i++)A[i][j] = 0;b. for (int i = 0; i < 100; i++)for (int j = 0; j < 100; j++)A[i][j] = 0;Answer:a. 5,000b. 509.8 Consider the following page reference string:1, 2, 3, 4, 2, 1, 5, 6, 2, 1, 2, 3, 7, 6, 3, 2, 1, 2, 3, 6.How many page faults would occur for the following replacement algorithms, assuming one, two, three, four, ?ve, six, or seven frames? Remember all frames are initially empty, so your ?rst unique pages will all cost one fault each.LRU replacementFIFO replacementOptimal replacement32 Chapter 9 Virtual MemoryAnswer:Number of frames LRU FIFO Optimal1 20 20 202 18 18 153 15 16 114 10 14 85 8 10 76 7 10 77 77 79.9 Suppose that you want to use a paging algorithm that requires a referencebit (such as second-chance replacement or working-set model), butthe hardware does not provide one. Sketch how you could simulate a reference bit even if one were not provided by the hardware, or explain why it is not possible to do so. If it is possible, calculate what the cost would be.Answer:You can use the valid/invalid bit supported in hardware to simulate the reference bit. Initially set the bit to invalid. O n ?rst reference a trap to the operating system is generated. The operating system will set a software bit to 1 and reset the valid/invalid bit to valid.9.10 You have devised a new page-replacement algorithm that you thinkmaybe optimal. In some contorte d test cases, Belady’s anomaly occurs. Is the new algorithm optimal? Explain your answer.Answer:No. An optimal algorithm will not suffer from Belady’s anomaly because —by de?nition—an optimal algorithm replaces the page that will notbe used for the long est time. Belady’s anomaly occurs when a pagereplacement algorithm evicts a page that will be needed in the immediatefuture. An optimal algorithm would not have selected such a page.9.11 Segmentation is similar to paging but uses variable-sized“pages.”De?netwo segment-replacement algorithms based on FIFO and LRU pagereplacement schemes. Remember that since segments are not the samesize, the segment that is chosen to be replaced may not be big enoughto leave enough consecutive locations for the needed segment. Consider strategies for systems where segments cannot be relocated, and thosefor systems where they can.Answer:a. FIFO. Find the ?rst segment large enough to accommodate the incoming segment. If relocation is not possible and no one segmentis large enough, select a combination of segments whose memoriesare contiguous, which are “closest to the ?rst of the list”andwhich can accommodate the new segment. If relocation is possible, rearrange the memory so that the ?rstNsegments large enough forthe incoming segment are contiguous in memory. Add any leftover space to the free-space list in both cases.Practice Exercises 33b. LRU. Select the segment that has not been used for the longestperiod of time and that is large enough, adding any leftover spaceto the free space list. If no one segment is large enough, selecta combination of the “oldest” segments that are contiguous inmemory (if relocation is not available) and that are large enough.If relocation is available, rearrange the oldest N segments to be contiguous in memory and replace those with the new segment.9.12 Consider a demand-paged computer system where the degree of multiprogramming is currently ?xed at four. The system was recently measured to determine utilization of CPU and the paging disk. The results are one of the following alternatives. For each case, what is happening? Can the degree of multiprogramming be increased to increase the CPU utilization? Is the paging helping?a. CPU utilization 13 percent; disk utilization 97 percentb. CPU utilization 87 percent; disk utilization 3 percentc. CPU utilization 13 percent; disk utilization 3 percentAnswer:a. Thrashing is occurring.b. CPU utilization is suf?ciently high to leave things alone, and increase degree of multiprogramming.c. Increase the degree of multiprogramming.9.13 We have an operating system for a machine that uses base and limit registers, but we have modi?ed the machine to provide a page table.Can the page tables be set up to simulate base and limit registers? How can they be, or why can they not be?Answer:The page table can be set up to simulate base and limit registers provided that the memory is allocated in ?xed-size segments. In this way, the base of a segment can be entered into the page table and the valid/invalid bit used to indicate that portion of the segment as resident in the memory. There will be some problem with internal fragmentation.9.27.Consider a demand-paging system with the following time-measured utilizations:CPU utilization 20%Paging disk 97.7%Other I/O devices 5%Which (if any) of the following will (probably) improve CPU utilization? Explain your answer.a. Install a faster CPU.b. Install a bigger paging disk.c. Increase the degree of multiprogramming.d. Decrease the degree of multiprogramming.e. Install more main memory.f. Install a faster hard disk or multiple controllers withmultiple hard disks.g. Add prepaging to the page fetch algorithms.h. Increase the page size.Answer: The system obviously is spending most of its time paging, indicating over-allocationof memory. If the level of multiprogramming is reduced resident processeswould page fault less frequently and the CPU utilization would improve. Another way toimprove performance would be to get more physical memory or a faster paging drum.a. Get a faster CPU—No.b. Get a bigger paging drum—No.c. Increase the degree of multiprogramming—No.d. Decrease the degree of multiprogramming—Yes.e. Install more main memory—Likely to improve CPU utilization as more pages canremain resident and not require paging to or from the disks.f. Install a faster hard disk, or multiple controllers with multiple hard disks—Also animprovement, for as the disk bottleneck is removed by faster response and morethroughput to the disks, the CPU will get more data more quickly.g. Add prepaging to the page fetch algorithms—Again, the CPU will get more datafaster, so it will be more in use. This is only the case if the paging action is amenableto prefetching (i.e., some of the access is sequential).h. Increase the page size—Increasing the page size will resultin fewer page faults ifdata is being accessed sequentially. If data access is more or less random, morepaging action could ensue because fewer pages can be kept in memory and moredata is transferred per page fault. So this change is as likely to decrease utilizationas it is to increase it.10.1、Is disk scheduling, other than FCFS scheduling, useful in asingle-userenvironment? Explain your answer.Answer: In a single-user environment, the I/O queue usually is empty. Requests generally arrive from a single process for one block or for a sequence of consecutive blocks. In these cases, FCFS is an economical method of disk scheduling. But LOOK is nearly as easy to program and will give much better performance when multiple processes are performing concurrent I/O, such as when aWeb browser retrieves data in the background while the operating system is paging and another application is active in the foreground.10.2.Explain why SSTF scheduling tends to favor middle cylindersover theinnermost and outermost cylinders.The center of the disk is the location having the smallest average distance to all other tracks.Thus the disk head tends to move away from the edges of the disk.Here is another way to think of it.The current location of the head divides the cylinders into two groups.If the head is not in the center of the disk and anew request arrives,the new request is more likely to be in the group that includes the center of the disk;thus,the head is more likely to move in that direction.10.11、Suppose that a disk drive has 5000 cylinders, numbered 0 to 4999. The drive is currently serving a request at cylinder 143, and the previous request was at cylinder 125. The queue of pending requests, in FIFO order, is86, 1470, 913, 1774, 948, 1509, 1022, 1750, 130Starting from the current head position, what is the total distance (in cylinders) that the disk arm moves to satisfy all the pending requests, for each of the following disk-scheduling algorithms?a. FCFSb. SSTFc. SCANd. LOOKe. C-SCANAnswer:a. The FCFS schedule is 143, 86, 1470, 913, 1774, 948, 1509, 1022, 1750, 130. The total seek distance is 7081.b. The SSTF schedule is 143, 130, 86, 913, 948, 1022, 1470, 1509, 1750, 1774. The total seek distance is 1745.c. The SCAN schedule is 143, 913, 948, 1022, 1470, 1509, 1750, 1774, 4999, 130, 86. The total seek distance is 9769.d. The LOOK schedule is 143, 913, 948, 1022, 1470, 1509, 1750, 1774, 130, 86. The total seek distance is 3319.e. The C-SCAN schedule is 143, 913, 948, 1022, 1470, 1509, 1750, 1774, 4999, 86, 130. The total seek distance is 9813.f. (Bonus.) The C-LOOK schedule is 143, 913, 948, 1022, 1470, 1509, 1750, 1774, 86, 130. The total seek distance is 3363.12CHAPTERFile-SystemImplementationPractice Exercises12.1 Consider a ?le currently consisting of 100 blocks. Assume that the ?lecontrol block (and the index block, in the case of indexed allocation) is already in memory. Calculate how many disk I/O operations are required for contiguous, linked, and indexed (single-level) allocation strategies, if, for one block, the following conditions hold. In the contiguous-allocation case, assume that there is no room to grow atthe beginning but there is room to grow at the end. Also assume thatthe block information to be added is stored in memory.a. The block is added at the beginning.b. The block is added in the middle.c. The block is added at the end.d. The block is removed from the beginning.e. The block is removed from the middle.f. The block is removed from the end.Answer:The results are:Contiguous Linked Indexeda. 201 1 1b. 101 52 1c. 1 3 1d. 198 1 0e. 98 52 0f. 0 100 012.2 What problems could occur if a system allowed a ?le system to be mounted simultaneously at more than one location?Answer:4344 Chapter 12 File-System ImplementationThere would be multiple paths to the same ?le, which could confuse users or encourage mistakes (deleting a ?le with one path deletes thele in all the other paths).12.3 Wh y must the bit map for ?le allocation be kept on mass storage, ratherthan in main memory?Answer:In case of system crash (memory failure) the free-space list would notbe lost as it would be if the bit map had been stored in main memory.12.4 Consider a system that supports the strategies of contiguous, linked, and indexed allocation. What criteria should be used in deciding which strategy is best utilized for a particular ?le?Answer:Contiguous—if ?le is usually accessed sequentially, if ?le isrelatively small.Linked—if ?le is large and usually accessed sequentially.Indexed—if ?le is large and usually accessed randomly.12.5 One problem with contiguous allocation is that the user must preallocate enough space for each ?le. If the ?le grows to be larger than thespace allocated for it, special actions must be taken. One solution to this problem is to de?ne a ?le structure consisting of an initial contiguous area (of a speci?ed size). If this area is ?lled, the operating system automatically de?nes an over?ow area that is linked to the initial contiguous area. If the over?ow area is ?lled, another over?ow areais allocated. Compare this implementation of a ?le with the standard contiguous and linked implementations.Answer:This method requires more overhead then the standard contiguousallocation. It requires less overheadthan the standard linked allocation. 12.6 How do caches help improve performance? Why do systems not use more or larger caches if they are so useful?Answer:Caches allow components of differing speeds to communicate moreef?ciently by storing data from the slower device, temporarily, ina faster device (the cache). Caches are, almost by de?nition, more expensive than the device they are caching for, so increasing the numberor size of caches would increase system cost.12.7 Why is it advantageous for the user for an operating system todynamically allocate its internal tables? What are the penalties to the operating system for doing so?Answer:Dynamic tables allow more ?exibility in system use growth —tablesare never exceeded, avoiding arti?cial use limits. Unfortunately, kernelstructures and code are more complicated, so there is more potentialfor bugs. The use of one resource can take away more system resources (by growing to accommodate the requests) than with static tables.Practice Exercises 4512.8 Explain how the VFS layer allows an operating system to support multiple types of ?le systems easily.Answer:VFS introduces a layer of indirection in the ?le system implementation.In many ways, it is similar to object-oriented programming techniques. System calls can be made generically (independent of ?le system type). Each ?le system type provides its function calls and data structuresto the VFS layer. A system call is translated into the proper speci?c functions for the target ?le system at the VFS layer. The calling program has no ?le-system-speci?c code, and the upper levels of the system callst ructures likewise are ?le system-independent. The translation at the VFS layer turns these generic calls into ?le-system-speci?c operations.。
Monotone

Monotone circuits for the majority functionShlomo Hoory Avner Magen†Toniann Pitassi†AbstractWe present a simple randomized construction of size O n3and depth53log n O1monotone circuits for the majority function on n variables.This result can be viewed as a reduction in the size anda partial derandomization of Valiant’s construction of an O n53monotone formula,[15].On the otherhand,compared with the deterministic monotone circuit obtained from the sorting network of Ajtai, Koml´o s,and Szemer´e di[1],our circuit is much simpler and has depth O log n with a small constant.The techniques used in our construction incorporate fairly recent results showing that expansion yields performance guarantee for the belief propagation message passing algorithms for decoding low-density parity-check(LDPC)codes,[3].As part of the construction,we obtain optimal-depth linear-size mono-tone circuits for the promise version of the problem,where the number of1’s in the input is promised to be either less than one third,or greater than two thirds.We also extend these improvements to general threshold functions.At last,we show that the size can be further reduced at the expense of increased depth,and obtain a circuit for the majority of size and depth about n1Department of Computer Science,University of British Columbia,Vancouver,Canada.†Department of Computer Science,University of Toronto,Toronto,Canada.1IntroductionThe complexity of monotone formulas/circuits for the majority function is a fascinating,albeit perplexing,problem in theoretical computer science.Without the monotonicity restriction,majority can be solvedwith simple linear-size circuits of depth O log n,where the best known depth(over binary AND,OR,NOT gates)is495log n O1[12].There are two fundamental algorithms for the majority function thatachieve logarithmic depth.Thefirst is a beautiful construction obtained by Valiant in1984[15]that achievesmonotone formulas of depth53log n O1and size O n53.The second algorithm is obtained from the celebrated sorting network constructed in1983by Ajtai,Koml´o s,and Szemer´e di[1].Restricting to binaryinputs and taking the middle output bit(median),reduces this network to a monotone circuit for the majorityfunction of depth K log n and size O n log n.The advantage of the AKS sorting network for majority is thatit is a completely uniform construction of small size.On the negative side,its proof is quite complicated andmore importantly,the constant K is huge:the best known constant K is about5000[11],and as observed byPaterson,Pippenger,and Zwick[12],this constant is important.Further converting the circuit to a formulayields a monotone formula of size O n K,which is roughly n5000.In order to argue about a quality of a solution to the problem,one should be precise about the differentresources and the tradeoffs between them.We care about the depth,the size,the number of random bitsfor a randomized construction,and formula vs circuit question.Finally,the conceptual simplicity of boththe algorithm and the correctness proof is also an important goal.Getting the best depth-size tradeoffs isperhaps the most sought after goal around this classical question,while achieving uniformity comes next. An interesting aspect of the problem is the natural way it splits into two subproblems,the solution to which gives a solution to the original problem.Problem I takes as input an arbitrary n-bit binary vector,and outputs an m-bit vector.If the input vector has a majority of1’s,then the output vector has at least a2/3fraction of 1’s,and if the input vector does not have a majority of1’s,then the output vector has at most a1/3fraction of1’s.Problem II is a promise problem that takes the m-bit output of problem I as its input.The output of Problem II is a single bit that is1if the input has at least a2/3fraction of1’s,and is a0if the input has at most a1/3fraction of1’s.Obviously the composition of these two functions solves the original majority problem.There are several reasons to consider monotone circuits that are constructed via this two-phase approach.First,Valiant’s analysis uses this viewpoint.Boppana’s later work[2]actually lower bounds each of thesesubproblems separately(although failing to provide lower bound for the entire problem).Finally,the secondsubproblem is of interest in its own right.Problem II can be viewed as an approximate counting problem,and thus plays an important role in many areas of theoretical computer science.Non monotone circuits forthis promise problem have been widely studied.Results The contribution of the current work is primarily in obtaining a new and simple construction ofmonotone circuits for the majority function of depth53log n and size O n3,hence significantly reducing the size of Valiant’s formula while not compromising at all the depth parameter.Further,for subproblem II as defined above,we supply a construction of a circuit size that is of a linear size,and it too does not compromise the depth compared to Valiant’s solution.A very appealing feature of this construction is that it is uniform,conditioned on a reasonable assumption about the existence of good enough expander graphs. To this end we introduce a connection between this circuit complexity question and another domain,namely message passing algorithms.The depth we achieve for the promise problem nearly matches the1954lower bound of Moore and Shannon[10].We further show how to generalize our solution to a general threshold function,and explore another optionin the tradeoffs between the different resources we use;specifically,we show that by allowing for a depth of roughly twice that of Valiant’s construction,we may get a circuit of size O n12Definitions and amplificationFor a monotone boolean function H on k inputs,we define its amplification function A H:0101 as A H p Pr H X1X k1,where X i are independent boolean random variables that are one with probability p.Valiant[15],considered the function H on four variables,which is the OR of two AND gates,H x1x2x3x4x1x2x3x4.The amplification function of H,depicted in Figure1,is A H p11p22,and has a non-trivialfixed point atβ512152.Let H k be the depth2k binary tree with alternating layers of AND and OR gates,where the root is labeled OR.Valiant’s construction uses the fact that A Hk is the composition of A H with itself k times.Therefore,H k probabilistically amplifiesβ∆β∆to βγεk∆βγεk∆,as long asγεk∆∆0.This implies that for any constantε0we can take2k33log n O1to probabilistically amplifyβΩ1nβΩ1n toε1ε,where33is any constant bigger thanαlogDefinition1.Let F be a boolean function F:01n01m,and let S01n be some subset of the inputs.We say that F deterministically amplifies p l p h to q l q h with respect to S,if for all inputs x S, the following promise is satisfied(we denote by x the number of ones in the vector x):F x q l m if x p l nF x q h m if x p h nNote that unlike the probabilistic amplification,deterministic amplification has to work for all inputs or scenarios in the given set S.From here on,whenever we simply say“amplification”we mean deterministic amplification.For an arbitrary small constantε0,the construction we give is composed of two independent phases that may be of independent interest.A circuit C1:01n01m for m O n that deterministically amplifiesβΩ1nβΩ1n toδ1δfor an arbitrarily small constantδ0.This circuit has size O n3and depth αεlog n O1.A circuit C2:01m01,such that C2x0if xδm and C2x1if x1δm,whereδ0is a sufficiently small constant.This circuit has size O m and depth2εlog m O1.Thefirst circuit C1is achieved by a simple probabilistic construction that resembles Valiant’s construction. We present two constructions for the second circuit,C2.Thefirst construction is probabilistic;the second construction is a simulation of a logarithmic number of rounds of a certain message passing algorithm on a good bipartite expander graph.The correctness is based on the analysis of a similar algorithm used to decode a low density parity check code(LDPC)on the erasure channel[3].Combining the two circuits together yields a circuit C:01n01for theβn-th threshold function. The circuit is of size O n3and depthα22εlog n O1.3Monotone circuits for MajorityIn this section we give a randomized construction of the circuit C:01n01such that C x is one if the portion of ones in x is at leastβn and zero otherwise.The circuit C has size O n3and depth2αεlog n O1for an arbitrary small constantε0.As we described before,we will describe C as the compositions of the circuits C1and C2whose parameters are given by the following two theorems: Theorem2.For everyεεc0,there exists a circuit C1:01n01m for m O n,of size O n3and depthαεlog n O1that deterministically amplifies all inputs fromβc nβc n toε1ε. Theorem3.For everyε0,there existsε0and a circuit C2:01n01,of size O n and depth 2εlog n O1that deterministically amplifies all inputs fromε1εto01.The two circuits use a generalization of the four input function H used in Valiant’s construction.For any integer d2,we define the function H d on d2inputs as the d-ary OR of d d-ary AND gates,i.e d i1d j1 x i j.Note that Valiant’s function H is just H2.Each of the circuits C1and C2is a layered circuit,where layer zero is the input,and each value at the i-th layer is obtained by applying H d to d2independently chosen inputs from layer i 1.However,the valuesof d we choose for C1and C2are different.For C1we have d2,while for C2we choose sufficiently large d dεto meet the depth requirement of the circuit.We let F n m F denote a random circuit mapping n inputs to m outputs,where F is afixed monotone boolean circuit with k inputs,and each of the m output bits is calculated by applying F to k independently chosen random inputs.We start with a simple lemma that relates the deterministic amplification properties of F to the probabilistic amplification function A F.1Lemma4.For anyεδ0,the random function F deterministically amplifies p l p h to A F p l1δA F p h1δwith respect to S01n with probability at least1ε,if:log S log1εmΩΘγ2εi1c nβγεγ2εi1c nThat is,we can chooseδas an increasing geometric sequence,starting fromΘ1n for i1,up toΘ1 for i logγ2εn.The implied layer size for error probability2n(which is much better than we need),is Θnδ2.Therefore,it decreases geometrically fromΘn3down toΘn.It is not difficult to see that after achieving the desired amplification fromβc n toβ∆0,only a constant number of layers is needed to get down toε.The corresponding value ofδin these last steps is a constant (that depends onε),and therefore,the required layer sizes are allΘn.Proof of Theorem3.The circuit C2is a composition of F n m1H d F m1m2H dF mt1m t H d,where d andthe layer sizes n m0m1m t1are suitably chosen parameters depending onε.We prove that with high probability such a circuit deterministically amplifies all inputs fromε1εto01.As before,we restrict our attention to the lower end of the promise problem and prove that C2outputs zero on all inputs with portion of ones smaller thanε.As in the circuit C1,the layer sizes must be sufficiently large to allow accurate computation.However, for the circuit C2,accurate computation does not mean that the portion of ones in each layer is close to its expected value.Rather,our aim is to keep the portion of ones bounded by afixed constantε,while making each layer smaller than the preceding one by approximately a factor of d.We continue this process until the layer size is constant,and then use a constant size circuit tofinish the computation.Therefore,since the number of layers of such a circuit is about log n log d,and the depth of the circuit for H d is2log d,the total depth is about2log n for large d.By the above discussion,it suffices to prove the following:For everyε0there exists a real number δ0and two integers d n0,such that for all n n0the random circuit F n m H d with m1εn d, deterministically amplifiesδtoδwith respect to all inputs,with failure probability at most1n.Since A Hδ11δd d dδd,the probability of failure for any specific input with portion of ones at most δ,is bounded by:mδmA Hδδm eamplification method to analyze the performance of a belief propagation message passing algorithm for decoding low density parity check(LDPC)codes.Today the use of belief propagation for decoding LDPC codes is one of the hottest topics in error correcting codes[9,14,13].Let G V L V R;E be a d regular bipartite graph with n vertices on each side,V L V R n.Consider the following message passing algorithm,where we think of the left and right as two players.The left player “plays AND”and the right player“plays OR”.At time zero the left player starts by sending one boolean message through each left to right edge,where the value of the message m uv from u V L to v V R is the input bit x u.Subsequently,the messages at time t0are calculated from the messages at time t 1.At odd times,given the left to right messages m uv,the right player calculates the right to left messages m vw, from v V R to w V L by the formula m vw u N v w m uv.That is,the right player sends a1along the edge from v V R to w V L if and only if at least one of the incoming messages/values(not including the incoming message from w)is1.Similarly,at even times the algorithm calculates the left to right messages m vw,v V L,w V R,from the right to left messages m uv,by the formula m vw u N v w m uv.That is,the left player sends a1along the edge from v V L to w V R if and only if all of the incoming messages/values (not including the incoming message from w)are1.We further need the following definitions.We call a left vertex bad at even time t if it transmits at least one message of value one at time t.Similarly,a right vertex is bad at odd time t if it is a right vertex that transmits at least one message of value zero at time t.We let b t be the number of bad vertices at time t.These definitions will be instrumental in providing a potential function measuring the progress of the message passing algorithm which is expressed in Lemma5.We say that a bipartite graph G V L V R;E isλe-expanding,if for any vertex set S V L(or S V R)of size at mostλn,N S e S.It will be convenient to denote the expansion of the set S by e S N S S. Lemma5.Consider the message passing algorithm using a d4regular expander graph with d1e d12.If b tλn d2then b t2b tη,whereηd1d1ηt and so b2t10for t log a d2d e gets,and the better the time guarantee above gets.How good are the expanders that we may use?One can show the existence of such expanders for sufficiently large d large,and e d c for an absolute constant c.The best known explicit construction that gets close to what we need,is the result of[4].However,that result does not suffice here for two reasons.Thefirst is that it only achieves expansion1εd for anyε0 and sufficiently large d depending onε.The second is that it only guarantees left-to-right expansion,while our construction needs both left-to-right and right-to-left expansion.We refer the reader to the survey[6] for further reading and background.For such expanders,ηd1d1log d1log d1iterations,all mes-sages contain the right answer,whereεcan be made arbitrarily small by choosing sufficiently large d.It remains to convert the algorithm into a monotone circuit,which introduces a depth-blowup of log d1 owing to the depth of a binary tree simulating a d1-ary gate.Thus we get a2εlog n-depth circuit for arbitrarily smallε0.The size is obviously dn depth O n log n.To get a linear circuit,further work is needed,which we now describe.The idea is to use a sequence of graphs G 0G G 1,where each graph is half the size of its preceding graph,but has the same degree and expansion parameters.We start the message passing algorithm using the graph G G 0,and every t 0rounds (each round consists of OR and then AND),we switch to the next graph in the sequence.Without the switch,the portion of bad vertices should decrease by a factor of ηt 0,every t 0rounds.We argue that each switch can be performed,while losing at most a constant factor.To describe the switch from G i to G i 1,we identify V L G i 1with an arbitrary half of the vertices V L G i ,and start the message passing algorithm on G i 1with the left to right messages from each vertex in V L G i 1,being the same as at the last round of the algorithm on G i .As the number of bad left vertices cannot increase at a switch,their portion,at most doubles.For the right vertices,the exact argument is slightly more involved,but it is clear that the portion of bad right vertices in the first round in G i 1,increases by at most a constant factor c ,compared with what it should have been,had there been no switch.(Precise calculation,yields c 2d η.)Therefore,to summarize,as the circuit consists of a geometrically decreasing sequence of blocks starting with a linear size block,the total size is linear as well.As for the depth,the amortized reduction in the portion of bad vertices per round,is by a factor of ηηc 1t 0.Therefore,the resulting circuit is only deeper than the one described in the previous paragraph,by a factor of log ηlog η.By choosing a sufficiently large value for t 0,we obtain:Theorem 6.For any ε0,there exists a 0such that for any n there exists a monotone circuit of depth 2εlog n O 1and size O n that solves a-promise problem.We note here that O log n depth monotone circuits for the a -promise problem can also be obtained from ε-halvers.These are building blocks used in the AKS network.However,our monotone circuits for the a -promise problem have two advantages.First,our algorithm relates this classical problem in circuit com-plexity to recent popular message passing algorithms.And second,the depth that we obtain is nearly ly,Moore and Shannon [10]prove that any monotone formula/circuit for majority requires depth 2log n O 1,and the lower bound holds for the a -promise problem as well.Proof of Lemma 5.(based on Burshtein and Miller [3])We consider only the case of bad left vertices.The proof for bad right vertices follows from the same proof,after exchanging ones with zeroes,ANDs with ORs,and lefts with rights.Let B V L be the set of bad leftvertices,and assume Bλd 2at some even time t and B the set of bad vertices at time t 2.We bound the size of B by considering separately B B and B B .Note that all sets considered in the proof have size at most λn ,and therefore expansion at least e.N(B’)To bound B B ,consider the set Q N B B N B N B B N B .Since vertices in Q are not adjacent to B ,then at time t 1they send right to left messages valued zero.On the other hand,any vertex in B B can receive at most one such zero message (otherwise all its messages at time t 2will be valuedzero and it cannot be in B).Therefore,since each vertex in Q must have at least one neighbour in B B,it follows that Q B B.Therefore,we have:N B B N B Q N B B B e B B B BOn the other hand,N B B e B B e B B B.Combining the above two inequalities,we obtain:B B e Be2B B1d12B(2) Combining inequalities(1)and(2)we get that:B B e B ed12Since e d12,and e B e,this yields the required bound:B B2d e d1As noted before in Section2,replacing the last2log n layers of Valiant’s tree with2log r n layers of r-ary AND/OR gates,results in an arbitrarily small increase in the depth of the corresponding formula for a large value of r.It is interesting to compare the expected behavior of the suggested belief-propagation algorithm to the behavior of the d1-ary tree.Assume that the graph G is chosen at random(in theconfiguration model),and that the number of rounds k is sufficiently small,d12k n.Then,almost surely the computation of all but o1fraction of the k-th round messages is performed by evaluating a d1-ary depth k trees.Moreover,introducing an additional o1error,one may assume that the leaves are independently chosen boolean random variables that are one with probability p,where p is the portion of ones in the input.This observation sheds some light on the performance of the belief propagation algorithm. However,our analysis proceeds far beyond the number of rounds for which a cycle free analysis can be applied.4Monotone formulas for threshold-k functionsConsider the case of the k-th threshold function,T k n,i.e.a function that is one on x01n if xk1and zero otherwise.We show that,by essentially the same techniques of Section3,we can construct monotone circuits to this more general problem.We assume henceforth that k n2,since otherwise, we construct the circuit T n1k n and switch AND with OR gates.For k nΘ1,the construction yields circuits of depth53log n O1and size O n3.However,when k o n,circuits are shallower and smaller (this not surprising fact is also discussed in[2]in the context of formulas).The construction goes as follows:(i)Amplify k n k1n toβΩ1kβΩ1k by randomly applying to the input a sufficiently large number of OR gates with arityΘn k(ii)AmplifyβΩ1kβΩ1k to O11O1using a variation of phase I,and(iii)Amplify O11O1to01using phase II.We now give a detailed description.For the sake of the section to follow,we require the following lemma which is more general than is needed for the results of this section.Lemma7.Let S01n,andε0.Then,for any k,there is a randomized construction of a monotone circuit that evaluates T k n correctly on all inputs from S and hasdepth log n23log k2εloglog S O1size O log S k nHere k min k n1k,and the constants of the O depend only onε.Proof.Let s log S,and let i be the OR function with arity i.Then An kk n11k n n k,while An k k1n11k1n n k.Hence An kk n is a constant bounded from zero andone.We further notice thatAn k k1nΘ1kIt is not hard to see that we can pick a constantρso that Aρn k knβΩ1k.Therefore,ρn k probabilistically amplify k n k1n toβΩ1kβΩingLemma4withδΘ1k and m sk2we get that F n mρn k amplifies k n k1n toβΩ1kβΩ1k with arbitrarily high probability.The depth required to implement the above circuit is log n k and the size is O skn.Next we apply a slight modification of phase I.The analysis there remains the same except that the starting point is separation guarantee ofΩ1k instead ofΩ1n,and log S is s instead of n.This leads to a circuit of depthαεlog k O1and of size O sk2,for an arbitrarily small constantε0.Also,we note that the output of this phase is of sizeΘs.Finally,we apply phase II,where the number of inputs isΘs instead ofΘn,to obtain an amplification from O11O1to01.This requires depth2εlog s O1and size O s,for an arbitrarily small constantε0.To guarantee the correctness of a monotone circuit for T n k,it suffices to check its output on inputs of weight k k1(as the circuit is monotone).Therefore,S n k n k1,implying that log S O k log n k. Therefore,we have:Theorem8.There is a randomized construction of a monotone circuit for T k n with:depth log n43log k O loglog n ksize O k2n log n kwhere k min k n1k,and the constants of the O are absolute.5Reducing the circuit sizeThe result obtained so far for the majority,is a monotone circuit of depth53log n O1and size O n3.In this section,we would like to obtain smaller circuit size,at the expense of increasing the depth somewhat. The crucial observation is that the size of our circuit depends linearly on the logarithm of the number of scenarios it has to handle.Therefore,applying a preprocessing stage to reduce the wealth of scenarios may save up to a factor of n in the circuit size.We propose a recursive construction that reduces the circuit size to about n1We chooseαi2σi1to equate1αiσi with3αi.This implies thatσi132σi1,and δi153δi22σi1,resulting in the following sequence:iαiσiδi2,and the sequence of δi tends to129896.Therefore,we have:Theorem9.There is a randomized construction of a monotone circuit for the majority of size n1There are two central open problems related to this work.First,is the promise version really simpler than majority?A lower bound greater than2log n on the communication complexity of mMaj-search would settle this question.Boppana[2]and more recent work[5]show lower bounds on a particular method for obtaining monotone formulas for majority.However we are asking instead for lower bounds on the size/depth of unrestricted monotone formulas/circuits.Secondly,the original question remains unresolved. Namely,we would like to obtain explicit uniform formulas for majority of optimal or near optimal size.A related problem is to come up with a natural(top-down)communication complexity protocol for mMaj-Search that uses O log n many bits.References[1]M.Ajtai,J.Koml´o s,and E.Szemer´e di.Sorting in c log n parallel binatorica,3(1):1–19,1983.[2]R.B.Boppana.Amplification of probabilistic boolean formulas.IEEE Symposium on Foundations ofComputer Science(FOCS),pages20–29,1985.[3]D.Burshtein and ler.Expander graph arguments for message-passing algorithms.IEEE Trans.Inform.Theory,47(2):782–790,2001.[4]M.Capalbo,O.Reingold,S.Vadhan,and A.Wigderson.Randomness conductors and constant-degreeexpansion beyond the degree2barrier.In Proceedings34th Symposium on Theory of Computing, pages659–668,2002.[5]M.Dubiner and U.Zwick.Amplification by read-once formulas.SIAM put.,26(1):15–38,1997.[6]S.Hoory,N.Linial,and A.Wigderson.Expander graphs and their applications.survey article toappear in the Bulletin of the AMS.[7]Mauricio Karchmer and Avi Wigderson.Monotone circuits for connectivity require super-logarithmicdepth.In Proceedings of the Twentieth Annual ACM Symposium on Theory of Computing,pages539–550,Chicago,IL,May1988.[8]M.Luby,M.Mitzenmacher,and A.Shokrollahi.Analysis of random processes via and-or tree evalu-ation.In ACM-SIAM Symposium on Discrete Algorithms(SODA),1998.[9]M.Luby,M.Mitzenmacher,A.Shokrollahi,and D.A.Spielman.Analysis of low density codes andimproved designs using irregular graphs.ACM Symposium on Theory of Computing(STOC),1998.[10]E.F.Moore and C.E.Shannon.Reliable circuits using less reliable relays.I,II.J.Franklin Inst.,262:191–208,281–297,1956.[11]M.S.Paterson.Improved sorting networks with O log N depth.Algorithmica,5(1):75–92,1990.[12]M.S.Paterson,N.Pippenger,and U.Zwick.Optimal carry save networks.In Boolean functioncomplexity(Durham,1990),volume169of London Math.Soc.Lecture Note Ser.,pages174–201.Cambridge Univ.Press,Cambridge,1992.[13]T.Richardson and R.Urbanke.Modern coding theory.Draft of a book.[14]T.Richardson and R.Urbanke.The capacity of low-density parity-check codes under message-passingdecoding.IEEE rm.Theory,47(2):599–618,2001.[15]L.G.Valiant.Short monotone formulae for the majority function.J.Algorithms,5(3):363–366,1984.。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
1.1
Syntactically Multilinear Arithmetic Circuits
An arithmetic circuit Φ over the field F and the set of variables X = {x1 , . . . , xn } is a directed acyclic graph as follows: Every vertex v in Φ is either of in-degree 0 or of in-degree 2. Every vertex v in Φ of in-degree 0 is labelled by either a variable in X or a field element in F. Every vertex v in Φ of in-degree 2 is labelled by either × or +. An arithmetic circuit Φ is called an arithmetic formula if Φ is a directed binary tree (the edges of an arithmetic formula are directed from the leaves to the root). Let Φ be an arithmetic circuit over the field F and the set of variables X . The vertices of Φ are also called gates. Every gate of in-degree 0 is called an input gate. Every gate of in-degree 2 labelled by × is called a product gate. Every gate of in-degree 2 labelled by + is called an addition gate. Every gate of out-degree 0 is called an output gate. For two gates u and v in Φ, if (u, v ) is an edge in Φ, then u is called a son of v , and v is called a father of u. The size of Φ, denoted |Φ|, is the number of edges in Φ. Since the in-degree of Φ is at most 2, the size of Φ is at most twice the number of gates in Φ. For a gate v in Φ, define Φv to be the sub-circuit of Φ rooted at v as follows: The gates of Φv are all the gates u in Φ such that there exists a directed path from u to v in Φ. The edges and labels of Φv are the same edges and labels of Φ (restricted to the set of gates of Φv ). An arithmetic circuit computes a polynomial in a natural way. For a gate v in Φ, define Φv ∈ F[X ] to be the polynomial computed by Φv as follows: If v is an input gate labelled by α ∈ F ∪ X , then Φv = α. If v is a product gate with sons v1 and v2 , then Φv = Φv1 · Φv2 . If v is an addition gate with sons v1 and v2 , then Φv = Φv1 + Φv2 . For a polynomial f ∈ F[X ], and a gate v in Φ, we say that v computes f if f = Φv . For a polynomial f ∈ F[X ], we say that Φ computes f if there exists a gate u in Φ that computes f .
∗
1
ISSN 1433-8092
multilinear arithmetic circuits. We prove an Ω(n4/3 / log2 n) lower bound for the size of syntactically multilinear arithmetic circuits computing an explicit polynomial.
Faculty of Mathematics and Computer Science, Weizmann Institute, Rehovot, Israel, and Microsoft Research. Email: ran.raz@weizmann.ac.il. Research supported by grants from the Binational Science Foundation (BSF), the Israel Science Foundation (ISF), and the Minerva Foundation. † Faculty of Computer Science, Technion, Israel. Email: shpilka@cs.technion.ac.il. ‡ Faculty of Mathematics and Computer Science, Weizmann Institute, Rehovot, Israel. Email: amir.yehudayoff@weizmann.ac.il. Research supported by grants from the Binational Science Foundation (BSF), the Israel Science Foundation (ISF), and the Minerva Foundation.
1.2
Background and Motivays to define multilinear arithmetic circuits: a syntactic definition, and a semantic definition, as described above. The semantic definition is a natural one, but the syntactic definition is 2
Electronic Colloquium on Computational Complexity, Report No. 60 (2006)
A Lower Bound for the Size of Syntactically Multilinear Arithmetic Circuits
Ran Raz
1
Introduction
Arithmetic circuits are the standard model for computing polynomials (see Section 1.1 for definition). Roughly speaking, given a set of variables X = {x1 , . . . , xn }, an arithmetic circuit uses additions and multiplications to compute a polynomial f in the set of variables X . Given a polynomial f , we are interested in the number of operations needed to compute f . The best lower bound known for the size of arithmetic circuits is the classical Ω(n log n) of Strassen [S73], and of Baur and Strassen [BS83]. Proving better lower bounds is an outstanding open problem. Better lower bounds are not known even for arithmetic circuits of depth 4 (over fields of characteristic different than 2). In this paper, we focus on a restricted class of arithmetic circuits, the class of syntactically
more convenient to work with. Note, for example, that given an arithmetic circuit Φ, deciding whether Φ is syntactically multilinear is straightforward, whereas it is not clear if one can decide whether Φ is semantically multilinear in deterministic polynomial time. We note also that similar distinction between semantic and syntactic definitions occur in other places in computer science (e.g., read k -times branching programs). Multilinear arithmetic circuits were defined by Nisan and Wigderson in [NW96]. The model of syntactically multilinear arithmetic formulas was defined in [R04a]. In [R04a], it is shown that any multilinear arithmetic formula computing the determinant (or the permanent) of an n × n matrix must be of size nΩ(log n) . Prior to our work, no lower bounds (better than the Ω(n log n) lower bound of Strassen, and of Baur and Strassen) for the size of syntactically multilinear arithmetic circuits were known. The techniques of [R04a] for proving super-polynomial lower bounds for the size of multilinear arithmetic formulas fail for circuits. In fact, [R04b] used these techniques to prove that syntactically multilinear arithmetic circuits are super-polynomially more powerful than multilinear arithmetic formulas. More specifically, there exists a polynomial f such that every multilinear arithmetic formula computing f is of size nΩ(log n) , and on the other hand, there exists a polynomial size syntactically multilinear arithmetic circuit computing f . Every multilinear polynomial f can be computed by a syntactically multilinear arithmetic circuit Φ, but Φ might not be the smallest arithmetic circuit computing f . However, computing a multilinear polynomial by an arithmetic circuit that is not syntactically multilinear is usually less intuitive, as cancellations of monomials are needed. A syntactically multilinear arithmetic circuit is semantically multilinear as well. However, it is still not known whether there is an efficient way to transform a semantically multilinear arithmetic circuit to a syntactically multilinear circuit. We note that a semantically multilinear arithmetic formula can be transformed without changing its size to a syntactically multilinear arithmetic formula that computes the same polynomial (see Section 2 in [R04a]). We do not know of any significant example of a semantically multilinear arithmetic circuit that is not syntactically multilinear. Finally, we note that two known classes of arithmetic circuits are contained in the class of syntactically multilinear arithmetic circuits: Pure arithmetic circuits (as defined by Nisan and Wigderson in [NW96], see also [RS05]) are a restricted type of syntactically multilinear arithmetic circuits. Monotone arithmetic circuits computing a multilinear polynomial are also syntactically multilinear (see [NW96]).