毕业设计外文资料翻译_文献英文原文
毕业论文(设计)外文文献翻译及原文

金融体制、融资约束与投资——来自OECD的实证分析R.SemenovDepartment of Economics,University of Nijmegen,Nijmegen(荷兰内梅亨大学,经济学院)这篇论文考查了OECD的11个国家中现金流量对企业投资的影响.我们发现不同国家之间投资对企业内部可获取资金的敏感性具有显著差异,并且银企之间具有明显的紧密关系的国家的敏感性比银企之间具有公平关系的国家的低.同时,我们发现融资约束与整体金融发展指标不存在关系.我们的结论与资本市场信息和激励问题对企业投资具有重要作用这种观点一致,并且紧密的银企关系会减少这些问题从而增加企业获取外部融资的渠道。
一、引言各个国家的企业在显著不同的金融体制下运行。
金融发展水平的差别(例如,相对GDP的信用额度和相对GDP的相应股票市场的资本化程度),在所有者和管理者关系、企业和债权人的模式中,企业控制的市场活动水平可以很好地被记录.在完美资本市场,对于具有正的净现值投资机会的企业将一直获得资金。
然而,经济理论表明市场摩擦,诸如信息不对称和激励问题会使获得外部资本更加昂贵,并且具有盈利投资机会的企业不一定能够获取所需资本.这表明融资要素,例如内部产生资金数量、新债务和权益的可得性,共同决定了企业的投资决策.现今已经有大量考查外部资金可得性对投资决策的影响的实证资料(可参考,例如Fazzari(1998)、 Hoshi(1991)、 Chapman(1996)、Samuel(1998)).大多数研究结果表明金融变量例如现金流量有助于解释企业的投资水平。
这项研究结果解释表明企业投资受限于外部资金的可得性。
很多模型强调运行正常的金融中介和金融市场有助于改善信息不对称和交易成本,减缓不对称问题,从而促使储蓄资金投着长期和高回报的项目,并且提高资源的有效配置(参看Levine(1997)的评论文章)。
因而我们预期用于更加发达的金融体制的国家的企业将更容易获得外部融资.几位学者已经指出建立企业和金融中介机构可进一步缓解金融市场摩擦。
毕设外文文献+翻译1

毕设外文文献+翻译1外文翻译外文原文CHANGING ROLES OF THE CLIENTS、ARCHITECTSAND CONTRACTORS THROUGH BIMAbstract:Purpose –This paper aims to present a general review of the practical implications of building information modelling (BIM) based on literature and case studies. It seeks to address the necessity for applying BIM and re-organising the processes and roles in hospital building projects. This type of project is complex due to complicated functional and technical requirements, decision making involving a large number of stakeholders, and long-term development processes.Design/methodology/approach–Through desk research and referring to the ongoing European research project InPro, the framework for integrated collaboration and the use of BIM are analysed.Findings –One of the main findings is the identification of the main factors for a successful collaboration using BIM, which can be recognised as “POWER”: product information sharing (P),organisational roles synergy (O), work processes coordination (W), environment for teamwork (E), and reference data consolidation (R).Originality/value –This paper contributes to the actual discussion in science and practice on the changing roles and processes that are required to develop and operate sustainable buildings with the support of integrated ICT frameworks and tools. It presents the state-of-the-art of European research projects and some of the first real cases of BIM application inhospital building projects.Keywords:Europe, Hospitals, The Netherlands, Construction works, Response flexibility, Project planningPaper type :General review1. IntroductionHospital building projects, are of key importance, and involve significant investment, and usually take a long-term development period. Hospital building projects are also very complex due to the complicated requirements regarding hygiene, safety, special equipments, and handling of a large amount of data. The building process is very dynamic and comprises iterative phases and intermediate changes. Many actors with shifting agendas, roles and responsibilities are actively involved, such as: the healthcare institutions, national and local governments, project developers, financial institutions, architects, contractors, advisors, facility managers, and equipment manufacturers and suppliers. Such building projects are very much influenced, by the healthcare policy, which changes rapidly in response to the medical, societal and technological developments, and varies greatly between countries (World Health Organization, 2000). In The Netherlands, for example, the way a building project in the healthcare sector is organised is undergoing a major reform due to a fundamental change in the Dutch health policy that was introduced in 2008.The rapidly changing context posts a need for a building with flexibility over its lifecycle. In order to incorporate life-cycle considerations in the building design, construction technique, and facility management strategy, a multidisciplinary collaboration is required. Despite the attempt for establishing integrated collaboration, healthcare building projects still facesserious problems in practice, such as: budget overrun, delay, and sub-optimal quality in terms of flexibility, end-user?s dissatisfaction, and energy inefficiency. It is evident that the lack of communication and coordination between the actors involved in the different phases of a building project is among the most important reasons behind these problems. The communication between different stakeholders becomes critical, as each stakeholder possesses different setof skills. As a result, the processes for extraction, interpretation, and communication of complex design information from drawings and documents are often time-consuming and difficult. Advanced visualisation technologies, like 4D planning have tremendous potential to increase the communication efficiency and interpretation ability of the project team members. However, their use as an effective communication tool is still limited and not fully explored. There are also other barriers in the information transfer and integration, for instance: many existing ICT systems do not support the openness of the data and structure that is prerequisite for an effective collaboration between different building actors or disciplines.Building information modelling (BIM) offers an integrated solution to the previously mentioned problems. Therefore, BIM is increasingly used as an ICT support in complex building projects. An effective multidisciplinary collaboration supported by an optimal use of BIM require changing roles of the clients, architects, and contractors; new contractual relationships; and re-organised collaborative processes. Unfortunately, there are still gaps in the practical knowledge on how to manage the building actors to collaborate effectively in their changing roles, and todevelop and utilise BIM as an optimal ICT support of the collaboration.This paper presents a general review of the practical implications of building information modelling (BIM) based on literature review and case studies. In the next sections, based on literature and recent findings from European research project InPro, the framework for integrated collaboration and the use of BIM are analysed. Subsequently, through the observation of two ongoing pilot projects in The Netherlands, the changing roles of clients, architects, and contractors through BIM application are investigated. In conclusion, the critical success factors as well as the main barriers of a successful integrated collaboration using BIM are identified.2. Changing roles through integrated collaboration and life-cycle design approachesA hospital building project involves various actors, roles, and knowledge domains. In The Netherlands, the changing roles of clients, architects, and contractors in hospital building projects are inevitable due the new healthcare policy. Previously under the Healthcare Institutions Act (WTZi), healthcare institutions were required to obtain both a license and a building permit for new construction projects and major renovations. The permit was issued by the Dutch Ministry of Health. The healthcare institutions were then eligible to receive financial support from the government. Since 2008, new legislation on the management of hospital building projects and real estate has come into force. In this new legislation, a permit for hospital building project under the WTZi is no longer obligatory, nor obtainable (Dutch Ministry of Health, Welfare and Sport, 2008). This change allows more freedom from the state-directed policy, and respectively,allocates more responsibilities to the healthcare organisations to deal with the financing and management of their real estate. The new policy implies that the healthcare institutions are fully responsible to man age and finance their building projects and real estate. The government?s support for the costs of healthcare facilities will no longer be given separately, but will be included in the fee for healthcare services. This means that healthcare institutions must earn back their investment on real estate through their services. This new policy intends to stimulate sustainable innovations in the design, procurement and management of healthcare buildings, which will contribute to effective and efficient primary healthcare services.The new strategy for building projects and real estate management endorses an integrated collaboration approach. In order to assure the sustainability during construction, use, and maintenance, the end-users, facility managers, contractors and specialist contractors need to be involved in the planning and design processes. The implications of the new strategy are reflected in the changing roles of the building actors and in the new procurement method.In the traditional procurement method, the design, and its details, are developed by the architect, and design engineers. Then, the client (the healthcare institution) sends an application to the Ministry of Healthto obtain an approval on the building permit and the financial support from the government. Following this, a contractor is selected through a tender process that emphasises the search for the lowest-price bidder. During the construction period, changes often take place due to constructability problems of the design and new requirements from the client.Because of the high level of technical complexity, and moreover, decision-making complexities, the whole process from initiation until delivery of a hospital building project can take up to ten years time. After the delivery, the healthcare institution is fully in charge of the operation of the facilities. Redesigns and changes also take place in the use phase to cope with new functions and developments in the medical world.The integrated procurement pictures a new contractual relationship between the parties involved in a building project. Instead of a relationship between the client and architect for design, and the client and contractor for construction, in an integrated procurement the client only holds a contractual relationship with the main party that is responsible for both design and construction. The traditional borders between tasks and occupational groups become blurred since architects, consulting firms, contractors, subcontractors, and suppliers all stand on the supply side in the building process while the client on the demand side. Such configuration puts the architect, engineer and contractor in a very different position that influences not only their roles, but also their responsibilities, tasks and communication with the client, the users, the team and other stakeholders.The transition from traditional to integrated procurement method requires a shift of mindset of the parties on both the demand and supply sides. It is essential for the client and contractor to have a fair and open collaboration in which both can optimally use their competencies. The effectiveness of integrated collaboration is also determined by the client?s capacity and strategy to organize innovative tendering procedures.A new challenge emerges in case of positioning an architect in a partnership with the contractor instead of with the client. In case of the architect enters a partnership with the contractor, an important issues is how to ensure the realisation of the architectural values as well as innovative engineering through an efficient construction process. In another case, the architect can stand at the client?s side in a strategic advisory role instead of being the designer. In this case, the architect?s responsibility is translating client?s requirements and wishes into the architectural values to be included in the design specification, and evaluating the contractor?s proposal against this. In any of this new role, the architect holds the responsibilities as stakeholder interest facilitator, custodian of customer value and custodian of design models.The transition from traditional to integrated procurement method also brings consequences in the payment schemes. In the traditional building process, the honorarium for the architect is usually based on a percentage of the project costs; this may simply mean that the more expensive the building is, the higher the honorarium will be. The engineer receives the honorarium based on the complexity of the design and the intensity of the assignment. A highly complex building, which takes a number of redesigns, is usually favourable for the engineers in terms of honorarium. A traditional contractor usually receives the commission based on the tender to construct the building at the lowest price by meeting the minimum specifications given by the client. Extra work due to modifications is charged separately to the client. After the delivery, the contractor is no longer responsible for the long-term use of the building. In the traditional procurement method, all risks are placed with theclient.In integrated procurement method, the payment is based on the achieved building performance; thus, the payment is non-adversarial. Since the architect, engineer and contractor have a wider responsibility on the quality of the design and the building, the payment is linked to a measurement system of the functional and technical performance of the building over a certain period of time. The honorarium becomes an incentive to achieve the optimal quality. If the building actors succeed to deliver a higher added-value thatexceed the minimum client?s requirements, they will receive a bonus in accordance to the client?s extra gain. The level of transparency is also improved. Open book accounting is an excellent instrument provided that the stakeholders agree on the information to be shared and to its level of detail (InPro, 2009).Next to the adoption of integrated procurement method, the new real estate strategy for hospital building projects addresses an innovative product development and life-cycle design approaches. A sustainable business case for the investment and exploitation of hospital buildings relies on dynamic life-cycle management that includes considerations and analysis of the market development over time next to the building life-cycle costs (investment/initial cost, operational cost, and logistic cost). Compared to the conventional life-cycle costing method, the dynamic life-cycle management encompasses a shift from focusing only on minimizing the costs to focusing on maximizing the total benefit that can be gained. One of the determining factors for a successful implementation of dynamic life-cycle management is the sustainable design of the building and building components, which means that the design carriessufficient flexibility to accommodate possible changes in the long term (Prins, 1992).Designing based on the principles of life-cycle management affects the role of the architect, as he needs to be well informed about the usage scenarios and related financial arrangements, the changing social and physical environments, and new technologies. Design needs to integrate people activities and business strategies over time. In this context, the architect is required to align the design strategies with the organisational, local and global policies on finance, business operations, health and safety, environment, etc.The combination of process and product innovation, and the changing roles of the building actors can be accommodated by integrated project delivery or IPD (AIA California Council, 2007). IPD is an approach that integrates people, systems, business structures and practices into a process that collaboratively harnesses the talents and insights of all participants to reduce waste and optimize efficiency through all phases of design, fabrication and construction. IPD principles can be applied to a variety of contractual arrangements. IPD teams will usually include members well beyond the basic triad of client, architect, and contractor. At a minimum, though, an Integrated Project should include a tight collaboration between the client, the architect, and the main contractor ultimately responsible for construction of the project, from the early design until the project handover. The key to a successful IPD is assembling a team that is committed to collaborative processes and is capable of working together effectively. IPD is built on collaboration. As a result, it can only be successful if the participants share and apply common values and goals.3. Changing roles through BIM applicationBuilding information model (BIM) comprises ICT frameworks and tools that can support the integrated collaboration based on life-cycle design approach. BIM is a digital representation of physical and functional characteristics of a facility. As such it serves as a shared knowledge resource for information about a facility forming a reliable basis for decisions during its lifecycle from inception onward (National Institute of Building Sciences NIBS, 2007). BIM facilitates time and place independent collaborative working. A basic premise of BIM is collaboration by different stakeholders at different phases of the life cycle of a facility to insert, extract, update or modify information in the BIM to support and reflect the roles of that stakeholder. BIM in its ultimate form, as a shared digital representation founded on open standards for interoperability, can become a virtual information model to be handed from the design team to the contractor and subcontractors and then to the client.BIM is not the same as the earlier known computer aided design (CAD). BIM goes further than an application to generate digital (2D or 3D) drawings. BIM is an integrated model in which all process and product information is combined, stored, elaborated, and interactively distributed to all relevant building actors. As a central model for all involved actors throughout the project lifecycle, BIM develops andevolves as the project progresses. Using BIM, the proposed design and engineering solutions can be measured against the client?s requirements and expected building performance. The functionalities of BIM to support the design process extend to multidimensional (nD), including: three-dimensional visualisation and detailing, clash detection, material schedule, planning, costestimate, production and logistic information, and as-built documents. During the construction process, BIM can support the communication between the building site, the factory and the design office– which is crucial for an effective and efficient prefabrication and assembly processes as well as to prevent or solve problems related to unforeseen errors or modifications. When the building is in use, BIM can be used in combination with the intelligent building systems to provide and maintain up-to-date information of the building performance, including the life-cycle cost.To unleash the full potential of more efficient information exchange in the AEC/FM industry in collaborative working using BIM, both high quality open international standards and high quality implementations of these standards must be in place. The IFC open standard is generally agreed to be of high quality and is widely implemented in software. Unfortunately, the certification process allows poor quality implementations to be certified and essentially renders the certified software useless for any practical usage with IFC. IFC compliant BIM is actually used less than manual drafting for architects and contractors, and show about the same usage for engineers. A recent survey shows that CAD (as a closed-system) is still the major form of technique used in design work (over 60 per cent) while BIM is used in around 20 percent of projects for architects and in around 10 per cent of projects for engineers and contractors.The application of BIM to support an optimal cross-disciplinary and cross-phase collaboration opens a new dimension in the roles and relationships between the building actors. Several most relevant issues are: the new role of a model manager; the agreement on the access right and IntellectualProperty Right (IPR); the liability and payment arrangement according to the type of contract and in relation to the integrated procurement; and the use of open international standards.Collaborative working using BIM demands a new expert role of a model manager who possesses ICT as well as construction process know-how (InPro, 2009). The model manager deals with the system as well as with the actors. He provides and maintains technological solutions required for BIM functionalities, manages the information flow, and improves the ICT skills of the stakeholders. The model manager does not take decisions on design and engineering solutions, nor the organisational processes, but his roles in the chain of decision making are focused on:the development of BIM, the definition of the structure and detail level of the model, and the deployment of relevant BIM tools, such as for models checking, merging, and clash detections;the contribution to collaboration methods, especially decision making and communication protocols, task planning, and risk management;and the management of information, in terms of data flow and storage, identification of communication errors, and decision or process (re-)tracking.Regarding the legal and organisational issues, one of the actual questions is: “In what way does the intellectual property right (IPR) in collaborative working using BIM differ from the IPR in a traditional teamwork?”. In terms of combine d work, the IPR of each element is at tached to its creator. Although it seems to be a fully integrated design, BIM actually resulted from a combination of works/elements; for instance: the outline of the building design, is created by the architect, the design for theelectrical system, is created by the electrical contractor, etc. Thus, in case of BIM as a combined work, the IPR is similar to traditional teamwork. Working with BIM with authorship registration functionalities may actually make it easier to keep track of the IPR.How does collaborative working, using BIM, effect the contractual relationship? On the one hand,collaborative working using BIM does not necessarily change the liability position in the contract nor does it obligate an alliance contract. The General Principles of BIM A ddendum confirms: …This does not effectuate or require a restructuring of contractual relationships or shifting of risks between or among the Project Participants other than as specifically required per the Protocol Addendum and its Attachments? (ConsensusDOCS, 2008). On the other hand, changes in terms of payment schemes can be anticipated. Collaborative processes using BIM will lead to the shifting of activities from to the early design phase. Much, if not all, activities in the detailed engineering and specification phase will be done in the earlier phases. It means that significant payment for the engineering phase, which may count up to 40 per cent of the design cost, can no longer be expected. As engineering work is done concurrently with the design, a new proportion of the payment in the early design phase is necessary.4. Review of ongoing hospital building projects using BIMIn The Netherlands, the changing roles in hospital building projects are part of the strategy, which aims at achieving a sustainable real estate in response to the changing healthcare policy. Referring to literature and previous research, the main factors that influence the success of the changing roles can be concluded as: the implementation of an integrated procurementmethod and a life-cycle design approach for a sustainable collaborative process; the agreement on the BIM structure and the intellectual rights; and the integration of the role of a model manager. The preceding sections have discussed the conceptual thinking on how to deal with these factors effectively. This current section observes two actual projects and compares the actual practice with the conceptual view respectively.The main issues, which are observed in the case studies, are: the selected procurement method and the roles of the involved parties within this method;the implementation of the life-cycle design approach;the type, structure, and functionalities of BIM used in the project;the openness in data sharing and transfer of the model, and the intended use of BIM in the future; and the roles and tasks of the model manager.The pilot experience of hospital building projects using BIM in the Netherlands can be observed at University Medical Centre St Radboud (further referred as UMC) and Maxima Medical Centre (further referred as MMC). At UMC, the new building project for the Faculty of Dentistry in the city of Nijmegen has been dedicated as a BIM pilot project. At MMC, BIM is used in designing new buildings for Medical Simulation and Mother-and-Child Centre in the city of Veldhoven.The first case is a project at the University Medical Centre (UMC) St Radboud. UMC is more than just a hospital. UMC combines medical services, education and research. More than 8500 staff and 3000 students work at UMC. As a part of the innovative real estate strategy, UMC has considered to use BIM for its building projects. The new development of the Faculty ofDentistry and the surrounding buildings on the Kapittelweg in Nijmegen has been chosen as a pilot project to gather practical knowledge and experience on collaborative processes with BIM support.The main ambition to be achieved through the use of BIM in the building projects at UMC can be summarised as follows: using 3D visualisation to enhance the coordination and communication among the building actors, and the user participation in design;integrating the architectural design with structural analysis, energy analysis, cost estimation, and planning;interactively evaluating the design solutions against the programme of requirements and specifications;reducing redesign/remake costs through clash detection during the design process; andoptimising the management of the facility through the registration of medical installations andequipments, fixed and flexible furniture, product and output specifications, and operational data.The second case is a project at the Maxima Medical Centre (MMC). MMC is a large hospital resulted from a merger between the Diaconessenhuis in Eindhoven and St Joseph Hospital in Veldhoven. Annually the 3,400 staff of MMC provides medical services to more than 450,000 visitors and patients. A large-scaled extension project of the hospital in Veldhoven is a part of its real estate strategy. A medical simulation centre and a women-and-children medical centre are among the most important new facilities within this extension project. The design has been developed using 3D modelling with several functionalities of BIM.The findings from both cases and the analysis are as follows.Both UMC and MMC opted for a traditional procurement method in which the client directly contracted an architect, a structural engineer, and a mechanical, electrical and plumbing (MEP) consultant in the design team. Once the design and detailed specifications are finished, a tender procedure will follow to select a contractor. Despite the choice for this traditional method, many attempts have been made for a closer and more effective multidisciplinary collaboration. UMC dedicated a relatively long preparation phase with the architect, structural engineer and MEP consultant before the design commenced. This preparation phase was aimed at creating a common vision on the optimal way for collaboration using BIM as an ICT support. Some results of this preparation phase are: a document that defines the common ambition for the project and the collaborative working process and a semi-formal agreement that states the commitment of the building actors for collaboration. Other than UMC, MMC selected an architecture firm with an in-house engineering department. Thus, the collaboration between the architect and structural engineer can take place within the same firm using the same software application.Regarding the life-cycle design approach, the main attention is given on life-cycle costs, maintenance needs, and facility management. Using BIM, both hospitals intend to get a much better insight in these aspects over the life-cycle period. The life-cycle sustainability criteria are included in the assignments for the design teams. Multidisciplinary designers and engineers are asked to collaborate more closely and to interact with the end-users to address life-cycle requirements. However, ensuring the building actors to engage in an integrated collaboration to generate sustainable design solutions that meet the life-cycle。
毕业设计外文文献翻译(原文+译文)

Environmental problems caused by Istanbul subway excavation and suggestionsfor remediation伊斯坦布尔地铁开挖引起的环境问题及补救建议Ibrahim Ocak Abstract:Many environmental problems caused by subway excavations have inevitably become an important point in city life. These problems can be categorized as transporting and stocking of excavated material, traffic jams, noise, vibrations, piles of dust mud and lack of supplies. Although these problems cause many difficulties,the most pressing for a big city like Istanbul is excava tion,since other listed difficulties result from it. Moreover, these problems are environmentally and regionally restricted to the period over which construction projects are underway and disappear when construction is finished. Currently, in Istanbul, there are nine subway construction projects in operation, covering approximately 73 km in length; over 200 km to be constructed in the near future. The amount of material excavated from ongoing construction projects covers approximately 12 million m3. In this study, problems—primarily, the problem with excavation waste(EW)—caused by subway excavation are analyzed and suggestions for remediation are offered.摘要:许多地铁开挖引起的环境问题不可避免地成为城市生活的重要部分。
毕业设计(论文)外文参考资料及译文

英文原文:Java is a simple, object-oriented, distributed, interpreted, robust security, structure-neutral, portable, high performance, multithreaded dynamic language. The main advantage of Java language, Java applications across hardware platforms and operating systems for transplant - this is because the JVM is installed on each platform can understand the same byte code. Java language and platform scalability is very strong. At the low end, Java language is the first open standards technology support enterprise one, support the use of XML and Web service can not stride business lines to share information and applications Cheng Xu.There are three versions of Java platform, which makes software developers, service providers and equipment manufacturers can target specific market development:1. Java SE form applications. Java SE includes support for Java Web services development classes, and for the Java Platform, Enterprise Edition (Java EE) to provide a basis. Most Java developers use Java SE 5, also known as Java 5.0 or "Tiger".2. Java EE formerly known as J2EE. Enterprise Edition to help develop and deploy portable, robust, scalable and secure server-side Java applications. Java SE Java EE is built on the foundation, which provides Web services, component model, management and communication API, can be used to achieve enterprise-class service-oriented architecture and Web 2.0 applications.3. Java ME formerly known as J2ME. Java ME devices in mobile and embedded applications running on a robust and flexible environment. Java ME includes flexible user interfaces, robust security model, and many built-in network protocols and networking that can be dynamically downloaded and extensive support for offline applications. Java ME-based application specification only write once and can be used in many devices and can use the native features of each device.Java language is simple. Java language syntax and the C language and C ++ language is very close, Java discarded the C++, rarely used, hard to understand the characteristics, such as operator overloading, multiple inheritance, the mandatory automatic type conversion. Java language does not use pointers, and provides automated waste collection. Java is an object-oriented language. Java language provides classes, interfaces and inheritance of the original language, for simplicity, only supports single inheritance between classes, but support multiple inheritance between interfaces and support classes and interfaces to achieve between the mechanism (keyword implements) . Java language fully supports dynamic binding, and C ++ language used only for dynamic binding of virtual functions. In short, Java language is a pure object-oriented programming language. Java language is distributed. Java language support for Internet application development, Java's RMI (remote method activation) mechanism is also an important means of developing distributed applications. Java language is robust. Java's strong type system, exception handling, automated waste collection is an important guarantee robust Java programs. Java language is safe. Java is often used in network environment, this, Java provides a security mechanism to prevent malicious code attacks.Java language is portable. This portability comes from the architecture neutrality. Java system itself is highly portable. Java language is multi-threaded. In the Java language, the thread is a special object, it must Thread class or the son (Sun) class to create. Java language support simultaneous execution of multiple threads, and provide synchronization mechanisms between threads (keyword synchronized).Java language features make Java an excellent application of unparalleled robustness and reliability, which also reduced application maintenance costs. Java on the full support of object technology and Java Platform API embedded applications to reduce development time and reduce costs. Java's compile once, run everywhere feature can make it anywhere available to provide an open architecture and multi-platform, low-cost way of transmitting information between. Hibernate Hibernate is a lightweight JDBC object package. It is an independent object persistence framework, and the App Server, and EJB is no necessary link. Hibernate can use JDBC can be used in any occasion, such as Java application, database access code, DAO interface implementation class, or even access the database inside a BMP code. In this sense, Hibernate, and EB is not a category of things that did not exist either-or relationship.Hibernate and JDBC is a closely related framework, the Hibernate and JDBC driver compatibility, and databases have some relationship, but the Java program and use it, and the App Server does not have any relationship, there was no compatibility issues. 1614Hibernate provides two Cache, first-level cache is a Session-level cache, which cache belongs to the scope of services. This level of cache by the hibernate managed without the need for intervention under normal circumstances; second-level cache is SessionFactory-level cache, it belongs to the process of range or scope of the cache cluster. This level of cache can be configured and changed, and can be dynamically loaded and unloaded. Hibernate query results also provide a query cache, it depends on the second level cache.When an application called Session's save (), update (), saveOrUpdate (), get () or load (), and the query interface call list (), iterate () or filter () method, if the Session cache does not exist a corresponding object, Hibernate will put the object to the first level cache. When cleaning the cache, Hibernate objects according to the state of the cache changes to synchronize update the database. Session for the application provides two methods of managing the cache: evict (Object obj): removed from the cache parameters of the specified persistent object. clear (): Empty the cache of all persistent objects.Hibernate second-level cache strategy general process is as follows:1) The condition when a query is always issued a select * from table_name where .... (Select all fields) such as SQL statement to query the database, an access to all of the data object.2) all the data objects to be placed under the ID to the second level cache.3) When the Hibernate object-based ID to access the data, the first check from the Session a cache; finding out, if the configuration of the secondary cache, then the secondary cache from the investigation; finding out, and then query the database, the results in accordance with the ID into the cache.4) remove, update and increase the time data, while updating the cache. Hibernate second against the conditions of the Query Cache.Hibernate object-relational mapping for the delay and non-delay object initialization. Non-lazy when reading an object and the object will be all read out together with other objects. This sometimes results in hundreds (if not thousands of words) select statement when reading the object implementation. This problem sometimes occurs when using the two-way relationship, often leading to the databases to be read during the initialization phase out. Of course, you can take the trouble to examine each object and other objects of Guanxi, and to the most expensive of the Shan Chu, but in the last, we may therefore lose Le ORM tool this Xiangzai obtained Bian Li.A cache and secondary cache of comparison: the first level cache second level cache data stored in the form of interrelated persistent objects the object of bulk data cache range of the scope of services, each transaction has a separate first-level cache process range or scope of the cluster, the cache is the same process or cluster to share on all matters within the concurrent access policies because each transaction has a separate first-level cache, concurrency problem does not occur without the need to provide concurrent access policy will be a number of matters simultaneous access to the same second-level cache data, it is necessary to provide appropriate concurrent access policies, to ensure that a particular transaction isolation level data expiration policies did not provide data expiration policies. Object in a cache will never expire, unless the application explicitly clear the cache or clear a specific object must provide data expiration policies, such as memory cache based on the maximum number of objects, allowing objects in the cache of the most a long time, and allowing the object in the cache the longest idle time of physical memory and hard disk memory storage medium. First of all bulk data objects stored in the memory-based cache, when the number of objects in memory to data expiration policy specified limit, the remaining objects will be written on the hard disk cache. Caching software implementation of the Hibernate Session is included in the realization of the cache provided by third parties, Hibernate provides only a cache adapter (CacheProvider). Used to plug into a particular cache in Hibernate. Way cache enabled applications by as long as the Session interface implementation save, update, delete, data loading and query the database operations, Hibernate will enable first-level cache, the data in the database in the form of an object copied to the cache For batch updates and bulk delete operations, if you do not want to enable first-level cache, you can bypass the Hibernate API, JDBC API directly to perform that operation. Users can type in a single class or a single set of second-level cache size on the configuration. If the instance of the class are frequently read but rarely modified, you can consider using a second-level cache. Only for a class or set of second-level cache is configured, Hibernate will run when an instance of it to the second-level cache. User management means the first level cache of physical media for the memory cache, because the memory capacity is limited, must pass the appropriate search strategies and retrieval methods to limit the number of objects loaded. Session of the evit () method can explicitly clear the cache a specific object, but this method is not recommended. Second-level cache memory andthe physical media can be a hard disk, so the second-level cache can store large amounts of data, data expiration policy maxElementsInMemory property values can control the number of objects in memory. Second-level cache management mainly includes two aspects: Select to use the second-level cache of persistent classes, set the appropriate concurrency strategy: Select the cache adapter, set the appropriate data expiration policies.One obvious solution is to use Hibernate's lazy loading mechanism provided. This initialization strategy is only invoked in an object-to-many or many to many relationship between its relationship only when read out of the object. This process is transparent to the developer, and only had a few requests for database operations, it will be more obvious performance have open. This will be by using the DAO pattern abstracts the persistence time of a major problem. Persistence mechanisms in order to completely abstract out all of the database logic, including open or closed session, can not appear in the application layer. The most common is the realization of the simple interface of some DAO implementation class to encapsulate the database logic completely. A fast but clumsy solution is to give up DAO mode, the database connection logic to add the application layer. This may be an effective small applications, but in large systems, this is a serious design flaw, preventing the system scalability.Struts2Struts2 is actually not a stranger to the Web frameworks, Struts2 is Webwork design ideas as the core, absorb Struts1 advantages, so that the Struts2 is the product of the integration Struts1 and Webwork.MVC Description: Struts2 WebWork is compatible with the MVC framework Struts1 and since, that the MVC framework on the MVC framework will have to make a brief, limited to a brief, if want to learn more about MVC can view the related knowledge document, or to find a Struts1 books, I believe the above is not rare on the length of MVC. Closer to home, in fact, Java the present situation of these frameworks, its ultimate goal is to contact coupling, whether Spring, Hibernate or the MVC framework, are designed to increase contact with coupling reuse. MVC contact with the coupling between View and Model. MVC consists of three basic parts: Model, View and Controller, these three parts work together to minimize the coupling to increase the scalability of the program and maintainability. Various parts of the implementation technology can be summarized as follows:1) Model: JavaBean, EJB's EntityBean2) View: JSP, Struts in TagLib3) Controller: Struts the ActionServlet, ActionTo sum up the advantages of MVC mainly about aspects:1) corresponds to multiple views can be a model. By MVC design pattern, a model that corresponds to multiple views, you can copy the code and the code to reduce the maintenance amount, if model changes, but also easy to maintain2) model the data returned and display logic separate. Model data can be applied to any display technology, for example, use the JSP page, Velocity templates, or directly from Excel documents, etc.3) The application is separated into three layers, reducing the coupling between the layers, providing application scalability4) The concept of layers is also very effective, because it put the different models and different views together, to complete a different request. Therefore, the control layer can be said to include the concept of user requests permission5) MVC more software engineering management. Perform their duties in different layers, each layer has the same characteristics of the components is beneficial tool by engineering and production management of program codeStruts2 Introduction: Struts2 Struts1 development appears to come from, but in fact Struts1 Struts2 and design ideas in the framework of the above is very different, Struts2 WebWork's design is based on the core, why not follow the Struts1 Struts2 design ideas After all, Struts1 in the current enterprise applications market is still very big in the, Struts1 some shortcomings:1) support the performance of a single layer2) coupled with the Servlet API serious, this could be the Execute method from the Action Statement which you can see them3) The code depends Struts1 API, there are invasive, this can be written when the Action class and look out FormBean, Action Struts in Action class must implement The reason for Struts2 WebWork's design for the core point is the recent upward trend of WebWork and play WebWork not Struts1 above those shortcomings, more MVC design ideas, and more conducive to reuse the code. Based on the above description can be read out, Struts2 architecture and architecture Struts1 very different, Struts1 is to use the ActionServlet as its central processor, Struts2 is using an interceptor (FilterDispatcher) as its central processor, so One benefit is to make Action class and Servlet API was isolated.Struts2 simple process flow is as follows:1) browser sends a request2) the processor to find the corresponding file under struts.xml the Action class to process the request3) WebWork interceptor chain applications automatically request common functions, such as: WorkFlow, Validation functions4) If Struts.xml Method configuration file parameters, then call the corresponding Action Method parameters in the Method class method, or call the Execute method to deal with common user request5) Action class method returns the results of the corresponding response to the browserStruts2 and Struts1 contrast:1) Action class impleme achieve the time to achieve any classes and interfaces, while providing a ActionSupport class Struts2, however, not required.2) Struts1 the Action class is the singleton pattern, must be designed into the thread-safe, Struts2 was generated for each request for an instance3) Struts1 the Action class dependence and the Servlet API, execute the method from its signature can be seen, execute method has two parameters Servlet HttpServletRequest and HttpServletResponse, Struts2 is not dependent on the ServletAPI4) Struts1 depends on the Servlet API the Web elements, therefore, of Action Struts1 when testing is difficult, it needs with other testing tools, Struts2 in Action can be as testing a number of other classes as Service Model layer test5) Struts1 of Action and the View through the ActionForm or its sub-class of data transmission, although there LazyValidationForm this ActionForm appearance, but still can not like the other levels as a simple POJO data transfer, and Struts2 would like expect change becomes a reality6) Struts1 binding of the JSTL, the preparation of convenience for the page, Struts2 integrates ONGL, you can use JSTL, Therefore, Struts2 is more powerful expression language underCompared with Struts2 WebWork: Struts2 actually WebWork2.3, however, Struts2 WebWork, or with a little difference:1) Struts2 IOC no longer support the built-in containers, use Spring's IOC container2) Struts2 Ajax for Webwork features some of the label to use Dojo to be replacedServletServlet is a server-side Java application, platform and protocol independent features that can generate dynamic Web pages. Customer requests to play it (Web browser or other HTTP client) and server response (HTTP server, database or application) of the middle layer. Servlet Web server is located inside the server-side Java applications started from the command line with the traditional application of different Java, Servlet loaded by the Web server, the Web server must include the Java Virtual Machine to support Servlet.HTTP Servlet using a HTML form to send and receive data. To create an HTTP Servlet, need to extend the HttpServlet class, the class is a special way to handle HTML forms GenericServlet a subclass. HTML form is <FORM> and </ FORM> tag definition. Form typically includes input fields (such as text input fields, check boxes, radio buttons and selection lists) and a button for submitting data. When submitting information, they also specify which server should implement the Servlet (or other program). HttpServlet class contains the init (), destroy (), service () and other methods. Where init () and destroy () method is inherited.init () method: In the Servlet life period, only run once init () method. It is executed when the server load Servlet. You can configure the server to start the server or the client's first visit to Servlet fashion into the Servlet. No matter how many clients to access Servlet, will not repeat the init (). The default init () method is usually to meet the requirements, but can also use custom init () method to overwrite it, typically the management server-side resources. For example, you may write a custom init () to be used only once a load GIF images, GIF images and improve the Servlet returns with the performance of multiple clients request. Another example is to initialize the database connection. The default init () method sets the Servlet initialization parameters, and use it's ServletConfig object parameter to start the configuration, all covered by init () method of the Servlet should call super.init () to ensure that stillperform these tasks. In the call to service () method before, make sure you have completed the init () method.service () method: service () method is the core of Servlet. Whenever a client requests a HttpServlet object, the object of the service () method must be called, and passed to this method a "request" (ServletRequest) objects and a "response" (ServletResponse) object as a parameter. Already exists in the HttpServlet service () method. The default service function is invoked with the HTTP request method to do the corresponding functions. For example, if the HTTP request method is GET, the default on the call to doGet (). Servlet Servlet support should do HTTP method override function. Because HttpServlet.service () method checks whether the request method calls the appropriate treatment, unnecessary coverage service () method. Just do cover the corresponding method on it.Servlet response to the following types: an output stream, the browser based on its content type (such as text / HTML) to explain; an HTTP error response, redirect to another URL, servlet, JSP.doGet () method: When a client through the HTML form to send a HTTP GET request or when a direct request for a URL, doGet () method is called. Parameters associated with the GET request to the URL of the back, and send together with this request. When the server does not modify the data, you should use doGet () method. doPost () method: When a client through the HTML form to send a HTTP POST request, doPost () method is called. Parameters associated with the POST request as a separate HTTP request from the browser to the server. When the need to modify the server-side data, you should use the doPost () method.destroy () method: destroy () method is only executed once, that is, stop and uninstall the server to execute the method of Servlet. Typically, the Servlet as part of the process server to shut down. The default destroy () method is usually to meet the requirements, but can also cover it, and typically manage server-side resources. For example, if the Servlet will be accumulated in the run-time statistics, you can write a destroy () method is used in Servlet will not load the statistics saved in the file. Another example is to close the database connection.When the server uninstall Servlet, it will in all service () method call is completed, or at a specified time interval after the call to destroy () method. Running a Servlet service () method may have other threads, so make sure the call destroy () method, the thread has terminated or completed.GetServletConfig () method: GetServletConfig () method returns a ServletConfig object, which used to return the initialization parameters and ServletContext. ServletContext interface provides information about servlet environment. GetServletInfo () method: GetServletInfo () method is an alternative method, which provides information on the servlet, such as author, version, copyright.When the server calls sevlet of Service (), doGet () and doPost () of these three methods are needed "request" and "response" object as a parameter. "Request" object to provide the requested information, and the "response" object to provide a response message will be returned to the browser as a communications channel.javax.servlet packages in the relevant classes for the ServletResponse andServletRequest, while the javax.servlet.http package of related classes for the HttpServletRequest and HttpServletResponse. Servlet communication with the server through these objects and ultimately communicate with the client. Servlet through call "request" object approach informed the client environment, server environment, information and all information provided by the client. Servlet can call the "response" object methods to send response, the response is ready to send back to clientJSPJavaServerPages (JSP) technology provides a simple and fast way to create a display content dynamically generated Web pages. Leading from the industry, Sun has developed technology related to JSP specification that defines how the server and the interaction between the JSP page, the page also describes the format and syntax.JSP pages use XML tags and scriptlets (a way to use script code written in Java), encapsulates the logic of generating page content. It labels in various formats (HTML or XML) to respond directly passed back to the page. In this way, JSP pages to achieve a logical page design and display their separation.JSP technology is part of the Java family of technologies. JSP pages are compiled into a servlet, and may call JavaBeans components (beans) or EnterpriseJavaBeans components (enterprise beans), so that server-side processing. Therefore, JSP technology in building scalable web-based applications play an important role.JSP page is not confined to any particular platform or web server. JSP specification in the industry with a wide range of adaptability.JSP technology is the result of collaboration with industry, its design is an open, industry standards, and support the vast majority of servers, browsers and related tools. The use of reusable components and tags replaced on the page itself relies heavily on scripting languages, JSP technology has greatly accelerated the pace of development. Support the realization of all the JSP to Java programming language-based scripting language, it has inherent adaptability to support complex operations.JqueryjQuery is the second prototype followed by a good Javascrīpt framework. Its purpose is: to write less code, do more.It is lightweight js library (compressed only 21k), which is less than the other js library which, it is compatible CSS3, is also compatible with all browsers (IE 6.0 +, FF 1.5 +, Safari 2.0 +, Opera 9.0 +).jQuery is a fast, simple javaScript library, allowing users to more easily dealwith HTML documents, events, to achieve animation effects, and provide easy AJAX for interactive web site.jQuery also has a larger advantage is that it is all documented, and various applications are very detailed, as well as many mature plug-ins available.jQuery's html page to allow users to maintain separate code and html content, that is, no need to insert in the html inside a pile of js to call the command, and you can just define id.jQuery is the second prototype followed by a good Javascrīpt framework. On theprototype I use small, simple and understood. However, after using the jquery immediately attracted by her elegance. Some people use such a metaphor to compare the prototype and jquery: prototype like Java, and jquery like a ruby. In fact I prefer java (less contact with Ruby Bale), but a simple jquery does have considerable practical appeal ah! I put the project in the framework jquery as its the only class package. Use the meantime there is also a little bit of experience, in fact, these ideas, in the jquery documentation above may also be speaking, but still it down to stop notes.译文:Java是一种简单的,面向对象的,分布式的,解释型的,健壮安全的,结构中立的,可移植的,性能优异、多线程的动态语言。
毕业设计外文文献翻译范文

毕业设计外文文献翻译专业学生姓名班级学号指导教师优集学院外文资料名称:Knowledge-Based Engineeri--ng Design Methodology外文资料出处:Int.J.Engng Ed.Vol.16.No.1附件: 1.外文资料翻译译文2.外文原文基于知识工程(KBE)设计方法D. E. CALKINS1.背景复杂系统的发展需要很多工程和管理方面的知识、决策,它要满足很多竞争性的要求。
设计被认为是决定产品最终形态、成本、可靠性、市场接受程度的首要因素。
高级别的工程设计和分析过程(概念设计阶段)特别重要,因为大多数的生命周期成本和整体系统的质量都在这个阶段。
产品成本的压缩最可能发生在产品设计的最初阶段。
整个生命周期阶段大约百分之七十的成本花费在概念设计阶段结束时,缩短设计周期的关键是缩短概念设计阶段,这样同时也减少了工程的重新设计工作量。
工程权衡过程中采用良好的估计和非正式的启发进行概念设计。
传统CAD工具对概念设计阶段的支持非常有限。
有必要,进行涉及多个学科的交流合作来快速进行设计分析(包括性能,成本,可靠性等)。
最后,必须能够管理大量的特定领域的知识。
解决方案是在概念设计阶段包含进更过资源,通过消除重新设计来缩短整个产品的时间。
所有这些因素都主张采取综合设计工具和环境,以在早期的综合设计阶段提供帮助。
这种集成设计工具能够使由不同学科的工程师、设计者在面对复杂的需求和约束时能够对设计意图达成共识。
那个设计工具可以让设计团队研究在更高级别上的更多配置细节。
问题就是架构一个设计工具,以满足所有这些要求。
2.虚拟(数字)原型模型现在需要是一种代表产品设计为得到一将允许一产品的早发展和评价的真实事实上原型的过程的方式。
虚拟样机将取代传统的物理样机,并允许设计工程师,研究“假设”的情况,同时反复更新他们的设计。
真正的虚拟原型,不仅代表形状和形式,即几何形状,它也代表如重量,材料,性能和制造工艺的非几何属性。
毕业设计(论文)外文翻译【范本模板】

华南理工大学广州学院本科生毕业设计(论文)翻译英文原文名Review of Vibration Analysis Methods for Gearbox Diagnostics and Prognostics中文译名对变速箱振动分析的诊断和预测方法综述学院汽车工程学院专业班级车辆工程七班学生姓名刘嘉先学生学号201130085184指导教师李利平填写日期2015年3月15日英文原文版出处:Proceedings of the 54th Meeting of the Society for Machinery Failure Prevention Technology, Virginia Beach,V A, May 1-4,2000,p. 623-634译文成绩:指导教师(导师组长)签名:译文:简介特征提取技术在文献中有描述;然而,大多数人似乎掩盖所需的特定的预处理功能。
一些文件没有提供足够的细节重现他们的结果,并没有一个全面的比较传统的功能过渡齿轮箱数据。
常用术语,如“残差信号”,是指在不同的文件不同的技术.试图定义了状态维修社区中的常用术语和建立所需的特定的预处理加工特性。
本文的重点是对所使用的齿轮故障检测功能。
功能分为五个不同的组基于预处理的需要。
论文的第一部分将提供预处理流程的概述和其中每个特性计算的处理方案。
在下一节中,为特征提取技术描述,将更详细地讨论每一个功能。
最后一节将简要概述的宾夕法尼亚州立大学陆军研究实验室的CBM工具箱用于齿轮故障诊断。
特征提取概述许多类型的缺陷或损伤会增加机械振动水平。
这些振动水平,然后由加速度转换为电信号进行数据测量。
原则上,关于受监视的计算机的健康的信息被包含在这个振动签名。
因此,新的或当前振动签名可以与以前的签名进行比较,以确定该元件是否正常行为或显示故障的迹象。
在实践中,这种比较是不能奏效的。
由于大的变型中,签名的直接比较是困难的。
相反,一个涉及从所述振动署名数据特征提取更多有用的技术也可以使用。
毕业设计外文原文及翻译

Thermal analysis for the feed drive system of a CNC machineAbstractA high-speed drive system generates more heat through friction at contact areas, such as the ball-screw and the nut, thereby causing thermal expansion which adversely affects machining accuracy. Therefore, the thermal deformation of a ball-screw is oneof the most important objects to consider for high-accuracy and high-speed machine tools. The objective of this work was to analyze the temperature increase and the thermal deformation of a ball-screw feed drive system. The temperature increase was measured using thermocouples, while a laser interferometer and a capacitance probe were applied to measure the thermal error of the ball-screw.Finite element method was used to analyze the thermal behavior of a ball-screw. The measured data were compared with numerical simulation results. Inverse analysis was applied to estimate the strength of the heat source from the measured temperature profile.The generated heat sources for different feed rates were investigated.Keywords:Machine tool; Ball-screw; Thermal error; Finite element method; Thermocouple1. IntroductionPrecise positioning systems with high speed, high resolution and long stroke become more important in ultra-precision machining. The development of high-speed feed drive systems has been a major issue in the machine-tool industry. A high-speed feed drive system reduces necessary non-cutting time. However, due to the backlash and friction force between the ball-screw and the nut, it is difficult to provide a highly precise feed drive system.Most current research is focused on the thermal error compensation of the whole machine tools. Thermally induced error is a time-dependent nonlinear process caused by nonuniform temperature variation in the machine structure. The interaction between the heat source location, its intensity, thermal expansion coefficient and the machine system configuration creates complex thermal behavior . Researchers have employed various techniques namely finite element methods,coordinate transformation methods, neural net-works etc., in modelling the thermal characteristicsA high-speed drive system generates more heat through friction at contact areas, such as the ball-screw and the nut, thereby causing thermal expansion which adversely affects machining accuracy. Therefore, the thermal deformation of a ball-screw is one of the most important objects to consider for high-accuracy and high-speed machine tools [5]. In order to achieve high-precision positioning, pre-load on the ball-screw is necessary to eliminate backlash. ball-screw pre-load also plays an important role in improving rigidity, noise, accuracy and life of the positioning stage [6]. However, pre-load also produces significant friction between the ball-screw and the nut that generates greater heat, leading to large thermal deformation of the ball-screw and causing low positioning accuracy. Consequently, the accuracy of the main system, such as a machine tool, is affected. There-fore, anoptimum pre-load of the ball-screw is one of the most important things to consider for machine tools with high accuracy and great rigidity.Only a few researchers have tackled this problem with some success. Huang used the multiple regression method to analyze the thermal deformation of a ball-screw feed drive system. Three temperature increases at front bearing, nut and back bearing were selected as independent variables of the analysis model. The multiple-regression model may be used to predict the thermal deformation of the ball-screw. Kim et al. Analyzed the temperature distribution along a ball-screw system using finite element methods with bilinear type of elements. Heat induced due to friction is the main source of deformation in a ball-screw system, the heat generated being dependent on the pre-load, the lubrication of the nut and the assembly conditions. The proposed FEM model was based on the assumption that the screw shaft and nut are a solid and hollow shaft respectively. Yun et al. used the modified lumped capacitance method and genius education algorithm to analyze the linear positioning error of the ball-screw.The objective of this work was to analyze the temperature increase and the thermal deformation of a ball-screw feed drive system. The temperature increase was measured using thermocouples while a laser interferometer and a capacitance probe were applied to meas-ure the thermal error of the ball-screw. Finite element method was also applied to simulate the thermal behavior of the ball-screw. Measured data were compared with numerical simulation results. Inverse analysis was applied to estimate the strength of the heat source from the measured temperature pro file. Generated heat sources for different feed rates were investigated.2 Experimental set-up and procedureIn this study, the object used to investigate the thermal characteristics of a ball-screw system is a machine center as shown in Fig. 1. The maximum rapid speed along thex-axis of the machine center is 40 m/min and the x-axis travel is 800 mm. The table repeatedly moved along the x-axis with a stroke of 600 mm. The main heat sourceFig. 1. Photograph of machine center.of the ball-screw system is the friction caused by a moving nut and rotating bearings. The produced temperature increase and thermal deformation were measured to study the thermal characteristics of the ball-screw system.In order to measure the temperature increase and the thermal deformation of a ball-screw system under long-term movement of the nut, experiments were performed with the arrangement shown in Fig. 2. Temperatures at nine points were measured as shown in Fig. 2a .Two thermocouples (numbered 1 and 8) were located on the rear and front bearing surfaces, respectively. They were used to measure the surface temperatures of these two support bearings. The last one (numbered 9) was used to measure the room temperature. The recorded room temperature was to eliminate the effect of environmental variation. These three thermocouples were used for continuous acquisition under moving conditions. The other six thermocouples (numbered 2 –7) were used to measure the surface temperatures of the ball-screw. Because the moving nut covered most of the ball-screw, thermocouples cannot be consistently fixed on the ball-screw. While temperature measurement was necessary, the ball-screw stopped running and these six thermocouples were quickly attached to specified locations of the ball-screw. Having collected the required data, the thermocouples were quickly removed.Thermal deformation errors were simultaneously measured with two methods. Because a thrust bearing is used on the driving side of the ball-screw, this end is considered to be fixed. A capacitance probe was installed next to the driven side of the ball-screw with a direction perpendicular to the side surface as shown in Fig. 2b. This probe was used to record the whole thermal deformation of the ball-screw. The values can be collected continuously during running conditions. The second method is used to measure the thermal error distribution at some specified time. Before the feed drive system started to operate, the original positional error distribution was measured with a laser interferometer (HP5528A). The table moved step-by-step (the increment ofFig. 2. Locations of measured points for (a) temperatures and (b) thermal errors.each step was 100 mm) and the positioning error was recorded at each step. Then the feed drive system started to operate and generate heat. After a certain time interval, the feed drive system stopped to measure thermal errors. In the same way, the positioning error distribution was again collected with the laser interferometer. Subtracting the actual error from the original error of the x-coordinates, the results are net thermal errors. Having collected the temperature increase (with thermocouples) and deformation distribution, the feed drive system starts running again.In this study, three feed rates (10, 15 and 20 m/min) along the x-axis and three different pre-loads (0, 150 and 300 kgf·cm) were used. The table moved along the x-axis in a reciprocating motion and the stroke was 600mm. The point temperatures and thermal errors were measured at sampling intervals of 10 min. Each stopping time was only about 10 s. These procedures were operated repeatedly until the temperature reached a steady state.3. Experimental results and discussionThe developed experimental setup was utilized for three constant feed rates (running at 10, 15 and 20m/min, respectively). The table reciprocated until point temperatures and thermal errors reached a steady state. Firstly, the ball-screw pre-load was zero and its thermal characteristics were studied. In Fig. 3, temperature variationsFig. 3. (a) Measured temperature increase and (b) thermal error over time for feed rate of 10 m/min and zero-pre-load.and thermal errors of the feed drive system are shown over time for a feed rate of 10 m/min. Measurements can also be made for feed rates of 15 and 20m/min. Themeasured data at a steady state are shown in Tables 1 and 2 . A brief discussion can be made as fol-lows.1. The higher feed rate produces larger frictional heat at the interface between the ball-screw and the nut. The frictional heat generated by the support bearings and the motor also increases with the feed rate. Therefore, the temperature of the ball-screw increases with the feed rate.2. The table travels over the middle part with a 600 mm stroke. The central part of the ball-screw reveals a higher temperature increase. Support bearings do not have high temperature increase because the bearing pre-load is zero.3. A higher rotational speed brings a larger thermal expansion for the ball-screw. The middle part of the screw has a slightly larger thermal expansion because of its higher temperature increase. However, this phenomenon is not obvious. The thermal error at some specified point of the ball-screw is approximately proportional to the distance between this point and the front end (the motor-driving side of the screw). Secondly, the ball-screw pre-load was set at 150kgf·cm and its thermal characteristics were studied. In Figs. 4 –5, temperature variations around the feed drive system and thermal errors are shown over time for feed rates of 10 and 15 m/min. Measured data are shown in Tables 1 and 2. Results reveal two interesting phenomena shown as follows.1. Temperature increases of measured points grow gradually until the ball-screw reaches a steady state except for the temperature increase of the bearing on the driven side. The temperature of this bearing quickly reaches a maximum value and then gradually drops.2. The thermal errors of P6, P7 and P8 are negative at the steady state. It means that these three points move to the driving side due to thermal expansion, while other points move to the driven side. Furthermore, the thermal errors of P4 to P8 show a gradual decrease after 60 min.These phenomena are different from previous results with no pre-load. Some experiments were carried out to study these phenomena. We found that the two bearing stands bent if the ball-screw was pre-loaded. After the pre-load was applied on the ball-screw, the original positional error distribution was measured using a laser interferometer. At this moment, the bending effects on error distribution were includ- Table 1Temperature distribution at steady state with different pre-loads and feed rates (unit: °C)Table 2Thermal error distribution at steady state with different pre-loads and feed rates (unit:µm)Fig. 4. (a) Measured temperature increase Fig. 5. (a) Measured temperature increase and (b) thermal error over time for feed and (b) thermal error over time for feed rate of 10 m/min and pre-load of rate of 15 m/min and pre-load of150kgf ·cm. 150kgf ·cm.-ed in the measured positioning error. The feed drive system starts to run and the ball-screw expands. The expansion relaxes the pre-load of the ball-screw and the bending deformation of two bearing stands. Therefore, the points on the driving side move closer to the motor, thereby thermal errors are negative, nevertheless, the points on the driven side move to the free end, thereby thermal errors are positive.The temperature change of the rear bearing was also investigated. A journal bearing was applied on the driven side and a thrust bearing was applied on the driving side. The pre-load of the ball-screw increases the pre-load of the bearing on the driven side. When the feed drive system runs, the bearing temperature on the driven side sharplyincreases due to the rising pre-load. However, the thermal expansion of the ball-screw relaxes the ball-screw and decreases pre-load of the bearing on the driven side. Therefore, the temperature gradually decreases to a steady state.Finally, the ball-screw pre-load was set to 300kgf·cm and its thermal characteristics were studied. In Figs. 6 and 7, temperature variations around the feed drive systemFig. 6. (a) Measured temperature increase and (b) thermal error over time for feed rate of 10 m/min and pre-load of 300kgf ·cm.Fig. 7. (a) Measured temperature increase and (b) thermal error over time for feed rate of 15 m/min and pre-load of 300kgf ·cm.and thermal errors are shown over time for feed rates of 10 and 15 m/min. The tendency with a 300kgf·cm is similar to that with a 150kgf·cm. Measured data are shown in Tables 1 and 2.4. Numerical simulationThe main heat source of a ball-screw system is the friction caused by a moving nut and the support bearings. In this study, temperature distribution was calculated using the FEM based on the following assumptions:1. The screw shaft is a solid cylinder.2. Friction heat generation between the moving nut and the screw shaft is uniform at contacting surface and is proportional to contacting time.3. Heat generation at support bearings is also constant per unit area and unit time.4. Convective heat coefficients are always constant during moving. The radiation term is neglected.The problem is de fined as transient heat conduction in non-deforming media without radiation. A classical form of the initial/boundary value problem is shown below:where is the internal heat generation rate, q the entering heat flux, a unit outward normal vector, the ambient temperature and h the convective heattrans-fer coefficient at a given boundary. A simplified heat transfer model of the ball-screw system is described in Fig. 8 along with the boundary conditions. The nut moves reciprocally with a stroke, s. The length of the nut is w. According to the previously mentioned assumption, No. 2, frictional heat fluxes on the balls-crew are shown as in Fig. 8b . Both ends of the balls-crew are subjected to frictional heat fluxq and q caused by the support bearings. Heat fluxes on rear and front ends are13 respectively. Other surfaces are subjected to convection heat transfer as shown in Fig.8c .To obtain an approximate solution, Eqs. (1)–(3) may be transformed through discretization into algebraic expressions that may be solved for unknowns. In orderto allow the replacement of the continuous system by an equivalent discrete system, the original domain is divided into elements. Four-node tetrahedral elements are chosen in this study. Elements and nodes of the balls-crew for FEM are shown in Fig.9.Once temperature distribution is obtained, the thermal expansion of the balls-crew may be predicted. In the case of linearly elastic isotropic three-dimensional solid, stress–strain relations are given by Hooke ’s law as [9]:of balls-crew.Fig. 9. Elements and nodes of ball-screw for FEM.where [C] is a matrix of elastic coefficients and 0ε→is the vector of initial strains. In the case of heating of an isotropic material, the initial strain vector is given by:where a is the coefficient of thermal expansion and T is the temperature change. Three unknowns 123,,q q and q are to be determined with inverse analysis. Firstly, initial guessing of these heat fluxes is applied in FEM simulation to obtain the temperature distribution of the balls-crew. If numerical results do not agree with the measured temperature distribution, the values of 123,,q q and q are adjusted iteratively until numerical and simulation results are in good agreement.Calculated values of 123,,q q and q for an un-pre-loaded ball-screw are listed in Table3. Measured and simulated temperature distributions for feed rates of 10, 15 and 20 m/min are indicated in Fig. 10. For each feed rate, it shows a good agreement between measured and simulated temperature distributions. The numerical program can also be used to simulate the thermal expansion of the ball-screw based on the calculated heatTable 3Values of heat flux at different locations (unit:2W m)/Fig. 10. Temperature increase from experimental measurement and numerical simulation for feed rate of (a) 10 m/min, (b) 15 m/min and (c) 20 m/min.fluxes. Measured and simulated thermal expansions of the ball-screw are compared as shown in Table 4. Thermal expansions also show good agreement with each other. From Table 3, the heat flux increases with the feed rate. Approximate linear relation can be found between the heat flux and the feed rate under the same operating condition.5. ConclusionsThis paper proposes a systematic method to investigate the thermal characteristics of a feed drive system. The approach measures the temperature increase and the thermal deformation under long-term movement of the working table. A simplified FEM model for the ball-screw was developed. The FEM model incorporated with themeasured temperature distribution was used to determine the strength of the frictional heat source by inverse analysis. The strength of the heat source was applied to the FEM model to calculate the thermal errors of the feed drive system. Calculated and measured thermal errors were found to agree with each other. From the results, the following conclusions can be drawn:1. The positional accuracy increases while closer to the driving side of the ball-screw. The thermal error increases with the distance to the driven side of the ball-screw. The maximum thermal error occurs at the driven side of the ball-screw (free end). This value can be taken as the total thermal error of the ball-screw and may be measured with a capacitance probe.2. The ball-screw pre-load raises the temperature increases of both support bearings, especially the bearing on the driven side. The surface temperature of the ball-screw decreases because the thermal effects relax the pre-load, thereby decreasing the friction between the nut and the ball-screw.3. The thermal expansion of the ball-screw increases with the feed rate, thereby increasing the positional error. However, the increasing pre-load reduces thermal errors and improves the positional accuracy of the feed drive system.4.Two bearing stands may bend if the ball-screw is pre-loaded. The thermal expansi Table 4Thermal errors at different feed rates-on relaxes the pre-load of the ball-screw and the bending deformation of two bearing stands. Therefore, the points on the motor side move closer to the motor and the thermal errors are negative; nevertheless, the points on the free side move to the free end and the thermal errors are positive.数控加工中心进给驱动系统的热分析摘要高速驱动系统在接触区域(如滚珠丝杠和螺母)通过摩擦产生大量的热,从而导致热膨胀,热膨胀严重地影响机械加工精度。
软件工程专业毕业设计外文文献翻译

软件工程专业毕业设计外文文献翻译1000字本文将就软件工程专业毕业设计的外文文献进行翻译,能够为相关考生提供一定的参考。
外文文献1: Software Engineering Practices in Industry: A Case StudyAbstractThis paper reports a case study of software engineering practices in industry. The study was conducted with a large US software development company that produces software for aerospace and medical applications. The study investigated the company’s software development process, practices, and techniques that lead to the production of quality software. The software engineering practices were identified through a survey questionnaire and a series of interviews with the company’s software development managers, software engineers, and testers. The research found that the company has a well-defined software development process, which is based on the Capability Maturity Model Integration (CMMI). The company follows a set of software engineering practices that ensure quality, reliability, and maintainability of the software products. The findings of this study provide a valuable insight into the software engineering practices used in industry and can be used to guide software engineering education and practice in academia.IntroductionSoftware engineering is the discipline of designing, developing, testing, and maintaining software products. There are a number of software engineering practices that are used in industry to ensure that software products are of high quality, reliable, and maintainable. These practices include software development processes, software configuration management, software testing, requirements engineering, and project management. Software engineeringpractices have evolved over the years as a result of the growth of the software industry and the increasing demands for high-quality software products. The software industry has developed a number of software development models, such as the Capability Maturity Model Integration (CMMI), which provides a framework for software development organizations to improve their software development processes and practices.This paper reports a case study of software engineering practices in industry. The study was conducted with a large US software development company that produces software for aerospace and medical applications. The objective of the study was to identify the software engineering practices used by the company and to investigate how these practices contribute to the production of quality software.Research MethodologyThe case study was conducted with a large US software development company that produces software for aerospace and medical applications. The study was conducted over a period of six months, during which a survey questionnaire was administered to the company’s software development managers, software engineers, and testers. In addition, a series of interviews were conducted with the company’s software development managers, software engineers, and testers to gain a deeper understanding of the software engineering practices used by the company. The survey questionnaire and the interview questions were designed to investigate the software engineering practices used by the company in relation to software development processes, software configuration management, software testing, requirements engineering, and project management.FindingsThe research found that the company has a well-defined software development process, which is based on the Capability Maturity Model Integration (CMMI). The company’s software development process consists of five levels of maturity, starting with an ad hoc process (Level 1) and progressing to a fully defined and optimized process (Level 5). The company has achieved Level 3 maturity in its software development process. The company follows a set of software engineering practices that ensure quality, reliability, and maintainability of the software products. The software engineering practices used by the company include:Software Configuration Management (SCM): The company uses SCM tools to manage software code, documentation, and other artifacts. The company follows a branching and merging strategy to manage changes to the software code.Software Testing: The company has adopted a formal testing approach that includes unit testing, integration testing, system testing, and acceptance testing. The testing process is automated where possible, and the company uses a range of testing tools.Requirements Engineering: The company has a well-defined requirements engineering process, which includes requirements capture, analysis, specification, and validation. The company uses a range of tools, including use case modeling, to capture and analyze requirements.Project Management: The company has a well-defined project management process that includes project planning, scheduling, monitoring, and control. The company uses a range of tools to support project management, including project management software, which is used to track project progress.ConclusionThis paper has reported a case study of software engineering practices in industry. The study was conducted with a large US software development company that produces software for aerospace and medical applications. The study investigated the company’s software development process,practices, and techniques that lead to the production of quality software. The research found that the company has a well-defined software development process, which is based on the Capability Maturity Model Integration (CMMI). The company uses a set of software engineering practices that ensure quality, reliability, and maintainability of the software products. The findings of this study provide a valuable insight into the software engineering practices used in industry and can be used to guide software engineering education and practice in academia.外文文献2: Agile Software Development: Principles, Patterns, and PracticesAbstractAgile software development is a set of values, principles, and practices for developing software. The Agile Manifesto represents the values and principles of the agile approach. The manifesto emphasizes the importance of individuals and interactions, working software, customer collaboration, and responding to change. Agile software development practices include iterative development, test-driven development, continuous integration, and frequent releases. This paper presents an overview of agile software development, including its principles, patterns, and practices. The paper also discusses the benefits and challenges of agile software development.IntroductionAgile software development is a set of values, principles, and practices for developing software. Agile software development is based on the Agile Manifesto, which represents the values and principles of the agile approach. The manifesto emphasizes the importance of individuals and interactions, working software, customer collaboration, and responding to change. Agile software development practices include iterative development, test-driven development, continuous integration, and frequent releases.Agile Software Development PrinciplesAgile software development is based on a set of principles. These principles are:Customer satisfaction through early and continuous delivery of useful software.Welcome changing requirements, even late in development. Agile processes harness change for the customer's competitive advantage.Deliver working software frequently, with a preference for the shorter timescale.Collaboration between the business stakeholders and developers throughout the project.Build projects around motivated individuals. Give them the environment and support they need, and trust them to get the job done.The most efficient and effective method of conveying information to and within a development team is face-to-face conversation.Working software is the primary measure of progress.Agile processes promote sustainable development. The sponsors, developers, and users should be able to maintain a constant pace indefinitely.Continuous attention to technical excellence and good design enhances agility.Simplicity – the art of maximizing the amount of work not done – is essential.The best architectures, requirements, and designs emerge from self-organizing teams.Agile Software Development PatternsAgile software development patterns are reusable solutions to common software development problems. The following are some typical agile software development patterns:The Single Responsibility Principle (SRP)The Open/Closed Principle (OCP)The Liskov Substitution Principle (LSP)The Dependency Inversion Principle (DIP)The Interface Segregation Principle (ISP)The Model-View-Controller (MVC) PatternThe Observer PatternThe Strategy PatternThe Factory Method PatternAgile Software Development PracticesAgile software development practices are a set ofactivities and techniques used in agile software development. The following are some typical agile software development practices:Iterative DevelopmentTest-Driven Development (TDD)Continuous IntegrationRefactoringPair ProgrammingAgile Software Development Benefits and ChallengesAgile software development has many benefits, including:Increased customer satisfactionIncreased qualityIncreased productivityIncreased flexibilityIncreased visibilityReduced riskAgile software development also has some challenges, including:Requires discipline and trainingRequires an experienced teamRequires good communicationRequires a supportive management cultureConclusionAgile software development is a set of values, principles, and practices for developing software. Agile software development is based on the Agile Manifesto, which represents the values and principles of the agile approach. Agile software development practices include iterative development, test-driven development, continuous integration, and frequent releases. Agile software development has many benefits, including increased customer satisfaction, increased quality, increased productivity, increased flexibility, increased visibility, and reduced risk. Agile software development also has some challenges, including the requirement for discipline and training, the requirement for an experienced team, the requirement for good communication, and the requirement for a supportive management culture.。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
毕业设计外文资料题目面对对象技术学院信息科学与工程学院专业计算机科学与技术班级计软1202学生刘桂斌学号20121214073指导教师史桂娴,王海燕二〇一六年三月二十日Object Technology2004, Vol.14 (2), pp.20.Object TechnologyTimothy A.BuddAbstract Object technology is a new approach to developing software that allows programmers to create objects, a combination of data and program instructions. This new technology has been steadily developed since the late 1960s and promises to be one of the major ingredients in the response to the ongoing software crisis.Keywords Object technology Optimization1.1 Introduction to OTThere exists a critical technology that is changing the way we conceive, build, use and evolve our computer systems. It is a technology that many companies are adopting to increase their efficiency, reduce costs and adapt to a dynamic marketplace. It is called Object Technology (OT).By allowing the integration of disparate and non compatible source, OT has the potential to precipitate a revolution in information systems design on a par with that caused in computer hardware by the introduction of the computer chip. Yet OT is not a newphenomenon. Development and product releases have been ongoing since its origin many years ago. However, the recent emphasis task of enterprise information technology integration has brought OT into the spotlight.OT promises to provide component-level software objects that can be quickly combined to build new applications that respond to changing business conditions. Once used, objects may be reused in other applications, lowering development costs and speeding up the development process. Because objects communicate by sending messages that can be understood by other objects, large integrated systems are easier to assemble.Each object is responsible for a specific function within either an application or a distributed system. That means that as the business changes, individual object may be easily upgraded, augmented or replaced, leaving the rest of the system untouched. This directly reduces the cost of maintenance and the timing and extendibility of new systems.1.2 OT-based ProductsThe current market for OT-based products can be divided into four major segments: ·Languages and programming tools·Developers’ toolkits·Object-Oriented database·Object-Oriented CASE toolsThe largest segment of the current market for OT-based products is languages andprogramming tools. Products in this area include language compliers for C++, Smalltalk, Common Lisp Object System (CLOS), Eiffel, Ada and Objective-C, as well as extensions to PASCAL and Modula-2.Products in this category are available from a variety of vendors. Increasingly, the trend in this group is to offer the language compliers with associated development tools as part of a complete development environment.Developers’ toolkits account for the next largest part of the OT market. These products are designed to develop a program that enables a developer to easily do one of two things. The first is interfacing an application to distributed environment. The second is developing a graphical screen through a product.By providing developers with higher level description language and reusablecomponents, products in this category give developers an easy and cost effective way to begin producing object-oriented systems.An important component in this category is the relatively new area of end-users tools. This element is important because organizing and analying the increasingly large amounts of data that computer systems are capable of collecting is a key problem.Object-oriented database management systems are one of the most interesting and rapidly growing segments of the OT market. A number of companies, including systems vendors like Digital and HP, and start-ups such as Object Design, Servio, and Objectivity, have all produced products.These products, dubbed ‖Objectbases‖, fill an important need by storing complexobjects as a single entity. The objectbase products allow objects to be stored, retrieved and shared in much the same way as data is stored in a relational database management system. The value of an objectbase, as opposed to a database, is best described as following: ―Object databases offer a better way to store objects because they provide all of the traditional database services without the overhead of disassembling and reassemblingobjects every time they are stored and retrieved. Compared with an object database, storing complex objects in a relational database is tedious at best. It’s like having to disassembling your car each night rather than just putting it into the gar age!‖Over the next few years, a shift from proprietary CASE implementations to those based on the object paradigm can be expected. This area has lagged growth from earlier projections. OT-based CASE tools will have to emerge as a viable product category to address the wide scale development of large systems. This category also include those tools that are methodological in nature.1.3 0bject-oriented ProgrammingObject-oriented programming (OOP) is a new approach to developing software that allows programmers to create objects, a combination of data and program instructions. Traditional programming methods keep data, such as files, independent of the programs that work with the data. Each traditional program, t5herfore, must define how the data will be used for that particular program. This often results in redundant programming code that must be changed every time the structure of the data is changed, such as when a new field is added to a file. With OOP, the program instructions and data are combined into objects that can be used repeatedly by programmers whenever they need them. Specificinstructions, called methods define how the object acts when it is used by a program.With OOP, programmers define classes of objects. Each class contains the methods that are unique to that class. Each class can have one or more subclasses. Each subclass contains the methods of its higher level classes plus whatever methods are unique to the subclass. The OOP capability to pass methods to lower levels is called ―inheritance‖.A specific instance of an object contains all methods from its higher level classes plus any methods that a unique to the object. When an OOP object is sent an instruction to do something, called a message, unlike a traditional program, the message does not have to tell the OOP object exactly what to do. What to do is defined by the methods that the OOP object contains or has inherited.Object—oriented programming can bring many advantages to users. It can bring productivity gains as high as 1000 to 1500 percent instead of the 10 or 15 percent gainsavailable from structured programming methods. It allows large complex systems to be built which are not economically feasible using traditional programming techniques. It allows program modifications to be made more easily. It could mean two different user interfaces within an application, one for the user who likes to type, and another for the users who just want to shout at the terminal.Objects can be viewed as reusable components, and once the programmer has developed a library of these components, he can minimize the amount of new coding required. One user envisions a commercial library of objects which could be purchased byprogrammers and reused for various applications. But creating a library is no simple task because the integrity of the original software design is critical. Reusability can be a mixed blessing for users, too, as a programmers has to be able to find the object he needs. But if productivity is your aim, reusability is worth the risks.The long-term productivity of systems is enhanced by object-oriented programming. Because of the modular nature of the code, programs are more malleable. This is particularly beneficial for applications that will be used for many years, during which company needs may change and make software modifications necessary.Software reliability can be improved by object-oriented programming. Since the objects are repeatedly tested in a variety of applications, bugs are more likely to be found and corrected. Object-oriented programming also has potential benefits in parallel processing. Execution speed under object oriented methods will improve with parallel processing.1.4 Object-oriented DBMSA shift toward object-oriented DBMSs does not have to replace relational DNMS. As its name implies, it is orientation rather than a full-blown DBMS model. As such, it can blend with and build on the relational schema.Object-oriented DBMSs integrate a variety of real-world data types –such as business procedures and policies, graphics, pictures, voice, and an non-tated text. Current relational products are not equipped to handle them efficiently. Data types in RDBMSs are more commonly record-oriented and expressed in numbers and text.Object orientation also makes contributions to application development efficiency.makes the data function, attributes, and relationships an integral part of the object. In this way, objects can be reused and replicated. You can query the data on its functions, attributes, and relationships.By contrast, most RDBMSs demand that the knowledge associated with the data be written into and maintained separately in each application program.Object orientation is going to be available in two forms: one for those who need and want a radical change, and one for those who want some of its advantages without going through a major conversion.The first form of object-oriented DBMS focused largely on the computer-aided design (CAD) market, which needed to store complex data types such as the graphics involved with an aircraft design.The second form is made up of the leading RDBMS vendors who support the concept of integrating object management capabilities whit their current line of relational products. Sybase, Inc, the first vendor to introduce an object-oriented capability,offers Sybase , which enables the user to program a limited number of business procedures along with the data types in a server’s database engine . Any client attempting a transaction that does not conform to these procedures is simply rejected by the database. That capability enables users to shorten the development cycle, since integrity logic and business rules no longer need to be programmed into each application.This approach reduces maintenance costs as well, since any changes in the procedure can be made once at the server level instead of several times within all the affected applications.Last, the server-level procedures increase the system’s performance, since the operations are taking place closer to where the data is actually stored.。