计算机系——外文翻译(中英对照-3000汉字左右)

毕业设计(论文)外文资料翻译

系别计算机信息与技术系

专业计算机科学与技术

班级

姓名

学号

外文出处

附件 1. 原文; 2. 译文

2012年3月

History of computing

Main article: History of computing hardware

The first use of the word "computer" was recorded in 1613, referring to a person who carried out calculations, or computations, and the word continued with the same meaning until the middle of the 20th century. From the end of the 19th century the word began to take on its more familiar meaning, a machine that carries out computations.

Limited-function early computers

The Jacquard loom, on display at the Museum of Science and Industry in Manchester, England, was one of the first programmable devices.

The history of the modern computer begins with two separate technologies, automated calculation and programmability, but no single device can be identified as the earliest computer, partly because of the inconsistent application of that term. A few devices are worth mentioning though, like some mechanical aids to computing, which were very successful and survived for centuries until the advent of the electronic calculator, like the Sumerian abacus, designed around 2500 BC of which a descendant won a speed competition against a modern desk calculating machine in Japan in 1946, the slide rules, invented in the 1620s, which were carried on five Apollo space missions, including to the moon and arguably the astrolabe and the Antikythera mechanism, an ancient astronomical computer built by the Greeks around 80 BC. The Greek mathematician Hero of Alexandria (c. 10–70 AD) built a mechanical theater which performed a play lasting 10 minutes and was operated by a complex system of ropes and drums that might be considered to be a means of deciding which parts of the mechanism performed which actions and when. This is the essence of programmability.

Around the end of the 10th century, the French monk Gerbert d'Aurillac brought back from Spain the drawings of a machine invented by the Moors that answered either Yes or No to the questions it was asked. Again in the 13th century, the monks Albertus Magnus and Roger Bacon built talking androids without any further development.

In 1642, the Renaissance saw the invention of the mechanical calculator, a device that could perform all four arithmetic operations without relying on human intelligence. The mechanical calculator was at the root of the development of

computers in two separate ways. Initially, it was in trying to develop more powerful and more flexible calculators that the computer was first theorized by Charles Babbage and then developed. Secondly, development of a low-cost electronic calculator, successor to the mechanical calculator, resulted in the development by Intel of the first commercially available microprocessor integrated circuit.

First general-purpose computers

In 1801, Joseph Marie Jacquard made an improvement to the textile loom by introducing a series of punched paper cards as a template which allowed his loom to weave intricate patterns automatically. The resulting Jacquard loom was an important step in the development of computers because the use of punched cards to define woven patterns can be viewed as an early, albeit limited, form of programmability.

In 1837, Charles Babbage was the first to conceptualize and design a fully programmable mechanical computer, his analytical engine. Limited finances and Babbage's inability to resist tinkering with the design meant that the device was never completed ; nevertheless his son, Henry Babbage, completed a simplified version of the analytical engine's computing unit (the mill) in 1888. He gave a successful demonstration of its use in computing tables in 1906. This machine was given to the Science museum in South Kensington in 1910.

In the late 1880s, Herman Hollerith invented the recording of data on a machine-readable medium. Earlier uses of machine-readable media had been for control, not data. "After some initial trials with paper tape, he settled on punched cards ..." To process these punched cards he invented the tabulator, and the keypunch machines. These three inventions were the foundation of the modern information processing industry. Large-scale automated data processing of punched cards was performed for the 1890 United States Census by Hollerith's company, which later became the core of IBM. By the end of the 19th century a number of ideas and technologies, that would later prove useful in the realization of practical computers, had begun to appear: Boolean algebra, the vacuum tube (thermionic valve), punched cards and tape, and the teleprinter.

During the first half of the 20th century, many scientific computing needs were met by increasingly sophisticated analog computers, which used a direct mechanical or electrical model of the problem as a basis for computation. However, these were not programmable and generally lacked the versatility and accuracy of modern digital

computers.

Alan Turing is widely regarded as the father of modern computer science. In 1936 Turing provided an influential formalisation of the concept of the algorithm and computation with the Turing machine, providing a blueprint for the electronic digital computer. Of his role in the creation of the modern computer, Time magazine in naming Turing one of the 100 most influential people of the 20th century, states: "The fact remains that everyone who taps at a keyboard, opening a spreadsheet or a word-processing program, is working on an incarnation of a Turing machine".

EDSAC was one of the first computers to implement the stored-program (von Neumann) architecture.

Die of an Intel 80486DX2 microprocessor (actual size: 12×6.75 mm) in its packaging.

The Atanasoff–Berry Computer (ABC) was the world's first electronic digital computer, albeit not programmable. Atanasoff is considered to be one of the fathers of the computer.Conceived in 1937 by Iowa State College physics professor John Atanasoff, and built with the assistance of graduate student Clifford Berry, the machine was not programmable, being designed only to solve systems of linear equations. The computer did employ parallel computation. A 1973 court ruling in a patent dispute found that the patent for the 1946 ENIAC computer derived from the Atanasoff–Berry Computer.

The first program-controlled computer was invented by Konrad Zuse, who built the Z3, an electromechanical computing machine, in 1941. The first programmable electronic computer was the Colossus, built in 1943 by Tommy Flowers.

George Stibitz is internationally recognized as a father of the modern digital computer. While working at Bell Labs in November 1937, Stibitz invented and built a relay-based calculator he dubbed the "Model K" (for "kitchen table", on which he had assembled it), which was the first to use binary circuits to perform an arithmetic operation. Later models added greater sophistication including complex arithmetic and programmability.

A succession of steadily more powerful and flexible computing devices were constructed in the 1930s and 1940s, gradually adding the key features that are seen in modern computers. The use of digital electronics (largely invented by Claude Shannon in 1937) and more flexible programmability were vitally important steps, but defining one point along this road as "the first digital electronic computer" is difficult.

Notable achievements include. Konrad Zuse's electromechanical "Z machines". The Z3 (1941) was the first working machine featuring binary arithmetic, including floating point arithmetic and a measure of programmability. In 1998 the Z3 was proved to be Turing complete, therefore being the world's first operational computer.

The non-programmable Atanasoff–Berry Computer (commenced in 1937, completed in 1941) which used vacuum tube based computation, binary numbers, and regenerative capacitor memory. The use of regenerative memory allowed it to be much more compact than its peers (being approximately the size of a large desk or workbench), since intermediate results could be stored and then fed back into the same set of computation elements.

The secret British Colossus computers (1943), which had limited programmability but demonstrated that a device using thousands of tubes could be reasonably reliable and electronically reprogrammable. It was used for breaking German wartime codes.

The Harvard Mark I (1944), a large-scale electromechanical computer with limited programmability.

The U.S. Army's Ballistic Research Laboratory ENIAC (1946), which used decimal arithmetic and is sometimes called the first general purpose electronic computer (since Konrad Zuse's Z3 of 1941 used electromagnets instead of electronics). Initially, however, ENIAC had an inflexible architecture which essentially required rewiring to change its programming.

Stored-program architecture

Replica of the Small-Scale Experimental Machine (SSEM), the world's first stored-program computer, at the Museum of Science and Industry in Manchester, England

Several developers of ENIAC, recognizing its flaws, came up with a far more flexible and elegant design, which came to be known as the "stored-program architecture" or von Neumann architecture. This design was first formally described by John von Neumann in the paper First Draft of a Report on the EDV AC, distributed in 1945. A number of projects to develop computers based on the stored-program architecture commenced around this time, the first of which was completed in 1948 at the University of Manchester in England, the Manchester Small-Scale Experimental Machine (SSEM or "Baby"). The Electronic Delay Storage Automatic Calculator

(EDSAC), completed a year after the SSEM at Cambridge University, was the first practical, non-experimental implementation of the stored-program design and was put to use immediately for research work at the university. Shortly thereafter, the machine originally described by von Neumann's paper—EDV AC—was completed but did not see full-time use for an additional two years.

Nearly all modern computers implement some form of the stored-program architecture, making it the single trait by which the word "computer" is now defined. While the technologies used in computers have changed dramatically since the first electronic, general-purpose computers of the 1940s, most still use the von Neumann architecture.

Beginning in the 1950s, Soviet scientists Sergei Sobolev and Nikolay Brusentsov conducted research on ternary computers, devices that operated on a base three numbering system of ?1, 0, and 1 rather than the conventional binary numbering system upon which most computers are based. They designed the Setun, a functional ternary computer, at Moscow State University. The device was put into limited production in the Soviet Union, but supplanted by the more common binary architecture.

Semiconductors and microprocessors

Computers using vacuum tubes as their electronic elements were in use throughout the 1950s, but by the 1960s had been largely replaced by semiconductor transistor-based machines, which were smaller, faster, cheaper to produce, required less power, and were more reliable. The first transistorised computer was demonstrated at the University of Manchester in 1953. In the 1970s, integrated circuit technology and the subsequent creation of microprocessors, such as the Intel 4004, further decreased size and cost and further increased speed and reliability of computers. By the late 1970s, many products such as video recorders contained dedicated computers called microcontrollers, and they started to appear as a replacement to mechanical controls in domestic appliances such as washing machines. The 1980s witnessed home computers and the now ubiquitous personal computer. With the evolution of the Internet, personal computers are becoming as common as the television and the telephone in the household.

Modern smartphones are fully programmable computers in their own right, and as of 2009 may well be the most common form of such computers in existenc.

历史的计算

主要文章:计算机硬件的历史

在第一次使用“计算机”这个词被记录在1613年,指的是对一个人进行了计算,或计算,与词的意思相同,直到继续20世纪中期。19世纪末开始其更熟悉的含义,一个机器进行计算。

Limited-function早期计算机

雅卡尔的展览上,在科学工业博物馆在曼彻斯特,是英国最早的一种可编程设备。

现代计算机的历史开始于两个独立的技术,自动计算和可编程,但没有一种单一的设备可以看成是最早的电脑,部分的原因是由于不应用那个术语。一些设备值得一提的是,虽然喜欢一些机械复制计算,很成功,并且存活了几个世纪,来临的电子计算器,像苏美尔算盘,大约在公元前2500年左右的设计赢得竞争的后代对现代书桌速度计算机器在1946年日本滑动规则,起源于1620年代,进行了五阿波罗太空机构,包括月球等机制,一个古老的天文计算机建造大约在公元前80年的希腊人。亚历山大的希腊数学家的英雄(公元前10 -公元70年)建造了一个机械剧院演了一出剧的持续10分钟的时间,是一个复杂的旋转系统操作的绳索和桶,可能会被认为是一种手段决定的哪个部位进行行为和机理的时候。这是可编程的本质。

十世纪末,法国僧侣带回来斯韦二氏西班牙画的机器发明的,要么摩尔人回答是或否的问题。在13世纪,僧侣们Albertus马格努斯和罗杰·培根说没有任何建造机器人的进一步发展。

1642年,文艺复兴时期的看见发明的机械计算器,一种装置能完成所有的四个算术运算,不用依赖人类的智力。机械计算器的根源是电脑的发展有两种方式。最初,是在试图发展更强大和更灵活的计算器,电脑是第一个被人们记得查尔斯·巴贝奇,然后理论发展。其次,开发的一种低成本的电子计算器继任者、机械计算器,造成了发展的第一个商业化可用英特尔处理器集成电路。

第一次通用计算机

1801年,约瑟夫玛丽做了一个改进提花织机介绍纺织一系列打孔纸质卡作为一个模板使织机编织自动错综复杂的模式。结果提花机是一个重要的发展历程中的电脑,因为使用穿孔卡片来定义编织模式可被看作是早,虽然有限,可编程的形式。

1837年,查尔斯·巴巴奇是第一个构思和设计一个完全可编程机械计算机,

他的分析引擎。有限的财政和巴巴奇未能抵抗摆弄设计意味着装置是根本没完成的;然而他的儿子,亨利·巴巴奇,完成了一个简化版本的分析引擎的运算单元(厂)于1888年。他给了一个成功的示范用于计算表于1906年。这台机器是给科学博物馆位于南肯辛顿于1910年。

在1880年代晚期,赫尔曼。霍尔瑞斯发明了记录的一种机器可以阅读的数据中。机器的使用媒体早些时候被控制,没有数据。“经过一段时间的试验,他住在纸带上穿孔卡片…“处理这些穿孔卡片他发明了针孔制表机。这三个发明是基础的现代信息加工业。大型自动化的数据处理进行了穿孔卡片为1890年美国人口普查由霍勒里斯的公司,后来成为了IBM的核心。19世纪末大量的新技术、新思路,后来被证明是有用的计算机实现的实践,已经开始出现:布尔代数,真空管(计算机阀),穿孔卡片和磁带,电传打字机。

在20世纪初,许多科学计算需要受到了越来越复杂的模拟计算机,它通过直接机械或电气模型为基础计算问题。然而,这不是编程,通常缺乏的通用性和准确性的现代数字计算机。

阿兰·图灵就被广泛认为是现代计算机科学之父。在1936年提供了一个有影响力的正规化涂林的概念和计算方法与涂林机器,提供电子数字计算机的蓝图。他的角色在创作现代计算机的时候,《时代》杂志在20世纪命名图灵的100位最具影响力的人物之一,声明:“事实是,每个人在一个键盘,打开水龙头电子数据表或一个字处理程序,正在涂林的化身”。

这Zuse虚拟的某种知识世界,1941年,被认为是世界上的第一个工作可编程、全自动计算机器。

EDSAC是其中第一个计算机来实现stored-program(冯。诺依曼)的建筑。

一个英特尔80486 DX2微处理器(实际尺寸:12×6.75毫米),在它的包装。 Atanasoff-Berry的计算机(ABC)是世界上最早的电子数字计算机,但不可编程的。阿坦那索夫被认为是其中一位电脑之父。它孕育于1937年由爱荷华州立大学物理学教授约翰?阿坦那索夫,修筑协助研究生克利夫?贝瑞,机器没有可编程的,只有解决系统在设计线性方程组。并采用计算机并行处理能力。1973年专利纠纷的法院判决书中发现的专利1946 ENIAC计算机来源于Atanasoff-Berry电脑。

第一个程控计算机发明的康拉德祖萨、谁建的虚拟的某种知识世界,机电一体化计算机器,在1941年。第一个可编程电子计算机是庞然大物,始建于1943年由汤米·花朵。

乔治是国际公认的父亲的现代数字计算机。在工作时,在贝尔实验室在1937年11月,Stibitz发明,搭建了一个基于继电器的计算器他称为“模型K”(“厨

房桌子”,但他聚集它),这是第一个使用二进制电路执行算术运算。最近的模型更大的复杂性增加包括复杂的算术和可编程。

一个接一个的稳定更强大和灵活的计算装置建于1940年代和1930年代,逐渐增加的主要特点是出现在现代的计算机。使用数码电子(主要是在1937年发明的克劳德夏侬)和更灵活的可编程是至关重要的步骤,但确定一个点沿着这条道路是“第一个电子数字计算机”是困难的。显著的成果包括。

康拉德·祖萨机电“Z机器”。这个虚拟的某种知识世界(1941)是第一个以二进制算术工作机,包括浮点运算和可编程的一个衡量标准。在1998年,虚拟的某种知识世界被证明是图灵完整,因此被世界上第一个操作电脑。

可编程的的计算机(开始于1937年,在1941年完成)采用真空管基础计算、二进制数字,和蓄热式电容器的记忆。使用蓄热式记忆允许他的作品被更紧凑的比它的同行出色(正在约的体积有一个大桌子或平台,因为中间结果可以存储,然后反馈到相同的一组计算元素。

英国的秘密"巨神像"电脑(1943年),有限但显示装置可编程使用成千上万的管可以合理可靠和电子系统中可编程门。它是用来破坏德国战时的代码。

哈佛马克一世(1944年),一个大型机电式计算机有限的可编程。

美国军队的弹道研究实验室ENIAC(1946),采用十进制算法和有时被称为第一次通用电子计算机(因为康拉德?祖萨的虚拟的某种知识世界1941使用电磁铁代替电子)。最初,然而,ENIAC有一个呆板的建筑本质要求改变其重新编程。

小规模的实验机的复制品(SSEM),世界上第一个可存储电脑,在科学工业博物馆在曼彻斯特,英格兰电子计分数字计算机的几个开发商,承认其缺点,想出了一个更加灵活和优雅的设计,这后来被称作“存储建筑”或冯·诺依曼结构。本设计是第一个正式的描述由约翰·冯·纽曼在纸的初稿EDVAC的一份报告显示,分布在1945年。许多项目建立计算机基于存储建筑开始在这个时候,其中第一本于1948年完工,在英国曼彻斯特大学的,这位小型实验机(SSEM或“宝贝”)。电子延迟存储自动计算器(EDSAC),完成后一年SSEM在剑桥大学,是第一个实用的、non-experimental设计和实施可存储计算机立即在投入使用为研究在本校工作。不久,这台机器原本冯。诺依曼的描述完成的,却没有看见paper-EDVAC-was 全职使用额外的两年。

几乎所有的现代计算机执行一些形式的stored-program架构,使之成为单一对方“计算机”这个词现在定义。而技术用于电脑有了显著的变化自从第一电子、通用计算机的1940年代,大部分仍然使用冯·诺依曼体系结构。

从1950年代开始,苏联科学家谢尔盖·萨辛进行研究,对三元Brusentsov 计算机、设备运行在一个基地,三个编码系统的?1,0,- 1,而不是传统的二进制编

码系统在大多数电脑的基础。他们设计了Setun,功能三元电脑,在莫斯科,州立大学。该设备已经投入有限公司生产在苏联,但取代二进制建筑更常见。

半导体和微处理器

计算机使用真空管为其在使用电子元件在整个1950年代,但到了1960年代已经很大程度地取代了半导体的电晶体的机械,这些机械是更小,更快,更便宜的生产,需要更少的电力,并且他们也更可在1970年代,集成电路技术和随后而来的微处理器,如英特尔4004,进一步减少体积和成本,进一步提高计算机的速度和可靠性。在1970年代晚期,许多产品,如录像机专用电脑中被称为微控制器,他们开始出现在更换到机械控制家电如洗衣机。在1980年代产生了家用电脑和现在到处都个人计算机。因特网的进化,个人电脑将成为一般的电视和电话的家务活动。

现代智能手机完全可编程计算机在他们自己的权力,在2009年很可能是最常见的计算机在手机等。

计算机专业外文文献及翻译

微软Visual Studio 1微软Visual Studio Visual Studio 是微软公司推出的开发环境,Visual Studio可以用来创建Windows平台下的Windows应用程序和网络应用程序,也可以用来创建网络服务、智能设备应用程序和Office 插件。Visual Studio是一个来自微软的集成开发环境IDE,它可以用来开发由微软视窗,视窗手机,Windows CE、.NET框架、.NET精简框架和微软的Silverlight支持的控制台和图形用户界面的应用程序以及Windows窗体应用程序,网站,Web应用程序和网络服务中的本地代码连同托管代码。 Visual Studio包含一个由智能感知和代码重构支持的代码编辑器。集成的调试工作既作为一个源代码级调试器又可以作为一台机器级调试器。其他内置工具包括一个窗体设计的GUI应用程序,网页设计师,类设计师,数据库架构设计师。它有几乎各个层面的插件增强功能,包括增加对支持源代码控制系统(如Subversion和Visual SourceSafe)并添加新的工具集设计和可视化编辑器,如特定于域的语言或用于其他方面的软件开发生命周期的工具(例如Team Foundation Server的客户端:团队资源管理器)。 Visual Studio支持不同的编程语言的服务方式的语言,它允许代码编辑器和调试器(在不同程度上)支持几乎所有的编程语言,提供了一个语言特定服务的存在。内置的语言中包括C/C + +中(通过Visual C++),https://www.360docs.net/doc/3919239579.html,(通过Visual https://www.360docs.net/doc/3919239579.html,),C#中(通过Visual C#)和F#(作为Visual Studio 2010),为支持其他语言,如M,Python,和Ruby等,可通过安装单独的语言服务。它也支持的 XML/XSLT,HTML/XHTML,JavaScript和CSS.为特定用户提供服务的Visual Studio也是存在的:微软Visual Basic,Visual J#、Visual C#和Visual C++。 微软提供了“直通车”的Visual Studio 2010组件的Visual Basic和Visual C#和Visual C + +,和Visual Web Developer版本,不需任何费用。Visual Studio 2010、2008年和2005专业版,以及Visual Studio 2005的特定语言版本(Visual Basic、C++、C#、J#),通过微软的下载DreamSpark计划,对学生免费。 2架构 Visual Studio不支持任何编程语言,解决方案或工具本质。相反,它允许插入各种功能。特定的功能是作为一个VS压缩包的代码。安装时,这个功能可以从服务器得到。IDE提供三项服务:SVsSolution,它提供了能够列举的项目和解决方案; SVsUIShell,它提供了窗口和用户界面功能(包括标签,工具栏和工具窗口)和SVsShell,它处理VS压缩包的注册。此外,IDE还可以负责协调和服务之间实现通信。所有的编辑器,设计器,项目类型和其他工具都是VS压缩包存在。Visual Studio 使用COM访问VSPackage。在Visual Studio SDK中还包括了管理软件包框架(MPF),这是一套管理的允许在写的CLI兼容的语言的任何围绕COM的接口。然而,MPF并不提供所有的Visual Studio COM 功能。

计算机专业外文文献翻译6

外文文献翻译(译成中文2000字左右): As research laboratories become more automated,new problems are arising for laboratory managers.Rarely does a laboratory purchase all of its automation from a single equipment vendor. As a result,managers are forced to spend money training their users on numerous different software packages while purchasing support contracts for each. This suggests a problem of scalability. In the ideal world,managers could use the same software package to control systems of any size; from single instruments such as pipettors or readers to large robotic systems with up to hundreds of instruments. If such a software package existed, managers would only have to train users on one platform and would be able to source software support from a single vendor. If automation software is written to be scalable, it must also be flexible. Having a platform that can control systems of any size is far less valuable if the end user cannot control every device type they need to use. Similarly, if the software cannot connect to the customer’s Laboratory Information Management System (LIMS) database,it is of limited usefulness. The ideal automation software platform must therefore have an open architecture to provide such connectivity. Two strong reasons to automate a laboratory are increased throughput and improved robustness. It does not make sense to purchase high-speed automation if the controlling software does not maximize throughput of the system. The ideal automation software, therefore, would make use of redundant devices in the system to increase throughput. For example, let us assume that a plate-reading step is the slowest task in a given method. It would make that if the system operator connected another identical reader into the system, the controller software should be able to use both readers, cutting the total throughput time of the reading step in half. While resource pooling provides a clear throughput advantage, it can also be used to make the system more robust. For example, if one of the two readers were to experience some sort of error, the controlling software should be smart enough to route all samples to the working reader without taking the entire system offline. Now that one embodiment of an ideal automation control platform has been described let us see how the use of C++ helps achieving this ideal possible. DISCUSSION C++: An Object-Oriented Language Developed in 1983 by BjarneStroustrup of Bell Labs,C++ helped propel the concept of object-oriented programming into the mainstream.The term ‘‘object-oriented programming language’’ is a familiar phrase that has been in use for decades. But what does it mean? And why is it relevant for automation software? Essentially, a language that is object-oriented provides three important programming mechanisms:

计算机系——外文翻译(中英对照-3000汉字左右)

毕业设计(论文)外文资料翻译 系别计算机信息与技术系 专业计算机科学与技术 班级 姓名 学号 外文出处 附件 1. 原文; 2. 译文 2012年3月

History of computing Main article: History of computing hardware The first use of the word "computer" was recorded in 1613, referring to a person who carried out calculations, or computations, and the word continued with the same meaning until the middle of the 20th century. From the end of the 19th century the word began to take on its more familiar meaning, a machine that carries out computations. Limited-function early computers The Jacquard loom, on display at the Museum of Science and Industry in Manchester, England, was one of the first programmable devices. The history of the modern computer begins with two separate technologies, automated calculation and programmability, but no single device can be identified as the earliest computer, partly because of the inconsistent application of that term. A few devices are worth mentioning though, like some mechanical aids to computing, which were very successful and survived for centuries until the advent of the electronic calculator, like the Sumerian abacus, designed around 2500 BC of which a descendant won a speed competition against a modern desk calculating machine in Japan in 1946, the slide rules, invented in the 1620s, which were carried on five Apollo space missions, including to the moon and arguably the astrolabe and the Antikythera mechanism, an ancient astronomical computer built by the Greeks around 80 BC. The Greek mathematician Hero of Alexandria (c. 10–70 AD) built a mechanical theater which performed a play lasting 10 minutes and was operated by a complex system of ropes and drums that might be considered to be a means of deciding which parts of the mechanism performed which actions and when. This is the essence of programmability. Around the end of the 10th century, the French monk Gerbert d'Aurillac brought back from Spain the drawings of a machine invented by the Moors that answered either Yes or No to the questions it was asked. Again in the 13th century, the monks Albertus Magnus and Roger Bacon built talking androids without any further development. In 1642, the Renaissance saw the invention of the mechanical calculator, a device that could perform all four arithmetic operations without relying on human intelligence. The mechanical calculator was at the root of the development of

计算机专业外文文献及翻译

计算机专业外文文献及翻译 微软Visual Studio 1微软Visual Studio 是微软公司推出的软软软境~可以用软建来平台下的 Visual Studio Visual StudioWindows 软用程序和软软用程序~也可以用软建软服软、智能软软软用程序和网来网 插件。WindowsOffice Visual 是一自微软的个来集成软软软境;,~可以用软软由它来微StudioIDEinteqrated development environment软软窗~软手机窗~、框架、精软架框和微软的支持的控制台和软Windows https://www.360docs.net/doc/3919239579.html,Silverlight 形用软界面的软用程序以及窗体软用程序~站网~软用程序和软服软网中的本地代软软同托管WindowsWeb 代软。 包含一由个智能感知和代软重构支持的代软软软器。集成的软软工作作软一源代软软软既个Visual Studio 软器又可以作软一台机器软软软器。其他置工具包括一软软内个窗体的软用程序~软软软软~软软软软~网数据软架GUI 构软软软。有乎各软面的件增强功能~包括增加软支持它几个插源代软控制系软;如和SubversionVisual ,添加新的工具集软软和可软化软软器~如并特定于域的软言或用于其他方面的软件软软生命周期SourceSafe 的工具;例如的客软端,软软软源管理器,。Team Foundation Server

支持不同的软程软言的服软方式的软言~允软代软软软器和软软器;在不同程 度上,支持它Visual Studio 几乎所有的软程软言~提供了一软言特定服软的存在。置的软言中包括个内中;通软C/C + +Visual C+,;通软,~,中;通软,,和,;作软+,https://www.360docs.net/doc/3919239579.html,Visual https://www.360docs.net/doc/3919239579.html,CVisual CFVisual Studio ,~软支持其他软言~如和等~可通软安软的软言服软。软也支持装独它的2010M,Python,Ruby 和软特定用软提供服软的也是存在的,微 XML/XSLT,HTML/XHTML ,JavaScriptCSS.Visual Studio软~,、,和。Visual BasicVisual JVisual CVisual C++ 微软提供了“直通软”的软件的和,和~Visual Studio 2010Visual BasicVisual CVisual C + +和版本~不需任何软用。、年和软软版~以及Visual Web DeveloperVisual Studio 201020082005 的特定软言版本;、、,、,,~通软微软的下软Visual Studio 2005Visual BasicC++CJ 软~软生免软。划学DreamSpark 2架构 不支持任何软程软言~解方案或工具本软。相反~允软入各软功能。特定的功决它插Visual Studio 能是作软一个软软包的代软。安软~软功能可以服软器得到。装个从提供三软服软,~VSIDESVsSolution它决提供了能软列软的软目和解方案~提供了口和用软界面功能;包括软软~工具软和工它窗; SVsUIShell 具口,和窗~软理它软软包的注。此外~册软可以软软软软和服软之软软软通信。所有的软软器~SVsShellVSIDE

计算机科学与技术专业外文翻译--数据库

外文原文: Database 1.1Database concept The database concept has evolved since the 1960s to ease increasing difficulties in designing, building, and maintaining complex information systems (typically with many concurrent end-users, and with a large amount of diverse data). It has evolved together with database management systems which enable the effective handling of databases. Though the terms database and DBMS define different entities, they are inseparable: a database's properties are determined by its supporting DBMS and vice-versa. The Oxford English dictionary cites[citation needed] a 1962 technical report as the first to use the term "data-base." With the progress in technology in the areas of processors, computer memory, computer storage and computer networks, the sizes, capabilities, and performance of databases and their respective DBMSs have grown in orders of magnitudes. For decades it has been unlikely that a complex information system can be built effectively without a proper database supported by a DBMS. The utilization of databases is now spread to such a wide degree that virtually every technology and product relies on databases and DBMSs for its development and commercialization, or even may have such embedded in it. Also, organizations and companies, from small to large, heavily depend on databases for their operations. No widely accepted exact definition exists for DBMS. However, a system needs to provide considerable functionality to qualify as a DBMS. Accordingly its supported data collection needs to meet respective usability requirements (broadly defined by the requirements below) to qualify as a database. Thus, a database and its supporting DBMS are defined here by a set of general requirements listed below. Virtually all existing mature DBMS products meet these requirements to a great extent, while less mature either meet them or converge to meet them. 1.2Evolution of database and DBMS technology The introduction of the term database coincided with the availability of direct-access storage (disks and drums) from the mid-1960s onwards. The term represented a contrast with the tape-based systems of the past, allowing shared interactive use rather than daily batch processing. In the earliest database systems, efficiency was perhaps the primary concern, but it was already recognized that there were other important objectives. One of the key aims was to make the data independent of the logic of application programs, so that the same data could be made available to different applications. The first generation of database systems were navigational,[2] applications typically accessed data by following pointers from one record to another. The two main data models at this time were the hierarchical model, epitomized by IBM's IMS system, and the Codasyl model (Network model), implemented in a number of

计算机科学与技术 外文翻译 英文文献 中英对照

附件1:外文资料翻译译文 大容量存储器 由于计算机主存储器的易失性和容量的限制, 大多数的计算机都有附加的称为大容量存储系统的存储设备, 包括有磁盘、CD 和磁带。相对于主存储器,大的容量储存系统的优点是易失性小,容量大,低成本, 并且在许多情况下, 为了归档的需要可以把储存介质从计算机上移开。 术语联机和脱机通常分别用于描述连接于和没有连接于计算机的设备。联机意味着,设备或信息已经与计算机连接,计算机不需要人的干预,脱机意味着设备或信息与机器相连前需要人的干预,或许需要将这个设备接通电源,或许包含有该信息的介质需要插到某机械装置里。 大量储存器系统的主要缺点是他们典型地需要机械的运动因此需要较多的时间,因为主存储器的所有工作都由电子器件实现。 1. 磁盘 今天,我们使用得最多的一种大量存储器是磁盘,在那里有薄的可以旋转的盘片,盘片上有磁介质以储存数据。盘片的上面和(或)下面安装有读/写磁头,当盘片旋转时,每个磁头都遍历一圈,它被叫作磁道,围绕着磁盘的上下两个表面。通过重新定位的读/写磁头,不同的同心圆磁道可以被访问。通常,一个磁盘存储系统由若干个安装在同一根轴上的盘片组成,盘片之间有足够的距离,使得磁头可以在盘片之间滑动。在一个磁盘中,所有的磁头是一起移动的。因此,当磁头移动到新的位置时,新的一组磁道可以存取了。每一组磁道称为一个柱面。 因为一个磁道能包含的信息可能比我们一次操作所需要得多,所以每个磁道划分成若干个弧区,称为扇区,记录在每个扇区上的信息是连续的二进制位串。传统的磁盘上每个磁道分为同样数目的扇区,而每个扇区也包含同样数目的二进制位。(所以,盘片中心的储存的二进制位的密度要比靠近盘片边缘的大)。 因此,一个磁盘存储器系统有许多个别的磁区, 每个扇区都可以作为独立的二进制位串存取,盘片表面上的磁道数目和每个磁道上的扇区数目对于不同的磁盘系统可能都不相同。磁区大小一般是不超过几个KB; 512 个字节或1024 个字节。

计算机科学外文翻译

Binomial heap In computer science, a binomial heap is a heap similar to a binary heap but also supports quick merging of two heaps. This is achieved by using a special tree structure. It is important as an implementation of the mergeable heap abstract data type(also called meldable heap), which is a priority queue supporting merge operation. Binomial tree A binomial heap is implemented as a collection of binomial trees (compare with a binary heap, which has a shape of a single binary tree). A binomial tree is defined recursively: ∙ A binomial tree of order 0 is a single node ∙ A binomial tree of order k has a root node whose children are roots of binomial trees of orders k−1, k−2, ..., 2, 1, 0 (in this order). Binomial trees of order 0 to 3: Each tree has a root node with subtrees of all lower ordered binomial trees, which have been highlighted. For example, the order 3 binomial tree is connected to an order 2, 1, and 0 (highlighted as blue, green and red respectively) binomial tree. A binomial tree of order k has 2k nodes, height k. Because of its unique structure, a binomial tree of order k can be constructed from two trees of order k−1 trivially by attaching one of them as the leftmost child of root of the

(整理)计算机_外文翻译_英文文献_中英版__仓库管理系统(_WMS_)

Warehouse Management Systems (WMS). The evolution of warehouse management systems (WMS) is very similar to that of many other software solutions. Initially a system to control movement and storage of materials within a warehouse, the role of WMS is expanding to including light manufacturing, transportation management, order management, and complete accounting systems. To use the grandfather of operations-related software, MRP, as a comparison, material requirements planning (MRP) started as a system for planning raw material requirements in a manufacturing environment. Soon MRP evolved into manufacturing resource planning (MRPII), which took the basic MRP system and added scheduling and capacity planning logic. Eventually MRPII evolved into enterprise resource planning (ERP), incorporating all the MRPII functionality with full financials and customer and vendor management functionality. Now, whether WMS evolving into a warehouse-focused ERP system is a good thing or not is up to debate. What is clear is that the expansion of the overlap in functionality between Warehouse Management Systems, Enterprise Resource Planning, Distribution Requirements Planning, Transportation Management Systems, Supply Chain Planning, Advanced Planning and Scheduling, and Manufacturing Execution Systems will only increase the level of confusion among companies looking for software solutions for their operations. Even though WMS continues to gain added functionality, the initial core functionality of a WMS has not really changed. The primary purpose of a WMS is to control the movement and storage of materials within an operation and process the associated transactions. Directed picking, directed replenishment, and directed put away are the key to WMS. The detailed setup and processing within a WMS can vary significantly from one software vendor to another, however the basic logic will use a combination of item, location, quantity, unit of measure, and order information to determine where to stock, where to pick, and in what sequence to perform these operations.

计算机专业中英文翻译外文翻译文献翻译

英文参考文献及翻译 Linux - Operating system of cybertimes Though for a lot of people , regard Linux as the main operating system to make up huge work station group, finish special effects of " Titanic " make , already can be regarded as and show talent fully. But for Linux, this only numerous news one of. Recently, the manufacturers concerned have announced that support the news of Linux to increase day by day, users' enthusiasm to Linux runs high unprecedentedly too. Then, Linux only have operating system not free more than on earth on 7 year this piece what glamour, get the favors of such numerous important software and hardware manufacturers as the masses of users and Orac le , Informix , HP , Sybase , Corel , Intel , Netscape , Dell ,etc. , OK? 1.The background of Linux and characteristic Linux is a kind of " free (Free ) software ": What is called free, mean users can obtain the procedure and source code freely , and can use them freely , including revise or copy etc.. It is a result of cybertimes, numerous technical staff finish its research and development together through Inte rnet, countless user is it test and except fault , can add user expansion function that oneself make conveniently to participate in. As the most outstanding one in free software, Linux has characteristic of the following: (1)Totally follow POSLX standard, expand the network operating system of supporting all AT&T and BSD Unix characteristic. Because of inheritting Unix outstanding design philosophy , and there are clean , stalwart , high-efficient and steady kernels, their all key codes are finished by Li nus Torvalds and other outstanding programmers, without any Unix code of AT&T or Berkeley, so Linu x is not Unix, but Linux and Unix are totally compatible. (2)Real many tasks, multi-user's system, the built-in network supports, can be with such seamless links as NetWare , Windows NT , OS/2 ,

计算机专业外文文献翻译

毕业设计(论文)外文文献翻译 (本科学生用) 题目:Plc based control system for the music fountain 学生姓名:_ ___学号:060108011117 学部(系): 信息学部 专业年级: _06自动化(1)班_ 指导教师: ___职称或学位:助教__ 20 年月日

外文文献翻译(译成中文1000字左右): 【主要阅读文献不少于5篇,译文后附注文献信息,包括:作者、书名(或论文题目)、出版社(或刊物名称)、出版时间(或刊号)、页码。提供所译外文资料附件(印刷类含封面、封 底、目录、翻译部分的复印件等,网站类的请附网址及原文】 英文节选原文: Central Processing Unit (CPU) is the brain of a PLC controller. CPU itself is usually one of the microcontrollers. Aforetime these were 8-bit microcontrollers such as 8051, and now these are 16-and 32-bit microcontrollers. Unspoken rule is that you’ll find mostly Hitachi and Fujicu microcontrollers in PLC controllers by Japanese makers, Siemens in European controllers, and Motorola microcontrollers in American ones. CPU also takes care of communication, interconnectedness among other parts of PLC controllers, program execution, memory operation, overseeing input and setting up of an output. PLC controllers have complex routines for memory checkup in order to ensure that PLC memory was not damaged (memory checkup is done for safety reasons).Generally speaking, CPU unit makes a great number of check-ups of the PLC controller itself so eventual errors would be discovered early. You can simply look at any PLC controller and see that there are several indicators in the form. of light diodes for error signalization. System memory (today mostly implemented in FLASH technology) is used by a PLC for a process control system. Aside form. this operating system it also contains a user program translated forma ladder diagram to a binary form. FLASH memory contents can be changed only in case where user program is being changed. PLC controllers were used earlier instead of PLASH memory and have had EPROM memory instead of FLASH memory which had to be erased with UV lamp and programmed on programmers. With the use of FLASH technology this process was greatly shortened. Reprogramming a program memory is done through a serial cable in a program for application development. User memory is divided into blocks having special functions. Some parts of a memory are used for storing input and output status. The real status of an input is stored either as “1”or as “0”in a specific memory bit/ each input or output has one corresponding bit in memory. Other parts of memory are used to store variable contents for variables used in used program. For example, time value, or counter value would be stored in this part of the memory. PLC controller can be reprogrammed through a computer (usual way), but also through manual programmers (consoles). This practically means that each PLC controller can programmed through a computer if you have the software needed for programming. Today’s transmission computers are ideal for reprogramming a PLC controller in factory itself. This is of great importance to industry. Once the system is corrected, it is also important to read the right program into a PLC again. It is also good to check from time to time whether program in a PLC has not changed. This helps to avoid hazardous situations in factory rooms (some automakers have established communication networks which regularly check programs in PLC controllers to ensure execution only of good programs). Almost every program for programming a PLC controller possesses various useful options such as: forced switching on and off of the system input/outputs (I/O lines),

相关文档
最新文档