云计算与物联网外文翻译文献

合集下载

物联网中英文对照外文翻译文献

物联网中英文对照外文翻译文献

中英文资料外文翻译Internet of Things1.the definition of connotationThe English name of the Internet of Things The Internet of Things, referred to as: the IOT.Internet of Things through the pass, radio frequency identification technology, global positioning system technology, real-time acquisition of any monitoring, connectivity, interactive objects or processes, collecting their sound, light, heat, electricity, mechanics, chemistry, biology, the location of a variety of the information you need network access through a variety of possible things and things, objects and people in the Pan-link intelligent perception of items and processes, identification and management. The Internet of Things IntelliSense recognition technology and pervasive computing, ubiquitous network integration application, known as the third wave of the world's information industry development following the computer, the Internet. Not so much the Internet of Things is a network, as Internet of Things services and applications, Internet of Things is also seen as Internet application development. Therefore, the application of innovation is the core of the development of Internet of Things, and 2.0 of the user experience as the core innovation is the soul of Things.2.The meaning of "material"Where the "objects" to meet the following conditions can be included in the scope of the "Internet of Things":1. Receiver have the appropriate information;2. Have a data transmission path;3. Have a certain storage capabilities;4. T o have the CPU;5.T o have the operating system;6. Have specialized applications;7. Have a data transmitter;8. Follow the communication protocol of Things;9. World Network, a unique number that can be identified.3. "Chinese style" as defined inInternet of Things (Internet of Things) refers to is the ubiquitous (Ubiquitous) terminal equipment (Devices) and facilities (Facilities), including with the "inner intelligence" sensors, mobile terminals, industrial systems, floor control system, the family of Intelligentfacilities, video surveillance systems, and external can "(Enabled), such as RFID, a variety of assets (the Assets), personal and vehicle carrying the wireless terminal" intelligent objects or animals "or" smart dust "(the Mote), through a variety of wireless and / or cable over long distances and / or short-range communication networks to achieve interoperability (M2M), application integration (the Grand Integration), and based on cloud computing, SaaS operation mode, in internal network (intranet), private network (e xtranet), and / or the Internet (Internet) environment, the use of appropriate information security mechanisms to provide a safe, controlled and even personalized real-time online monitoring, retrospective positioning, alarm linkage, command and control plan management, remote control, security, remote repair and maintenance, online upgrades, statistical reporting, decision support, the leadership of the desktop (showcase of the Cockpit Dashboard) management and service functions, "Everything," "efficient, energy saving, security environmental protection, "" possession, control, Camp integration [1].4.EU definitionIn September 2009, the Internet of Things and enterprise environments held in Beijing, China-EU Seminar on the European Commission and Social Media Division RFID Division is responsible for Dr. Lorent Ferderix, given the EU's definition of things: the Internet of Things is a dynamic global network infrastructure, it has a standards-based and interoperable communication protocols, self-organizing capabilities, including physical and virtual "objects" of identity, physical attributes, virtual features and smart interface and seamless integration of information networks . Internet of Things Internet and media, the Internet and business Internet one, constitute the future of the Internet.5.changeThe Internet of Things (Internet of Things) the word universally recognized at home and abroad Ashton, Professor of the MIT Auto-ID Center in 1999 first proposed to study RFID. The report of the same name released in 2005, the International T elecommunication Union (ITU), the definition and scope of the Internet of Things has been a change in the coverage of a larger expansion, no longer refers only to the Internet of Things based on RFID technology.Since August 2009, Premier Wen Jiabao put forward the "Experience China" Internet of Things was officially listed as a national one of the five emerging strategic industries, to write the "Government Work Report" Internet of Things in China has been the great concern of the society as a whole degree of concern is unparalleled in the United States, European Union, as well as other countries.The concept of Internet of Things is not so much a foreign concept, as it has been the concept of a "Made in China", his coverage of the times, has gone beyond the scope of the 1999 Ashton professor and the 2005 ITU report referred to, Internet of Things has been labeled a "Chinese style" label.6.BackgroundThe concept of Internet of Things in 1999. Internet-based, RFID technology and EPC standards, on the basis of the computer Internet, the use of radio frequency identification technology, wireless data communication technology, a global items of information to real-time sharing of the physical Internet "Internet of things" (referred to as the Internet of Things) , which is also the basis of the first round of the China Internet of Things boom set off in 2003.The sensor network is built up based on sensing technology network. Chinese Academy of Sciences in 1999 on the start sensor network research and has made some achievements in scientific research, the establishment of applicable sensor network.1999, held in the United States, mobile computing and networking International Conference, "The sensor network is a development opportunity facing humanity in the next century. In 2003, the United States, "T echnology Review" proposed sensor network technology will be future changes ten people's lives first.November 17, 2005, the WSIS held in Tunis (WSIS), the International T elecommunication Union released ITU Internet Report 2005: Internet of Things ", citing the concept of the" Internet of things ". The report pointed out that the ubiquitous "Internet o f Things" communication era is approaching, all the objects in the world, from tires to toothbrushes, from housing to the tissue via the Internet, take the initiative to be exchanged. Radio Frequency Identification (RFID), sensor technology, nanotechnology, intelligent embedded technology will be more widely used.According to the description of the ITU, the era of things, a short-range mobile transceivers embedded in a variety of daily necessities, human beings in the world of information and communication will receive a new communication dimension, from any time communication between people of the place of connection extended to the communication connection between persons and things and things and things. The Internet of Things concept of the rise, largely due to the International T elecommunication Union (ITU), the title of Internet of Things 2005 annual Internet Report. However, the ITU report the lack of a clear definition of Things.Domestic Internet of Things is also there is no single standard definition, but the Internet of Things In essence, the Internet of Things is a polymer application of modern information technology to a certain stage of development and technological upgrading of various sensing technology modern network technology and artificial intelligence and automation technology aggregation and integration of applications, so that the human and material wisdom of dialogue to create a world of wisdom. Because the development of the Internet of Things technology, involving almost all aspects of IT, innovative application and development of a polymer, systematic, and therefore be called revolutionary innovation of information industry. Summed up the nature of the Internet of Things is mainly reflected in three aspects: First, the Internet features that need to be networked objects must be able to achieve the interoperability of the Internet; identification and communication features, that is included in the Internet of Things "objects" must to have the functions of automatic identification and physical objects communication (M2M); intelligent features, the network system should have automated, self-feedback and intelligent control features January 28, 2009, Obama became the President of the United States, held with U.S.business leaders a "round table", as one of the only two representatives, IBM CEO Sam Palmisano for the first time that "the wisdom of the Earth" this concept, it is recommended that the new government to invest in a new generation of intelligent infrastructure.February 24, 2009 news, IBM Greater China CEO money crowd called "Smarter Planet" strategy announced in the forum 2009IBM.This concept was put forth, that is the great concern of the United States from all walks of life, and even analysts believe that IBM's vision is very likely to rise to U.S. national strategy, and caused a sensation in the world. IBM believes that the industry, the next phase of the mission is to make full use of the new generation of IT technology in all walks of life among specifically, is the embedded sensors and equipment to the power grid, railways, bridges, tunnels, highways, buildings, water supply systems dams, oil and gas pipelines and other objects, and is generally connected to the formation of Things.Strategy conference, IBM, and implant the concept of "wisdom" in the implementation of the infrastructure, strong, not only in the short term to stimulate the economy, promote employment, and in a short period of time for China to build a mature wisdom infrastructure platform.IBM "Smarter Planet" strategy will set off again after the wave of Internet technology industrial revolution. Former IBM CEO Lou Gerstner has raised an important point of view, every 15 years, a revolution in computing model. This judgment is the same as Moore's Law accurately call it a "15-year cycle Law". Before and after 1965, changes to the mainframe as a symbol, 1980 marked by the popularization of personal computers, 1995, the Internet revolution. Each such technological change are caused by the enterprise, industry and even the national competitive landscape of major upheaval and change. T o a certain extent in the Internet revolution is ripening by the "information superhighway" strategy. 1990s, the Clinton administration plan for 20 years, $ 200 billion to -4000 billion, construction of the U.S. National Information Infrastructure, to create a huge economic and social benefits.T oday, the "Smarter Planet" strategy by many Americans that there are many similarities with the "information superhighway", the same they revive the economy, a key strategy for competitive advantage. The strategy can be set off, not only for the UnitedStates, such as the Internet revolution was the wave of technological and economic concern, more attention from the world."Internet of Things prospects are very bright, it will dramatically change our current way of life." Demonstration director of the Center of Nanjing University of Aeronautics and Astronautics, National Electrical and Electronic Zhao Guoan said. Industry experts said that the Internet of things to our life personification of the things became a kind of human.Goods (goods) in the world of physical objects associated with each other "exchange", without the need for human intervention. The Internet of Things using radio frequency identification (RFID) technology, to achieve the interconnection and sharing of the automatic identification of goods (products) and information through the computer Internet. It can be said that the Internet of Things depict the world is full of intelligent. In the world of Internet of Things, material objects connected to the dragnet.The second session, held at Peking University in November 2008, China Mobile Government Seminar "Knowledge Society and Innovation 2.0", the experts made the mobile technology, the Internet of Things technology led to the development of economic and social form, innovative forms of change, and promote the The next generation of innovation for the knowledge society as the core of user experience (innovative 2.0) the formation of innovation and development of the form to pay more attention to the user to focus on people-oriented. Research institutions is expected to 10 years, the Internet of Things may be mass adoption of this technology will develop into one of thousands of yuan-scale high-tech market, the industry than the Internet 30 times.It is learned that the things industry chain can be broken down into the identity, perception, processing and information transfer, four links, each link of the key technologies for the wireless transmission network of RFID, sensors, smart chip and telecom operators. EPOSS in the "Internet of Things in 2020" report, an analysis predicted that the future development of the Internet of Things will go through four stages, 2010, RFID is widely used in the field of logistics, retail and pharmaceutical objects interconnect 2010 to 2015, 2015 ~ In 2020, the object into the semi-intelligent, intelligent objects into 2020.As the vanguard of the Internet of Things, RFID has become the most concerned about the technology market. The data show that the global RFID market size in 2008 from$ 4.93 billion in 2007 rose to $ 5.29 billion, this figure covers all aspects of the RFID market, including tags, readers and other infrastructure, software and services. RFID card and card-related infrastructure will account for 57.3 percent of the market, reaching $ 3.03 billion. Application from financial and security industries will drive the market growth of RFID cards. Analysys International forecasts, the Chinese RFID market size in 2009 will reach 5.0 billion, a CAGR of 33%, in which the electronic tag is more than 3.8 billion yuan, the reader close to 700 million yuan, software and services market to reach 500 million yuan pattern.MEMS is the abbreviation of the micro-electromechanical systems, MEMS technology is built on the basis of micro / nano, the market prospect is broad. The main advantage of the MEMS sensor is the small size, large-scale mass production cost reduction, mainly used in two major areas of automoti ve and consumer electronics. Under ICInsight the latest report is expected in 2007-2012, global sales of semiconductor sensors and actuators based on MEMS will reach 19 percent compound annual growth rate (CAGR), compared with $ 4.1 billion in 2007 to five years will achieve $ 9.7 billion in annual sales. 7.PrincipleInternet of Things is on the basis of the computer Internet, RFID, wireless data communications technology, to construct a cover everything in the world's "Internet of Things". In this network, the goods (products) to each other "exchange", without the need for human intervention. Its essence is the use of radio frequency identification (RFID) technology to achieve the interconnection and sharing of the automatic identification of goods (products) and information through the computer Internet.The Internet of Things is a very important technology is radio frequency identification (RFID) technology. RFID is radio frequency identification (Radio Frequency Identification) technology abbreviation, is an automatic identification technology in the 1990s began to rise, the more advanced a non-contact identification technology. The development of RFID technology based on a simple RFID system, combined with existing network technology, database technology, middleware technology, to build a one composed by a large number of networked readers and numerous mobile label, much larger than the Internet of Things trend.RFID, It is able to let items "speak" a technique. In the "Internet of Things" concept, RFID tags are stored in the specification and interoperability information collected automatically by wireless data communications network to a central information system, to achieve the identification of goods (products), and then through the open computer network for information exchange and sharing, items "transparent" management.The information technology revolution in the Internet of Things is referred to as IT mobile Pan of a specific application. Internet of Things through IntelliSense, identification technology and pervasive computing, ubiquitous network convergence applications, breaking the conventional thinking before, human beings can achieve ubiquitous computing and network connectivity [3]. The traditional thinking has been the separation of physical infrastructure and IT infrastructure: on the one hand, airports, roads, buildings, while on the other hand, the data center, PC, broadband. In the era of the "Internet of Things", reinforced concrete, cable with the chip, broadband integration into a unified infrastructure, in this sense, the infrastructure is more like a new site of the Earth, the world really works it, which including economic management, production operation, social and even personal life. "Internet of Things" makes it much more refined and dynamic management of production and life, to manage the future of the city to achieve the status of "wisdom" to improve resource utilization and productivity levels, and improve the relationship between man and nature.8.Agency1, institution-buildingAs the first national Internet of Things industry community organizations - the application of professional Committee of China Electronic Chamber of Things technology products (referred to as: "objects of the IPCC"), the Ministry of Civil Affairs in June 2010, preliminary approved by the Ministry of August being reported that the Ministry of Civil Affairs for final approval.2, the main taskServe as a bridge between business and government to assist the Government of the industry guidance, coordination, consultation and services to help members to reflect the business requirements to the Government; coordinate the relationship between enterprisesto strengthen technical cooperation, product distribution, the elimination of vicious competition ; supervision of members the correct implementation of national laws and regulations, to regulate the industry; member of information communication technology products, cooperation, resource sharing, capital operation, and promote the app lication of Internet of Things technologies and products, and promote the Internet of Things industrial scale , co-development.9.ConstructionInternet of Things in the practical application to carry out requires the involvement of all walks of life, and need the guidance of the national government as well as related regulations and policies to assist the launching of the Internet of Things has the scale, broad participation, management, technical, and material properties, etc. other features, the technical problem is the most crucial issues of Things billion Bo logistics consulting, Internet of Things technology is an integrated technology, a system not yet which company has overall responsibility for network planning and construction of the entire system, theoretical studies have commenced in all walks of life and the practical application is limited to within the industry. The key is on the planning and design and research and development of the Internet of Things research in the field of RFID, sensors, embedded software, and transmission of data calculation. In general, to carry out the steps of the Internet of things mainly as follows:(1) identified the object attributes, properties, including static and dynamic properties of the static property can be stored directly in the label, the dynamic properties need to start with sensors to detect real-time;(2) the need to identify the equipment to complete the reading of object attributes, and information into a data format suitable for network transmission;(3) the object of information transmitted over the network to the information processing center (processing center may be distributed, such as home computers or mobile phones, may also be centralized, such as China Mobile IDC) by the processing center to complete the object communication calculation.10.key areasInternet of Things 4 key areas:(1) RFID;(2) sensor network;(3) The M2M;(4) integration of the two.11.TrendIndustry experts believe that the Internet of things on the one hand can improve economic efficiency and significant cost savings; the other hand, can provide technical impetus to global economic recovery. Currently, the United States, the European Union are all invested heavily in-depth study to explore the Internet of Things. The country is also highly concerned about the emphasis of Things, Industry and Information T echnology Ministry in conjunction with the relevant departments are conducting research in a new generation of IT to the formation of policies and measures to support the development of a new generation of IT.China Mobile CEO Wang Jianzhou has repeatedly mentioned the Internet of Things will become the focus of future development of China Mobile. He will be invited to T aiwan to produce RFID, sensors and bar code manufacturers and China Mobile. According to him, the use of the Internet of Things technology, Shanghai Mobile has a number of industrial customers tailor the data collection, transmission, processing and business management in one set of wireless application solutions. The latest data show that Shanghai Mobile has more than 100,000 chips mounted on a taxi, bus, various forms of matter networking applications in all walks of prowess, to ensure the orderly operation of the city. During the Shanghai World Expo, "the bus services through" will be fully applied to the Shanghai public transport system, the smooth flow traffic to the most advanced technology to protect Expo area; for logistics transportation management, e-logistics ", will provide users with real-time accurate information of Cargo, vehicle tracking and positioning, the transport path selection, logistics network design and optimization services greatly enhance the comprehensive competitiveness of logistics enterprises.In addition, the popularization of the "Internet of Things" for the number of animals, plants and machinery, sensors and RFID tags of items and related interface devices will greatly exceed the number of mobile phones. The promotion of the Internet of Things willbecome a drive to promote economic development for the industry to open up a potential development opportunities. According to the current demand on the Internet of Things, in recent years, billions of sensors and electronic tags, which will greatly promote the production of IT components, while increasing the number of job opportunities.According to reports, it is necessary to truly build an effective Internet of things, there are two important factors. First, the scale, only with the scale to make the items of intelligence play a role. For example, a city of one million vehicles, if we only 10000 vehicles installed on the smart system, it is impossible to form an intelligent transportation system; two mobility items are usually not static, but in the state of the movement , we must maintain the items in the state of motion, and even high-speed motion state can at any time for dialogue.FORRESTER of the authority of the U.S. advisory body predicted that 2020, the world of business of the Internet of Things, compared with the business of interpersonal communication, will reach 30 to 1, so the "Internet of Things" is known to be the next one trillion communications services.Internet of Things heat wave Why is rapidly growing in China? Internet of Things in China rapid rise thanks to the several advantages of our country in terms of things.In the early 1999 launched the Internet of Things core sensor network technology research, R & D level in the world; the second, sensor network field in the world, China is the standard one of the dominant country, the patent owner; third China is one of the countries to achieve a complete industrial chain of Things; Fourth, China's wireless communications network and broadband coverage provides a solid infrastructure to support the development of the Internet of Things; Fifth, China has become the world's first the three major economies, with strong economic strength to support the development of the Internet of Things.12.MythThe current understanding of the Internet of things there are a lot of misunderstanding, which is also a direct impact on our understanding of Things on the development of the logistics industry, it is necessary first to distinguish errors, clarify our thinking.One sensor networks or RFID network equivalent of Things. The fact that sensortechnology, or RFID technology, or are simply one of the information collection technology. In addition to the sensor technology and RFID technology, GPS, video recognition, infrared, laser, scanning can be achieved automatically identify physical objects to communicate technical information collection technology can become the Internet of Things. Sensor networks or RFID network is just an application of Things, but not all of Things.Second, the Internet of Things as a myriad of unlimited extension of the Internet of Things as a completely open for all things, all of the interconnections, all shared Internet platform.In fact, the Internet of Things is not simple infinite extension of the global sharing of the Internet. Even if the Internet is also not only refers to we typically think of the international sharing computer network, Internet, WAN and LAN. Internet of Things can be both an extension of our usual sense of the Internet to the matter; LAN, professional can also be based on real needs and industrial applications. The reality is not necessary and can not make all the items networking; no need to make professional, LAN must be connected to the global Internet sharing platform. Of things in the future the Internet will be very different from the professional network of similar smart logistics, smart transportation, smart grid; the intelligence community and other local area network is the largest use of space.T er, that the ubiquitous network of the Internet of Things Internet of Things, and therefore the Internet of Things is a castle in the air, is difficult to achieve the technology. In fact the Internet of things are real, many of the primary Internet of Things applications already for our services. The Internet of Things concept is introduced in many real-world applications based on polymeric integrated innovation, pre-existing network with the Internet of Things, intelligent, automated system, summarized and upgrading it upgraded from a higher perspective our knowledge.Four of Things as a basket, and everything installed inside; based on self-awareness, and only be able to interact, communication products as the Internet of Things applications. For example, just embedded some of the sensors, to become the so-called Internet of Things appliances; products labeled with RFID tags, became the Internet of Things applications.es。

物联网工程中英文对照外文翻译文献

物联网工程中英文对照外文翻译文献

中英文对照外文翻译(文档含英文原文和中文翻译)Android: A Programmer’s Guide1 What Is Android1.1 Key Skills & Concepts● History of embedded device programming● Explanation of Open Handset Alliance● First look at the Android home screenIt can be said that, for a while, traditional desktop application developers have been spoiled. This is not to say that traditional desktop application development is easier than other forms of develop ment. However, as traditional desktop application developers, we have had the ability to create alm ost any kind of application we can imagine. I am including myself in this grouping because I got my start in desktop programming.One aspect that has made desktop programming more accessible is that we have had the ability to interact with the desktop operating system, and thus interact with any underlying hardware, pretty1freely (or at least with minimal exceptions). This kind of freedom to program independently, how ever, has never really been available to the small group of programmers who dared to venture int o the murky waters of cell phone development.NOTE :I refer to two different kinds of developers in this discussion: traditional desktop applicati on developers, who work in almost any language and whose end product, applications, are built to run on any “desktop” operating system; and Android developers, J ava developers who develop for the Android platform. This is not for the purposes of saying one is by any means better or wors e than the other. Rather, the distinction is made for purposes of comparing the development styles and tools of desktop operating system environments to the mobile operating system environment1.2 Brief History of Embedded Device ProgrammingFor a long time, cell phone developers comprised a small sect of a slightly larger group of developers known as embedded device developers. Seen as a less “glamorous” sibling to desktop—and later web—development, embedded device development typically got the proverbial short end of the stick as far as hardware and operating system features, because embedded device manufacturers were notoriously stingy on feature support.Embedded device manufacturers typically needed to guard their hardware secrets closely, so they gave embedded device developers few libraries to call when trying to interact with a specific device. Embedded devices differ fro m desktops in that an embedded device is typically a “computer on a chip.” For example, consider your standard television remote control; it is not really seen as an overwhelming achievement of technological complexity. When any button is pressed, a chip interprets the signal in a way that has been programmed into the device. This allows the device to know what to expect from the input device (key pad), and how to respond to those commands (for example, turn on the television). This is a simple form of embedded device programming. However, believe it or not, simple devices such as these are definitely related to the roots of early cell phone devices and development.Most embedded devices ran (and in some cases still run) proprietary operating systems. The reason for choosing to create a proprietary operating system rather than use any consumer system was really a product of necessity. Simple devices did not need very robust and optimized operating systems.As a product of device evolution, many of the more complex embedded devices, such as early PDAs, household security systems, and GPSs, moved to somewhat standardized operating system platforms about five years ago. Small-footprint operating systems such as Linux, or even an embedded version of Microsoft Windows, have become more prevalent on many embedded devices. Around this time in device evolution, cell phones branched from other embedded devices onto their own path. This branching is evident whenyou examine their architecture.Nearly since their inception, cell phones have been fringe devices insofar as they run on proprietary software—software that is owned and controlled by the manufacturer, and is almost always considered to be a “closed” system. The practice of manufacturers using proprietary operating systems began more out of necessity than any other reason. That is, cell phone manufacturers typically used hardware that was completely developed in-house, or at least hardware that was specifically developed for the purposes of running cell phone equipment. As a result, there were no openly available, off-the-shelf software packages or solutions that would reliably interact with their hardware. Since the manufacturers also wanted to guard very closely their hardware trade secrets, some of which could be revealed by allowing access to the software level of the device, the common practice was, and in most cases still is, to use completely proprietary and closed software to run their devices. The downside to this is that anyone who wanted to develop applications for cell phones needed to have intimate knowledge of the proprietary environment within which it was to run. The solution was to purchase expensive development tools directly from the manufacturer. This isolated many of the “homebrew” develo pers.NOTE:A growing culture of homebrew developers has embraced cell phone application development. The term “homebrew” refers to the fact that these developers typically do not work for a cell phone development company and generally produce small, one-off products on their own time.Another, more compelling “necessity” that kept cell phone development out of the hands of theeveryday developer was the hardware manufacturers’ solution to the “memory versus need” dilemma. Until recently, cell phones did little more than execute and receive phone calls, track your contacts, and possiblysend and receive short text messages; not really the “Swiss army knives” of technology they are today.Even as late as 2002, cell phones with cameras were not commonly found in the hands of consumers.By 1997, small applications such as calculators and games (Tetris, for example) crept their way ontocell phones, but the overwhelming function was still that of a phone dialer itself. Cell phones had not yetbecome the multiuse, multifunction personal tools they are today. No one yet saw the need for Internetbrowsing, MP3 playing, or any of the multitudes of functions we are accustomed to using today. It ispossible that the cell phone manufacturers of 1997 did not fully perceive the need consumers would havefor an all-in-one device. However, even if the need was present, a lack of device memory and storagecapacity was an even bigger obstacle to overcome. More people may have wanted their devices to be all-in-one tools, but manufacturers still had to climb the memory hurdle.To put the problem simply, it takes memory to store and run applications on any device, cell phones included. Cell phones, as a device, until recently did not have the amount of memory available to them thatwould facilitate the inclusion of “extra” programs. Within the last two years, the price of memory hasreached very low levels. Device manufacturers now have the ability to include more memory at lowerprices. Many cell phones now have more standard memory than the average PC had in the mid-1990s. So,now that we have the need, and the memory, we can all jump in and develop cool applications for cellphones around the world, right? Not exactly.Device manufacturers still closely guard the operating systems that run on their devices. While a fewhave opened up to the point where they will allow some Java-based applications to run within a smallenvironment on the phone, many do not allow this. Even the systems that do allow some Java apps to rundo not allow the kind of access to the “core” system that standard desktop developers are accustomed to having.1.3 Open Handset Alliance and AndroidThis barrier to application development began to crumble in November of 2007 when Google, under theOpen Handset Alliance, released Android. The Open Handset Alliance is a group of hardware and softwaredevelopers, including Google, NTT DoCoMo, Sprint Nextel, and HTC, whose goal is to create a more opencell phone environment. The first product to be released under the alliance is the mobile device operatingsystem, Android.With the release of Android, Google made available a host of development tools and tutorials to aid would-be developers onto the new system. Help files, the platform software development kit (SDK), and even a developers’ community can be found at Google’s Android website, This site should be your starting point, and I highly encourage you to visit the site.NOTE :Google, in promoting the new Android operating system, even went as far as to create a $10million contest looking for new and exciting Android applications.While cell phones running Linux, Windows, and even PalmOS are easy to find, as of this writing, nohardware platforms have been announced for Android to run on. HTC, LG Electronics, Motorola, andSamsung are members of the Open Handset Alliance, under which Android has been released, so we canonly hope that they have plans for a few Android-based devices in the near future. With its release inNovember 2007, the system itself is still in a software-only beta. This is good news for developers because it gives us a rare advance look at a future system and a chance to begin developing applications that willrun as soon as the hardware is released.NOTE:This strategy clearly gives the Open Handset Alliance a big advantage over other cell phone operating system developers, because there could be an uncountable number of applications available immediately for the first devices released to run Android.Introduction to AndroidAndroid, as a system, is a Java-based operating system that runs on the Linux 2.6 kernel. The system is very lightweight and full featured. Android applications are developed using Java and can be ported rather easily to the new platform. If you have not yet downloaded Java or are unsure about which version you need, I detail the installation of the development environment in Chapter 2. Other features of Android include an accelerated 3-D graphics engine (based on hardware support), database support powered by SQLite, and an integrated web browser.If you are familiar with Java programming or are an OOP developer of any sort, you are likely used to programmatic user interface (UI) development—that is, UI placement which is handled directly within the program code. Android, while recognizing and allowing for programmatic UI development, also supports the newer, XML-based UI layout. XML UI layout is a fairly new concept to the average desktop developer. I will cover both the XML UI layout and the programmatic UI development in the supporting chapters of this book.One of the more exciting and compelling features of Android is that, because of its architecture, third-partyapplications—including those that are “home grown”—are executed with the same system priority as those that are bundled with the core system. This is a major departure from most systems, which give embeddedsystem apps a greater execution priority than the thread priority available to apps created by third-partydevelopers. Also, each application is executed within its own thread using a very lightweight virtualmachine.Aside from the very generous SDK and the well-formed libraries that are available to us to develop with,the most exciting feature for Android developers is that we now have access to anything the operatingsystem has access to. In other words, if you want to create an application that dials the phone, you haveaccess to the phone’s dialer; if you want to create an application that utilizes the phone’s internal GPS (ifequipped), you have access to it. The potential for developers to create dynamic and intriguing applicationsis now wide open.On top of all the features that are available from the Android side of the equation, Google has thrown insome very tantalizing features of its own. Developers of Android applications will be able to tie their applications into existing Google offerings such as Google Maps and the omnipresent Google Search.Suppose you want to write an application that pulls up a Google map of where an incoming call isemanating from, or you want to be able to store common search results with your contacts; the doors ofpossibility have been flung wide open with Android.Chapter 2 begins your journey to Android development. You will learn the how’s and why’s of usingspecific development environments or integrated development environments (IDE), and you will downloadand install the Java IDE Eclipse.2 Application: Hello World2.1 Key Skills & Concepts● Creating new Android projects● Working with Views● Using a TextView● Modifying the main.xml file● Running applications on the Android EmulatorIn this chapter, you will be creating your first Android Activity. This chapter examines theapplication-building process from start to finish. I will show you how to create an Android project inEclipse, add code to the initial files, and run the finished application in the Android Emulator. The resultingapplication will be a fully functioning program running in an Android environment.Actually, as you move through this chapter, you will be creating more than one Android Activity.Computer programming tradition dictates that your first application be the typical Hello World! application,so in the first section you will create a standard Hello World! application with just a blank background andthe “Hello World!” text. Then, for the sake of enabling you to get to know the language better, the next section explains in detail the files automatically created by Android for your Hello World! application. You will create two iterations of this Activity, each using different techniques for displaying information to the screen. You will also create two different versions of a Hello World! application that will display an image that delivers the “Hello World!” message. This will give you a good introduction to the controls and inner workings of Android.NOTE:You will often see “application” and “Activity” used interchangeably. The difference between the two is that an application can be composed of multiple Activities, but one application must have at leastone Activity. Each “window” or screen of your application is a separate Activity. Therefore, if you create a fairly simple application with only one screen of data (like the Hello World! application in this chapter),that will be one Activity. In future chapters you will create applications with multiple Activities.To make sure that you get a good overall look at programming in Android, in Chapter 6 you will createboth of these applications in the Android SDK command-line environment for Microsoft Windows andLinux. In other words, this chapter covers the creation process in Eclipse, and Chapter 6 covers the creationprocess using the command-line tools. Therefore, before continuing, you should check that your Eclipseenvironment is correctly configured. Review the steps in Chapter 3 for setting the PATH statement for theAndroid SDK. You should also ensure that the JRE is correctly in your PATH statement.TIP:If you have configuration-related issues while attempting to work with any of the command-lineexamples, try referring to the configuration steps in Chapters 2 and 3; and look at the Android SDK documentation.2.2 Creating Your First Android Project in EclipseTo start your first Android project, open Eclipse. When you open Eclipse for the first time, it opens toan empty development environment (see Figure 5-1), which is where you want to begin. Your first task isto set up and name the workspace for your application. Choose File | New | Android Project, which willlaunch the New Android Project wizard.CAUTION Do not select Java Project from the New menu. While Android applications are written in Java, and you are doing all of your development in Java projects, this option will create a standard Java application. Selecting Android Project enables you to create Android-specific applications.If you do not see the option for Android Project, this indicates that the Android plugin for Eclipse was not fully or correctly installed. Review the procedure in Chapter 3 for installing the Android plugin for Eclipse to correct this.2.3 The New Android Project wizard creates two things for youA shell application that ties into the Android SDK, using the android.jar file, and ties the project intothe Android Emulator. This allows you to code using all of the Android libraries and packages, and alsolets you debug your applications in the proper environment.Your first shell files for the new project. These shell files contain some of the vital application blocksupon which you will be building your programs. In much the same way as creating a Microsoft .NETapplication in Visual Studio generates some Windows-created program code in your files, using the Android Project wizard in Eclipse generates your initial program files and some Android-created code. Inaddition, the New Android Project wizard contains a few options, shown next, that you must set to initiate your Android project. For the Project Name field, for purposes of this example, use the titleHelloWorldText. This name sufficiently distinguishes this Hello World! project from the others that youwill be creating in this chapter.In the Contents area, keep the default selections: the Create New Project inWorkspace radio button should be selected and the Use Default Location check box should be checked.This will allow Eclipse to create your project in your default workspace directory. The advantage ofkeeping the default options is that your projects are kept in a central location, which makes ordering,managing, and finding these projects quite easy. For example, if you are working in a Unix-basedenvironment, this path points to your $HOME directory.If you are working in a Microsoft Windows environment, the workspace path will beC:/Users/<username>/workspace, as shown in the previous illustration. However, for any number of reasons, you may want to uncheck the Use Default Location check box and select a different location for your project. One reason you may want to specify a different location here is simply if you want to choose a location for this specific project that is separate from other Android projects. For example, you may want to keep the projects that you create in this book in a different location from projects that you create in the future on your own. If so, simply override the Location option to specify your own custom location directory for this project.3 Application FundamentalsAndroid applications are written in the Java programming language. The compiled Java code — along with any data and resource files required by the application — is bundled by the aapt tool into an Androidpackage, an archive file marked by an .apk suffix. This file is the vehicle for distributing the application and installing it on mobile devices; it's the file users download to their devices. All the code in a single .apk file is considered to be one application.In many ways, each Android application lives in its own world:1. By default, every application runs in its own Linux process. Android starts the process when any of the application's code needs to be executed, and shuts down the process when it's no longer needed and system resources are required by other applications.2. Each process has its own virtual machine (VM), so application code runs in isolation from the code of all other applications.3. By default, each application is assigned a unique Linux user ID. Permissions are set so that the application's files are visible only to that user and only to the application itself — although there are ways to export them to other applications as well.It's possible to arrange for two applications to share the same user ID, in which case they will be able to see each other's files. To conserve system resources, applications with the same ID can also arrange to run in the same Linux process, sharing the same VM.3.1 Application ComponentsA central feature of Android is that one application can make use of elements of other applications (provided those applications permit it). For example, if your application needs to display a scrolling list of images and another application has developed a suitable scroller and made it available to others, you can call upon that scroller to do the work, rather than develop your own. Application have four types of components:(1)ActivitiesAn activity presents a visual user interface for one focused endeavor the user can undertake. For example, an activity might present a list of menu items users can choose from or it might display photographs along with their captions. A text messaging application might have one activity that shows a list of contacts to send messages to, a second activity to write the message to the chosen contact, and other activities to review old messages or change settings. Though they work together to form a cohesive user interface, each activity is independent of the others. Each one is implemented as a subclass of the Activity base class.An application might consist of just one activity or, like the text messaging application just mentioned, it may contain several. What the activities are, and how many there are depends, of course, on the application and its design. Typically, one of the activities is marked as the first one that should be presented to the user when the application is launched. Moving from one activity to another is accomplished by having the current activity start the next one.Each activity is given a default window to draw in. Typically, the window fills the screen, but it might be smaller than the screen and float on top of other windows. An activity can also make use of additional windows —— for example, a pop-up dialog that calls for a user response in the midst of the activity, or a windowswindow that presents users with vital information when they select a particular item on-screen.The visual content of the window is provided by a hierarchy of views — objects derived from the base View class. Each view controls a particular rectangular space within the window. Parent views contain and organize the layout of their children. Leaf views (those at the bottom of the hierarchy) draw in the rectangles they control and respond to user actions directed at that space. Thus, views are where the activity's interaction with the user takes place.For example, a view might display a small image and initiate an action when the user taps that image. Android has a number of ready-made views that you can use — including buttons, text fields, scroll bars, menu items, check boxes, and more.A view hierarchy is placed within an activity's window by the Activity.setContentView() method. The content view is the View object at the root of the hierarchy. (See the separate User Interface document for more information on views and the hierarchy.)(2)ServicesA service doesn't have a visual user interface, but rather runs in the background for an indefinite period of time. For example, a service might play background music as the user attends to other matters, or it might fetch data over the network or calculate something and provide the result to activities that need it. Each service extends the Service base class.A prime example is a media player playing songs from a play list. The player application would probably have one or more activities that allow the user to choose songs and start playing them. However, the musicplayback itself would not be handled by an activity because users will expect the music to keep playing even after they leave the player and begin something different. To keep the music going, the media player activity could start a service to run in the background. The system would then keep the music playback service running even after the activity that started it leaves the screen.It's possible to connect to (bind to) an ongoing service (and start the service if it's not already running). While connected, you can communicate with the service through an interface that the service exposes. For the music service, this interface might allow users to pause, rewind, stop, and restart the playback.Like activities and the other components, services run in the main thread of the application process. So that they won't block other components or the user interface, they often spawn another thread for time-consuming tasks (like music playback). See Processes and Threads, later.(3)Broadcast receiversA broadcast receiver is a component that does nothing but receive and react to broadcast announcements. Many broadcasts originate in system code — for example, announcements that the timezone has changed, that the battery is low, that a picture has been taken, or that the user changed a language preference. Applications can also initiate broadcasts — for example, to let other applications know that some data has been downloaded to the device and is available for them to use.An application can have any number of broadcast receivers to respond to any announcements it considers important. All receivers extend the BroadcastReceiver base class.Broadcast receivers do not display a user interface. However, they may start an activity in response to the information they receive, or they may use the NotificationManager to alert the user. Notifications can get the user's attention in various ways —— flashing the backlight, vibrating the device, playing a sound, and so the user's attention in various wayson. They typically place a persistent icon in the status bar, which users can open to get the message.(4)Content providersA content provider makes a specific set of the application's data available to other applications. The data can be stored in the file system, in an SQLite database, or in any other manner that makes sense. The content provider extends the ContentProvider base class to implement a standard set of methods that enable other applications to retrieve and store data of the type it controls. However, applications do not call these methods directly. Rather they use a ContentResolver object and call its methods instead. A ContentResolver can talk to any content provider; it cooperates with the provider to manage any interprocess communication that's involved.See the separate Content Providers document for more information on using content providers. Whenever there's a request that should be handled by a particular component, Android makes sure that the application process of the component is running, starting it if necessary, and that an appropriate instance of the component is available, creating the instance if necessary.3.2 Activating components: intentsContent providers are activated when they're targeted by a request from a ContentResolver. The other three components — activities, services, and broadcast receivers — are activated by asynchronous messages called intents. An intent is an Intent object that holds the content of the message. For activities and services, it names the action being requested and specifies the URI of the data to act on, among other things. For example, it might convey a request for an activity to present an image to the user or let the user edit some text. For broadcast receivers, theIntent object names the action being announced. For example, it might announce to interested parties that the camera button has been pressed.。

Hadoop云计算外文翻译文献

Hadoop云计算外文翻译文献

Hadoop云计算外文翻译文献(文档含中英文对照即英文原文和中文翻译)原文:Meet HadoopIn pioneer days they used oxen for heavy pulling, and when one ox couldn’t budge a log, they didn’t try to grow a larger ox. We shouldn’t be trying for bigger computers, but for more systems of computers.—Grace Hopper Data!We live in the data age. It’s not easy to measure the total volume of data stored electronically, but an IDC estimate put the size of the “digital universe” at 0.18 zettabytes in2006, and is forecasting a tenfold growth by 2011 to 1.8 zettabytes. A zettabyte is 1021 bytes, or equivalently one thousand exabytes, one million petabytes, or one billion terabytes. That’s roughly the same order of magnitude as one disk drive for every person in the world.This flood of data is coming from many sources. Consider the following:• The New York Stock Exchange generates about one terabyte of new trade data perday.• Facebook hosts approximately 10 billion photos, taking up one petabyte of storage.• , the genealogy site, stores around 2.5 petabytes of data.• The Internet Archive stores around 2 petabytes of data, and is growing at a rate of20 terabytes per month.• The Large Hadron Collider near Geneva, Switzerland, will produce about 15 petabytes of data per year.So there’s a lot of data out there. But you are probably wondering how it affects you.Most of the data is locked up in the largest web properties (like search engines), orscientific or financial institutions, isn’t it? Does the advent of “Big Data,” as it is being called, affect smaller organizations or individuals?I argue that it does. Take photos, for example. My wife’s grandfather was an avid photographer, and took photographs throughout his adult life. His entire corpus of medium format, slide, and 35mm film, when scanned in at high-resolution, occupies around 10 gigabytes. Compare this to the digital photos that my family took last year,which take up about 5 gigabytes of space. My family is producing photographic data at 35 times the rate my wife’s grandfather’s did, and the rate is increasing every year as it becomes easier to take more and more photos.More generally, the digital streams that individuals are producing are growing apace. Microsoft Research’s MyLifeBits project gives a glimpse of archiving of pe rsonal information that may become commonplace in the near future. MyLifeBits was an experiment where an individual’s interactions—phone calls, emails, documents were captured electronically and stored for later access. The data gathered included a photo taken every minute, which resulted in an overall data volume of one gigabyte a month. When storage costs come down enough to make it feasible to store continuous audio and video, the data volume for a future MyLifeBits service will be many times that.The t rend is for every individual’s data footprint to grow, but perhaps more importantly the amount of data generated by machines will be even greater than that generated by people. Machine logs, RFID readers, sensor networks, vehicle GPS traces, retail transactions—all of these contribute to the growing mountain of data.The volume of data being made publicly available increases every year too. Organizations no longer have to merely manage their own data: success in the future will be dictated to a large extent by their ability to extract value from other organizations’ data.Initiatives such as Public Data Sets on Amazon Web Services, , and exist to foster the “information commons,” where data can be freely (or in the case of AWS, for a modest price) shared for anyone to download and analyze. Mashups between different information sources make for unexpected and hitherto unimaginable applications.Take, for example, the project, which watches the Astrometry groupon Flickr for new photos of the night sky. It analyzes each image, and identifies which part of the sky it is from, and any interesting celestial bodies, such as stars or galaxies. Although it’s still a new and experimental service, it shows the kind of things that are possible when data (in this case, tagged photographic images) is made available andused for something (image analysis) that was not anticipated by the creator.It has been said that “More data usually beats better algorithms,” which is to say that for some problems (such as recommending movies or music based on past preferences),however fiendish your algorithms are, they can often be beaten simply by having more data (and a less sophisticated algorithm).The good news is that Big Data is here. The bad news is that we are struggling to store and analyze it.Data Storage and AnalysisThe problem is simple: while the storage capacities of hard drives have increased massively over the years, access speeds--the rate at which data can be read from drives--have not kept up. One typical drive from 1990 could store 1370 MB of data and had a transfer speed of 4.4 MB/s, so you could read all the data from a full drive in around five minutes. Almost 20years later one terabyte drives are the norm, but the transfer speed is around 100 MB/s, so it takes more than two and a half hours to read all the data off the disk.This is a long time to read all data on a single drive and writing is even slower. The obvious way to reduce the time is to read from multiple disks at once. Imagine if we had 100 drives, each holding one hundredth of the data. Working in parallel, we could read the data in under two minutes.Only using one hundredth of a disk may seem wasteful. But we can store one hundred datasets, each of which is one terabyte, and provide shared access to them. We can imagine that the users of such a system would be happy to share access in return for shorter analysis times, and, statistically, that their analysis jobs would be likely to be spread over time, so they wouldn`t interfere with each other too much.There`s more to being able to read and write data in parallel to or from multiple disks, though. The first problem to solve is hardware failure: as soon as you start using many pieces of hardware, the chance that one will fail is fairly high. A common way of avoiding data loss is through replication: redundant copies of the data are kept by the system so that in the event of failure, there is another copy available. This is how RAID works, for instance, although Hadoop`s filesystem, the Hadoop Distributed Filesystem (HDFS),takes a slightly different approach, as you shall see later. The second problem is that most analysis tasks need to be able to combine the data in some way; data read from one disk may need to be combined with the data from any of the other 99 disks. Various distributed systems allow data to be combined from multiple sources, but doing this correctly is notoriously challenging. MapReduce provides a programming model that abstracts the problem from disk reads and writes, transforming it into a computation over sets of keys and values. We will look at the details of this model in later chapters, but the important point for the present discussion is that there are two parts to the computation, the map a nd the reduce, and it’s the interface between the two where the “mixing” occurs. Like HDFS, MapReduce has reliability built-in.This, in a nutshell, is what Hadoop provides: a reliable shared storage and analysis system. The storage is provided by HDFS, and analysis by MapReduce. There are other parts to Hadoop, but these capabilities are its kernel.Comparison with Other SystemsThe approach taken by MapReduce may seem like a brute-force approach. The premise is that the entire dataset—or at least a good portion of it—is processed for each query. But this is its power. MapReduce is a batch query processor, and the ability to run an ad hoc query against your whole dataset and get the results in a reasonable time is transformative. It changes the way you think about data, and unlocks data that was previously archived on tape or disk. It gives people the opportunity to innovate with data. Questions that took too long to get answered before can now be answered, which in turn leads to new questions and new insights.For example, Mailtrust, Rackspace’s mail division, used Hadoop for processing email logs. One ad hoc query they wrote was to find the geographic distribution of their users.In their words: This data was so useful that we’ve scheduled the MapReduce job to run monthly and we will be using this data to help us decide which Rackspace data centers to place new mail servers in as we grow. By bringing several hundred gigabytes of data together and having the tools to analyze it, the Rackspace engineers were able to gain an understanding of the data that they otherwise would never have had, and, furthermore, they were able to use what they had learned to improve the service for their customers. You can read more about how Rackspace uses Hadoop in Chapter 14.RDBMSWhy can’t we use databases with lots of disks to do large-scale batch analysis? Why is MapReduce needed? The answer to these questions comes from another trend in disk drives: seek time is improving more slowly than transfer rate. Seeking is the process of moving the disk’s head to a particular place on the disk to read or write data. It characterizes the latency of a disk operation, whereas the transfer rate corresponds to a disk’s bandwidth.If the data access pattern is dominated by seeks, it will take longer to read or write large portions of the dataset than streaming through it, which operates at the transfer rate. On the other hand, for updating a small proportion of records in a database, a traditional B-Tree (the data structure used in relational databases, which is limited by the rate it can perform seeks) works well. For updating the majority of a database, a B-Tree is less efficient than MapReduce, which uses Sort/Merge to rebuild the database.In many ways, MapReduce can be seen as a complement to an RDBMS. (The differences between the two systems are shown in Table 1-1.) MapReduce is a good fit for problems thatneed to analyze the whole dataset, in a batch fashion, particularly for ad hoc analysis. An RDBMS is good for point queries or updates, where the dataset has been indexed to deliver low-latency retrieval and update times of a relatively small amount of data. MapReduce suits applications where the data is written once, and read many times, whereas a relational database is good for datasets that are continually updated.Table 1-1. RDBMS compared to MapReduceTraditional RDBMS MapReduceData size Gigabytes PetabytesAccess Interactive and batch BatchWrite once, read many times Updates Read and write manytimesStructure Static schema Dynamic schemaIntegrity High LowScaling Nonlinear LinearAnother difference between MapReduce and an RDBMS is the amount of structure in the datasets that they operate on. Structured data is data that is organized into entities that have a defined format, such as XML documents or database tables that conform to a particular predefined schema. This is the realm of the RDBMS. Semi-structured data, on the other hand, is looser, and though there may be a schema, it is often ignored, so it may be used only as a guide to the structure of the data: for example, a spreadsheet, in which the structure is the grid of cells, although the cells themselves may hold anyform of data. Unstructured data does not have any particular internal structure: for example, plain text or image data. MapReduce works well on unstructured or semistructured data, since it is designed to interpret the data at processing time. In other words, the input keys and values for MapReduce are not an intrinsic property of the data, but they are chosen by the person analyzing the data.Relational data is often normalized to retain its integrity, and remove redundancy. Normalization poses problems for MapReduce, since it makes reading a record a nonlocaloperation, and one of the central assumptions that MapReduce makes is that it is possible to perform (high-speed) streaming reads and writes.A web server log is a good example of a set of records that is not normalized (for example, the client hostnames are specified in full each time, even though the same client may appear many times), and this is one reason that logfiles of all kinds are particularly well-suited to analysis with MapReduce.MapReduce is a linearly scalable programming model. The programmer writes two functions—a map function and a reduce function—each of which defines a mapping from one set of key-value pairs to another. These functions are oblivious to the size of the data or the cluster that they are operating on, so they can be used unchanged for a small dataset and for a massive one. More importantly, if you double the size of the input data, a job will run twice as slow. But if you also double the size of the cluster, a job will run as fast as the original one. This is not generally true of SQL queries.Over time, however, the differences between relational databases and MapReduce systems are likely to blur. Both as relational databases start incorporating some of the ideas from MapReduce (such as Aster Data’s and Greenplum’s databases), and, from the other direction, as higher-level query languages built on MapReduce (such as Pig and Hive) make MapReduce systems more approachable to traditional database programmers.Grid ComputingThe High Performance Computing (HPC) and Grid Computing communities have been doing large-scale data processing for years, using such APIs as Message Passing Interface (MPI). Broadly, the approach in HPC is to distribute the work across a cluster of machines, which access a shared filesystem, hosted by a SAN. This works well for predominantly compute-intensive jobs, but becomes a problem when nodes need to access larger data volumes (hundreds of gigabytes, the point at which MapReduce really starts to shine), since the network bandwidth is the bottleneck, and compute nodes become idle.MapReduce tries to colocate the data with the compute node, so data access is fast since it is local. This feature, known as data locality, is at the heart of MapReduce and is the reason for its good performance. Recognizing that network bandwidth is the most precious resource in a data center environment (it is easy to saturate network links by copying data around),MapReduce implementations go to great lengths to preserve it by explicitly modelling network topology. Notice that this arrangement does not preclude high-CPU analyses in MapReduce.MPI gives great control to the programmer, but requires that he or she explicitly handle the mechanics of the data flow, exposed via low-level C routines and constructs, such as sockets, as well as the higher-level algorithm for the analysis. MapReduce operates only at the higher level: the programmer thinks in terms of functions of key and value pairs, and the data flow is implicit.Coordinating the processes in a large-scale distributed computation is a challenge. The hardest aspect is gracefully handling partial failure—when you don’t know if a remote process has failed or not—and still making progress with the overall computation. MapReduce spares the programmer from having to think about failure, since the implementation detects failed map or reduce tasks and reschedules replacements on machines that are healthy. MapReduce is able to do this since it is a shared-nothing architecture, meaning that tasks have no dependence on one other. (This is a slight oversimplification, since the output from mappers is fed to the reducers, but this is under the control of the MapReduce system; in this case, it needs to take more care rerunning a failed reducer than rerunning a failed map, since it has to make sure it can retrieve the necessary map outputs, and if not, regenerate them by running the relevant maps again.) So from the programmer’s point of view, the order in which the tasks run doesn’t matter. By contrast, MPI programs have to explicitly manage their own checkpointing and recovery, which gives more control to the programmer, but makes them more difficult to write.MapReduce might sound like quite a restrictive programming model, and in a sense itis: you are limited to key and value types that are related in specified ways, and mappers and reducers run with very limited coordination between one another (the mappers pass keys and values to reducers). A natural question to ask is: can you do anything useful or nontrivial with it?The answer is yes. MapReduce was invented by engineers at Google as a system for building production search indexes because they found themselves solving the same problem over and over again (and MapReduce was inspired by older ideas from the functional programming, distributed computing, and database communities), but it has since been used for many other applications in many other industries. It is pleasantly surprising to see the range of algorithms that can be expressed in MapReduce, from image analysis, to graph-based problems,to machine learning algorithms. It can’t solve every problem, of course, but it is a general data-processing tool.You can see a sample of some of the applications that Hadoop has been used for in Chapter 14.Volunteer ComputingWhen people first hear about Hadoop and MapReduce, they oft en ask, “How is it different from SETI@home?” SETI, the Search for Extra-Terrestrial Intelligence, runs a project called SETI@home in which volunteers donate CPU time from their otherwise idle computers to analyze radio telescope data for signs of intelligent life outside earth. SETI@home is the most well-known of many volunteer computing projects; others include the Great Internet Mersenne Prime Search (to search for large prime numbers) and Folding@home (to understand protein folding, and how it relates to disease).Volunteer computing projects work by breaking the problem they are trying to solve into chunks called work units, which are sent to computers around the world to be analyzed. For example, a SETI@home work unit is about 0.35 MB of radio telescope data, and takes hours or days to analyze on a typical home computer. When the analysis is completed, the results are sent back to the server, and the client gets another work unit. As a precaution to combat cheating, each work unit is sent to three different machines, and needs at least two results to agree to be accepted.Although SETI@home may be superficially similar to MapReduce (breaking a problem into independent pieces to be worked on in parallel), there are some significant differences. The SETI@home problem is very CPU-intensive, which makes it suitable for running on hundreds of thousands of computers across the world, since the time to transfer the work unit is dwarfed by the time to run the computation on it. Volunteers are donating CPU cycles, not bandwidth.MapReduce is designed to run jobs that last minutes or hours on trusted, dedicated hardware running in a single data center with very high aggregate bandwidth interconnects. By contrast, SETI@home runs a perpetual computation on untrusted machines on the Internet with highly variable connection speeds and no data locality.译文:初识Hadoop古时候,人们用牛来拉重物,当一头牛拉不动一根圆木的时候,他们不曾想过培育个头更大的牛。

云计算研究现状文献综述及外文文献

云计算研究现状文献综述及外文文献

本文档包括该专题的:外文文献、文献综述文献标题:An exploratory study on factors affecting the adoption of cloud computing by information professionals作者:Aharony, Noa期刊:The Electronic Library, 33(2), 308-328.年份:2015一、外文文献An exploratory study on factors affecting the adoption of cloud computing byinformation professionals(影响云计算采用与否的一个探索性研究)Aharony, NoaPurpose- The purpose of this study explores what factors may influence information professionals to adopt new technologies, such as cloud computing in their organizations. The objectives of this study are as follows: to what extent does the technology acceptance model (TAM) explain information professionals intentions towards cloud computing, and to what extent do personal characteristics, such as cognitive appraisal and openness to experience, explain information professionals intentions to use cloud computing.Design/methodology/approach- The research was conducted in Israel during the second semester of the 2013 academic year and encompassed two groups of information professionals: librarians and information specialists. Researchers used seven questionnaires to gather the following data: personal details, computer competence, attitudes to cloud computing, behavioral intention, openness to experience, cognitive appraisal and self-efficacy. Findings- The current study found that the behavioral intention to use cloud computing was impacted by several of the TAM variables, personal characteristics and computer competence.Originality/value- The study expands the scope of research about the TAM by applying it to information professionals and cloud computing and highlights the importance of individual traits, such as cognitive appraisal, personal innovativeness, openness to experience and computer competence when considering technology acceptance. Further, the current study proposes that if directors of information organizations assume that novel technologies may improve their organizations' functioning, they should be familiar with both the TAM and the issue of individual differences. These factors may help them choose the most appropriate workers.Keywords: Keywords Cloud computing, TAM, Cognitive appraisal, Information professionals, Openness to experienceIntroductionOne of the innovations that information technology (IT) has recently presented is thephenomenon of cloud computing. Cloud computing is the result of advancements in various technologies, including the Internet, hardware, systems management and distributed computing (Buyya et al. , 2011). Armbrust et al. (2009) suggested that cloud computing is a collection of applications using hardware and software systems to deliver services to end users via the Internet. Cloud computing offers a variety of services, such as storage and different modes of use (Leavitt, 2009). Cloud computing enables organizations to deliver support applications and avoid the need to develop their own IT systems (Feuerlicht et al. , 2010).Due to the growth of cloud computing use, the question arises as to what factors may influence information professionals to adopt new technologies, such as cloud computing, in their organizations. Assuming that using new technologies may improve the functioning of information organizations, this study seeks to explore if information professionals, who often work with technology and use it as an important vehicle in their workplace, are familiar with technological innovations and whether they are ready to use them in their workplaces. As the phenomenon of cloud computing is relatively new, there are not many surveys that focus on it and, furthermore, no one has so far focussed on the attitudes of information professionals towards cloud computing. The research may contribute to an understanding of the variables that influence attitudes towards cloud computing and may lead to further inquiry in this field.The current study uses the well-known technology acceptance model (TAM), a theory for explaining individuals' behaviours towards technology (Davis, 1989; Venkatesh, 2000), as well as personal characteristics, such as cognitive appraisal and openness to new experiences, as theoretical bases from which we can predict factors which may influence information professionals adopting cloud computing in their workplaces. The objectives of this study are to learn the following: the extent to which the TAM explains information professionals' attitudes towards cloud computing, and the extent to which personal characteristics, such as cognitive appraisal and openness to experiences, explain the intention of information professionals to use cloud computing.Theoretical backgroundCloud computingResearchers have divided cloud computing into three layers: Software as a Service (SaaS), Platform as a Service (PaaS) and Infrastructure as a Service (IaaS). SaaS has changed the concept of software as a product to that of a service instead. The software runs in the cloud and the user can access it via the Internet to work on an application. PaaS enables powerful tools for developers to create the applications, without having to deal with concerns about the infrastructure. IaaS provides complete infrastructure resources (e.g. servers, software, network equipment and storage). With IaaS, consumers do not have to purchase the latest technology, perform maintenance, upgrade software or buy software licenses (Anuar et al. , 2013). Cloud computing deployment can be divided into four types: private clouds, public clouds, community clouds and hybrid clouds (Mell and Grance, 2011). Public clouds have open access, private clouds run within organizations, community clouds containresources that are shared with others in the community and hybridclouds encompass two or more cloud models. Anuar et al. (2013) presented the main characteristics of cloud computing: flexible scale that enables flexible-scale capabilities for computing; virtualization that offers a new way of getting computing resources remotely, regardless of the location of the user or the resources; high trust , as the cloud offers more reliability to end users than relying on local resources; versatility , because cloud services can serve different sectors in various disciplines use the same cloud; and on demand service , as end users can tailor their service needs and pay accordingly.As cloud computing is relatively new, there are not a lot of surveys that focus on it. Several researchers conducted in-depth interviews investigating respondents' attitudes towards keeping their virtual possessions in the online world (Odom et al. , 2012). Teneyuca (2011) reported on a survey of cloud computing usage trends that included IT professionals as respondents. Results revealed preferences for virtualization and cloud computing technologies. However, the major reasons for cloud computing adoption being impeded were the lack of cloud computing training (43 per cent) and security concerns (36 per cent). Another report showed that nearly 40 per cent of Americans think that saving data to their hard drive is more secure than saving it to a cloud (Teneyuca, 2011). A further study (Ion et al., 2011) explored private users' privacy attitudes and beliefs about cloud computing in comparison with those in companies. Anuar et al. (2013) investigated cloud computing in an academic institution, claiming that cloud computing technology enhances performance within the academic institution. A study that was carried out in the education arena examined factors that led students to adopt cloud computing technology (Behrend et al. , 2010). Technology acceptance modelThe TAM (Davis, 1989) is a socio-technical model which aims to explain user acceptance of an information system. It is based on the theory of reasoned action (TRA) (Fishbein and Ajzen, 1975) which seeks to understand how people construct behaviours. The model suggests that technology acceptance can be explained according to the individual's beliefs, attitudes and intentions (Davis, 1989). The TAM hypothesizes that one's intention is the best predictor of usage behaviour and suggests that an individual's behavioural intention to use technology is determined by two beliefs: perceived usefulness (PU) and perceived ease of use (PEOU). PU refers to the individual's perception that using a technology will improve performance and PEOU addresses a user's perceptions that using a particular system would be free of effort (Davis, 1989). The current study concentrates on PEOU as the researchers wanted to examine if information professionals' perceptions about new technology is affected by its simplicity and friendly interface. Earlier research mainly investigated personal behaviour to use new information systems and technology in the following: corporate environments (Gefen and Straub, 1997);Web shopping (Chang et al. , 2002; Lin and Lu, 2000);education, particularly e-learning (Park, 2009) and m-learning (Aharony, 2014); and the library arena (Aharony, 2011; Park et al. , 2009).Personal innovativenessA construct which may contribute to information professionals' intention behaviour to use cloud computing is personal innovativeness, a major characteristic in innovation diffusion research in general (Agarwal and Prasad, 1998; Rogers, 1983, 1995). Agarwal and Prasad (1998) have coined the term "personal innovativeness in the domain of IT" (PIIT), which describes a quite stable characteristic of the individual across situational considerations. Previous studies found that personal innovativeness is a significant determinant of PEOU, as well as of PU (Agarwal and Karahanna, 2000; Lewis et al. , 2003). Several researchers have suggested that innovative people will search for intellectually or sensorially stimulating experiences (Uray and Dedeoglu, 1997).Openness to experienceAnother variable that may predict respondents' perspectives towards cloud computing is openness to experience which addresses the tendency to search for new and challenging experiences, to think creatively and to enjoy intellectual inquiries (McCrae and Sutin, 2009). People who are highly open to experience are perceived as also open to new challenges, thoughts and emotions (McCrae and Costa, 2003). Studies reported that there is a positive relation between openness to experience and intelligence tests (Gignac et al. , 2004). According to Weiss et al. (2012), challenging transitions may influence differently those who are high or low in openness to experience. Those who are high may approach these situations with curiosity, emphasizing the new possibilities offered to them. However, those who are low in openness may be threatened and try to avoid them by adhering to predictable environments. Various researchers note that people who are high in openness to experience are motivated to resolve new situations (McCrae, 1996; Sorrentino and Roney, 1999). Furthermore, openness to experience is associated with cognitive flexibility and open-mindedness (McCrae and Costa, 1997), and negatively associated with rigidity, uncertainty and inflexibility (Hodson and Sorrentino, 1999). Thus, people who are less open to experience tend to avoid novelty and prefer certainty. Studies reveal that openness to experience declines in the later years (Allemand et al. , 2007; Donnellan and Lucas, 2008).Challenge and threatThe following section will focus on the personality characteristics of challenge and threat that might affect information professionals' behavioural intention to use cloud computing. Challenge and threat are the main variables of a unidimensional, bipolar motivational state. They are the result of relative evaluations of situational demands and personal resources that are influenced both by cognitive and affective processes in motivated performance situations (Vick et al. , 2008). According to Lazarus and Folkman (1984), challenge refers to the potential for growth or gain and is characterized by excitement and eagerness, while threat addresses potential harm and is characterized by anxiety, fear and anger. Situations that suggest low demands and high resources are described as challenging, while those that suggest high demands and low resources are perceived as threatening (Seginer, 2008). In general, challenge or threat can take place in situations such as delivering a speech, taking a test, sports competitions or performing with another person on a cooperative or competitive task.The challenge appraisal suggests that with effort, the demands of the situation can be overcome (Lazarus et al. , 1980; Park and Folkman, 1997). On the other hand, threat appraisal indicates potential danger to one's well-being or self-esteem (Lazarus, 1991; Lazarus and Folkman, 1984), as well as low confidence in one's ability to cope with the threat (Bandura, 1997; Lazarus, 1991; Lazarus and Folkman, 1984). Different studies (Blascovich et al. , 2002; Blascovich and Mendes, 2000; Lazarus and Folkman, 1984; Lazarus et al. , 1980) have found that challenge leads to positive feelings associated with enjoyment, better performance, eagerness and anticipation of personal rewards or benefits. Several studies which focussed on the threat and challenge variable were carried out in the library and information science environment as well (Aharony, 2009, 2011).Self-efficacyAn additional variable which may influence individuals' behavioural intention to use cloud computing is self-efficacy. The concept of self-efficacy was developed in the discipline of "social learning theory" by Bandura (1997). Self-efficacy addresses individuals' beliefs that they possess the resources and skills needed to perform and succeed in a specific task. Therefore, individuals' previous performance and their perceptions of relevant resources available may influence self-efficacy beliefs (Bandura, 1997). Self-efficacy is not just an ability perception, it encompasses the motivation and effort required to complete the task and it helps determine which activities are required, the effort in pursuing these activities and persistence when facing obstacles (Bandura, 1986, 1997). The construct of self-efficacy is made up of four principal sources of information:"mastery experience" refers to previous experience, including success and failure; "vicarious experience" addresses observing the performances, successes and failures of others;"social persuasion" includes verbal persuasion from peers, colleagues and relatives; and"physiological and emotional states" from which people judge their strengths, capabilities and vulnerabilities (Bandura, 1986, 1994, 1995).As self-efficacy is based on self-perceptions regarding different behaviours, it is considered to be situation specific. In other words, a person may exhibit high levels of self-efficacy within one domain, while exhibiting low levels within another (Cassidy and Eachus, 2002). Thus, self-efficacy has generated research in various disciplines such as medicine, business, psychology and education (Kear, 2000; Lev, 1997; Schunk, 1985; Koul and Rubba, 1999). Computer self-efficacy is a sub-field of self-efficacy. It is defined as one's perceived ability to accomplish a task with the use of a computer (Compeau and Higgins, 1995). Various studies have noted that training and experience play important roles in computer self-efficacy (Compeau and Higgins, 1995; Kinzie et al. , 1994; Stone and Henry, 2003). Several studies have investigated the effect of computer self-efficacy on computer training performance (Compeau and Higgins, 1995) and on IT use (Easley et al. , 2003).HypothesesBased on the study objectives and assuming that PEOU, personal innovativeness,cognitive appraisal and openness to experience may predict information professionals' behavioural intention to use cloud computing, the underlying assumptions of this study are as follows:H1. High scores in respondent PEOU will be associated with high scores in their behavioural intention to use cloud computing.H2. High scores in respondents' personal innovativeness will be associated with high scores in their behavioural intention to use cloud computing.H3. Low scores in respondents' threat and high scores in respondents' challenge will be associated with high scores in their behavioural intention to use cloud computing. H4. High scores in respondents' self-efficacy will be associated with high scores in their behavioural intention to use cloud computing.H5. High scores in respondents' openness to experience will be associated with high scores in their behavioural intention to use cloud computing.H6. High scores in respondents' computer competence and in social media use will be associated with high scores in their behavioural intention to use cloud computing. MethodologyData collectionThe research was conducted in Israel during the second semester of the 2013 academic year and encompassed two groups of information professionals: librarians and information specialists. The researchers sent a message and a questionnaire to an Israeli library and information science discussion group named "safranym", which included school, public and academic librarians, and to an Israeli information specialist group named "I-fish", which consists of information specialists that work in different organizations. Researchers explained the study's purpose and asked their members to complete the questionnaire. These two discussion groups consist of about 700 members; 140 responses were received, giving a reply percentage of 20 per cent. Data analysisOf the participants, 25 (17.9 per cent) were male and 115 (82.1 per cent) were female. Their average age was 46.3 years.MeasuresThe current study is based on quantitative research. Researchers used seven questionnaires to gather the following data: personal details, computer competence, attitudes towards cloud computing, behavioural intention, openness to experience, cognitive appraisal and self-efficacy.The personal details questionnaire had two statements. The computer competence questionnaire consisted of two statements rated on a 5-point Likert scale (1 = strongest disagreement; 5 = strongest agreement). The cloud computing attitude questionnaire, based on Liuet al. (2010), was modified for this study and consisted of six statements rated on a seven-point Likert scale (1 = strongest disagreement; 7 = strongest agreement). A principal components factor analysis using Varimax rotation with Kaiser Normalization was conducted and explained 82.98 per cent of the variance. Principal components factor analysis revealed two distinct factors. The first related to information professionals' personal innovativeness (items 2, 3 and 5), and the second to information professionals' perceptions about cloud computing ease ofuse (PEOU) (items 1, 4, and 6); the values of Cronbach's Alpha were 0.89 and 0.88, respectively.The behavioural intention questionnaire, based on Liu et al. (2010), was modified for this study and consisted of three statements rated on a six-point Likert scale (1 = strongest disagreement; 6 = strongest agreement). Its Cronbach's Alpha was 0.79. The openness to experience questionnaire was derived from the Big Five questionnaire (John et al. , 1991) and consisted of eight statements rated on a five-point Likert scale (1 = strongest disagreement; 5 = strongest agreement); Cronbach's Alpha was 0.81. The cognitive appraisal questionnaire measured information professionals' feelings of threat versus challenge when confronted with new situations. It consisted of 10 statements rated on a six-point scale (1 = fully disagree; 6 = fully agree). This questionnaire was previously used (Aharony, 2009, 2011; Yekutiel, 1990) and consisted of two factors: threat (items 1, 2, 3, 5, 7 and 8) and challenge (items 4, 6, 9 and 10). Cronbach's Alpha was 0.70 for the threat factor and 0.89 for the challenge factor.The self-efficacy questionnaire was based on Askar and Umay's (2001) questionnaire and consisted of 18 statements rated on a five-point scale (1 = fully disagree; 5 = fully agree); Cronbach's Alpha was 0.96.FindingsTo examine the relationship between openness to experience, cognitive appraisal (threat, challenge and self-efficacy), TAM variables (personal innovativeness and PEOU), and behavioural intention to use cloud computing, researchers performed Pearson correlations, which are given in Table I.Table I presents significant correlations between research variables and the dependent variable (behavioural intention to use cloud computing). All correlations are positive, except the one between threat and behavioural intention to use cloud computing. Hence, the higher these measures, the greater the behavioural intention to use cloud computing. A significant negative correlation was found between threat and the dependent variable. Therefore, the more threatened respondents are, the lower is their behavioural intention to use cloud computing.Regarding the correlations between research variables, significant positive correlations were found between openness to experience and challenge, self-efficacy, personal innovativeness and PEOU. A significant negative correlation was found between openness to experience and threat. That is, the more open to experience respondents are, the more challenged they are, the higher is their self-efficacy, personal innovativeness, and PEOU and the less threatened they are. In addition, significant negative correlations were found between threat and self-efficacy, personal innovativeness and PEOU. We can conclude that the more threatened respondents are, the less they are self-efficient, personally innovative and the less they perceive cloud computing as easy to use. Significant positive correlations were also found between self-efficacy and personal innovativeness and PEOU. Thus, the more self-efficient respondents are, the more personally innovative they are and the more they perceive cloud computing as easy to use.The study also examined two variables associated with computer competence:computer use and social media use. Table II presents correlations between these two variables and the other research variables.Significant, high correlations were found between computer competence variables and openness to experience, self-efficacy, personal innovativeness, PEOU and behavioural intention to use cloud computing. Hence, the higher respondents' computer competence, the more they are open to experience, self-efficient and personally innovative, and perceive cloud computing as easy to use, the higher is their behavioural intention to use cloud computing.Researchers also examined relationships with demographic variables. To examine the relationship between age and other research variables, the researchers performed Pearson correlations. A significant negative correlation was found between age and PEOU, r = -0.21, p < 0.05. We may assume that the younger the respondents are, the more they perceive cloud computing as easy to use. To examine whether there are differences between males and females concerning the research variables, a MANOV A was performed and did not reveal a significant difference between the two groups concerning research variables, F (7,130) = 1.88, p > 0.05.The researchers also conducted a hierarchical regression using behavioural intention to use cloud computing as a dependent variable. The predictors were entered as five steps:respondents' openness to experience;respondents' computer competence (computer use and social media use);cognitive appraisal (threat, challenge and self-efficacy);TAM variables (personal innovativeness and PEOU); andinteractions with the TAM variables.The entrance of the four first steps was forced, while the interactions were done according to their contribution to the explained variance of behavioural intention to use cloud computing. The regression explained 54 per cent of behavioural intention to use cloud computing. Table III presents the standardized and unstandardized coefficients of the hierarchical regression of respondents' behavioural intention to use cloud computing.The first step introduced the openness variable that contributed significantly by adding 13 per cent to the explained variance of behavioural intention to use cloud computing. The beta coefficient of the openness variable is positive; hence, the more open to experience respondents are, the higher is their behavioural intention to use cloud computing. The second step introduced the two computer competence variables (computer use and social media use) which contributed 5 per cent to the explained variance of behavioural intention. Of these two variables, only the social media variable contributed significantly and its beta coefficient was positive. In other words, the more respondents use social media, the higher is their behavioural intention to use cloud computing. Note that Pearson correlations found significant positive correlations between these two variables and behavioural intention to use cloud computing. It seems that because of the correlation between these two variables, r = 0.33, p < 0.001, the computer use variable did not contribute to the regression.As the third step, researchers added respondents' personal appraisal variables (threat and challenge, and self-efficacy), and this also contributed significantly by adding 25 per cent to the explained variance of behavioural intention. The beta coefficients of challenge and of self-efficacy were positive, while that of threat was negative. Therefore, we may conclude that the more respondents perceived themselves as challenged and self-efficient, and the less they perceived themselves as threatened, the higher is their behavioural intention to use cloud computing. The inclusion of this step caused a decrease in the [beta] size of the openness to experience variable that changed it into an insignificant one, and may suggest a possibility of mediation. Sobel tests indicated that self-efficacy mediates between openness to experience and behavioural intention (z = 4.68, p < 0.001). Hence, the more respondents are open to experience, the higher is their self-efficacy and, as a result, the higher is their behavioural intention to use cloud computing.The fourth step added the TAM variables (respondents' PEOU and personal innovation), and this also contributed significantly by adding 9 per cent to the explained variance of behavioural intention to use cloud computing. The beta coefficient of this variable was positive; therefore, the more respondents perceived themselves to be personally innovative and cloud computing as easy to use, the higher is their behavioural intention to use cloud computing. Note that in this step there was a decrease in the [beta] size of self-efficacy. Sobel tests indicated that of the two variables, PEOU mediates between self-efficacy and behavioural intention (z = 4.77, p < 0.001). Thus, the more respondents perceive themselves as self-efficient, the higher they perceive cloud computing's PEOU and, as a result, the higher is their behavioural intention to use it.As the fifth step, researchers added the interaction between computer use X personal innovativeness. This interaction added 2 per cent to the explained variance of behavioural intention to use cloud computing and is presented in Figure 1.Figure 1 shows a correlation between personal innovation and behavioural intention to use cloud computing among respondents who are low and high in computer use. This correlation is higher among respondents who are low in computer use, [beta] = . 40, p < 0.05, than among those who are high in computer use, [beta] = 0.04, p < 0.05. It seems that especially among participants who are low in computer use, the higher their personal innovativeness, the higher is their behavioural intention to use cloud computing.DiscussionThe present research explored the extent to which the TAM and personal characteristics, such as threat and challenge, self-efficacy and openness to experience, explain information professionals' perspectives on cloud computing. Researchers divided the study hypotheses into three categories. The first (consisting of H1 -H2 ) refers to the TAM, the second (H3 -H5 ) to personality characteristics and, finally, H6 to computer competence. All hypotheses were accepted. Regarding the first category of hypotheses, results show that both were accepted. Findings suggest that high scores in PEOU and personal innovativeness are associated with high scores in respondents' intention to adopt cloud computing. These findings can be associated with previous。

云计算外文翻译参考文献

云计算外文翻译参考文献

云计算外文翻译参考文献(文档含中英文对照即英文原文和中文翻译)原文:Technical Issues of Forensic Investigations in Cloud Computing EnvironmentsDominik BirkRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyAbstract—Cloud Computing is arguably one of the most discussedinformation technologies today. It presents many promising technological and economical opportunities. However, many customers remain reluctant to move their business IT infrastructure completely to the cloud. One of their main concerns is Cloud Security and the threat of the unknown. Cloud Service Providers(CSP) encourage this perception by not letting their customers see what is behind their virtual curtain. A seldomly discussed, but in this regard highly relevant open issue is the ability to perform digital investigations. This continues to fuel insecurity on the sides of both providers and customers. Cloud Forensics constitutes a new and disruptive challenge for investigators. Due to the decentralized nature of data processing in the cloud, traditional approaches to evidence collection and recovery are no longer practical. This paper focuses on the technical aspects of digital forensics in distributed cloud environments. We contribute by assessing whether it is possible for the customer of cloud computing services to perform a traditional digital investigation from a technical point of view. Furthermore we discuss possible solutions and possible new methodologies helping customers to perform such investigations.I. INTRODUCTIONAlthough the cloud might appear attractive to small as well as to large companies, it does not come along without its own unique problems. Outsourcing sensitive corporate data into the cloud raises concerns regarding the privacy and security of data. Security policies, companies main pillar concerning security, cannot be easily deployed into distributed, virtualized cloud environments. This situation is further complicated by the unknown physical location of the companie’s assets. Normally,if a security incident occurs, the corporate security team wants to be able to perform their own investigation without dependency on third parties. In the cloud, this is not possible anymore: The CSP obtains all the power over the environmentand thus controls the sources of evidence. In the best case, a trusted third party acts as a trustee and guarantees for the trustworthiness of the CSP. Furthermore, the implementation of the technical architecture and circumstances within cloud computing environments bias the way an investigation may be processed. In detail, evidence data has to be interpreted by an investigator in a We would like to thank the reviewers for the helpful comments and Dennis Heinson (Center for Advanced Security Research Darmstadt - CASED) for the profound discussions regarding the legal aspects of cloud forensics. proper manner which is hardly be possible due to the lackof circumstantial information. For auditors, this situation does not change: Questions who accessed specific data and information cannot be answered by the customers, if no corresponding logs are available. With the increasing demand for using the power of the cloud for processing also sensible information and data, enterprises face the issue of Data and Process Provenance in the cloud [10]. Digital provenance, meaning meta-data that describes the ancestry or history of a digital object, is a crucial feature for forensic investigations. In combination with a suitable authentication scheme, it provides information about who created and who modified what kind of data in the cloud. These are crucial aspects for digital investigations in distributed environments such as the cloud. Unfortunately, the aspects of forensic investigations in distributed environment have so far been mostly neglected by the research community. Current discussion centers mostly around security, privacy and data protection issues [35], [9], [12]. The impact of forensic investigations on cloud environments was little noticed albeit mentioned by the authors of [1] in 2009: ”[...] to our knowledge, no research has been published on how cloud computing environments affect digital artifacts,and on acquisition logistics and legal issues related to cloud computing env ironments.” This statement is also confirmed by other authors [34], [36], [40] stressing that further research on incident handling, evidence tracking and accountability in cloud environments has to be done. At the same time, massive investments are being made in cloud technology. Combined with the fact that information technology increasingly transcendents peoples’ private and professional life, thus mirroring more and more of peoples’actions, it becomes apparent that evidence gathered from cloud environments will be of high significance to litigation or criminal proceedings in the future. Within this work, we focus the notion of cloud forensics by addressing the technical issues of forensics in all three major cloud service models and consider cross-disciplinary aspects. Moreover, we address the usability of various sources of evidence for investigative purposes and propose potential solutions to the issues from a practical standpoint. This work should be considered as a surveying discussion of an almost unexplored research area. The paper is organized as follows: We discuss the related work and the fundamental technical background information of digital forensics, cloud computing and the fault model in section II and III. In section IV, we focus on the technical issues of cloud forensics and discuss the potential sources and nature of digital evidence as well as investigations in XaaS environments including thecross-disciplinary aspects. We conclude in section V.II. RELATED WORKVarious works have been published in the field of cloud security and privacy [9], [35], [30] focussing on aspects for protecting data in multi-tenant, virtualized environments. Desired security characteristics for current cloud infrastructures mainly revolve around isolation of multi-tenant platforms [12], security of hypervisors in order to protect virtualized guest systems and secure network infrastructures [32]. Albeit digital provenance, describing the ancestry of digital objects, still remains a challenging issue for cloud environments, several works have already been published in this field [8], [10] contributing to the issues of cloud forensis. Within this context, cryptographic proofs for verifying data integrity mainly in cloud storage offers have been proposed,yet lacking of practical implementations [24], [37], [23]. Traditional computer forensics has already well researched methods for various fields of application [4], [5], [6], [11], [13]. Also the aspects of forensics in virtual systems have been addressed by several works [2], [3], [20] including the notionof virtual introspection [25]. In addition, the NIST already addressed Web Service Forensics [22] which has a huge impact on investigation processes in cloud computing environments. In contrast, the aspects of forensic investigations in cloud environments have mostly been neglected by both the industry and the research community. One of the first papers focusing on this topic was published by Wolthusen [40] after Bebee et al already introduced problems within cloud environments [1]. Wolthusen stressed that there is an inherent strong need for interdisciplinary work linking the requirements and concepts of evidence arising from the legal field to what can be feasibly reconstructed and inferred algorithmically or in an exploratory manner. In 2010, Grobauer et al [36] published a paper discussing the issues of incident response in cloud environments - unfortunately no specific issues and solutions of cloud forensics have been proposed which will be done within this work.III. TECHNICAL BACKGROUNDA. Traditional Digital ForensicsThe notion of Digital Forensics is widely known as the practice of identifying, extracting and considering evidence from digital media. Unfortunately, digital evidence is both fragile and volatile and therefore requires the attention of special personnel and methods in order to ensure that evidence data can be proper isolated and evaluated. Normally, the process of a digital investigation can be separated into three different steps each having its own specificpurpose:1) In the Securing Phase, the major intention is the preservation of evidence for analysis. The data has to be collected in a manner that maximizes its integrity. This is normally done by a bitwise copy of the original media. As can be imagined, this represents a huge problem in the field of cloud computing where you never know exactly where your data is and additionallydo not have access to any physical hardware. However, the snapshot technology, discussed in section IV-B3, provides a powerful tool to freeze system states and thus makes digital investigations, at least in IaaS scenarios, theoretically possible.2) We refer to the Analyzing Phase as the stage in which the data is sifted and combined. It is in this phase that the data from multiple systems or sources is pulled together to create as complete a picture and event reconstruction as possible. Especially in distributed system infrastructures, this means that bits and pieces of data are pulled together for deciphering the real story of what happened and for providing a deeper look into the data.3) Finally, at the end of the examination and analysis of the data, the results of the previous phases will be reprocessed in the Presentation Phase. The report, created in this phase, is a compilation of all the documentation and evidence from the analysis stage. The main intention of such a report is that it contains all results, it is complete and clear to understand. Apparently, the success of these three steps strongly depends on the first stage. If it is not possible to secure the complete set of evidence data, no exhaustive analysis will be possible. However, in real world scenarios often only a subset of the evidence data can be secured by the investigator. In addition, an important definition in the general context of forensics is the notion of a Chain of Custody. This chain clarifies how and where evidence is stored and who takes possession of it. Especially for cases which are brought to court it is crucial that the chain of custody is preserved.B. Cloud ComputingAccording to the NIST [16], cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications and services) that can be rapidly provisioned and released with minimal CSP interaction. The new raw definition of cloud computing brought several new characteristics such as multi-tenancy, elasticity, pay-as-you-go and reliability. Within this work, the following three models are used: In the Infrastructure asa Service (IaaS) model, the customer is using the virtual machine provided by the CSP for installing his own system on it. The system can be used like any other physical computer with a few limitations. However, the additive customer power over the system comes along with additional security obligations. Platform as a Service (PaaS) offerings provide the capability to deploy application packages created using the virtual development environment supported by the CSP. For the efficiency of software development process this service model can be propellent. In the Software as a Service (SaaS) model, the customer makes use of a service run by the CSP on a cloud infrastructure. In most of the cases this service can be accessed through an API for a thin client interface such as a web browser. Closed-source public SaaS offers such as Amazon S3 and GoogleMail can only be used in the public deployment model leading to further issues concerning security, privacy and the gathering of suitable evidences. Furthermore, two main deployment models, private and public cloud have to be distinguished. Common public clouds are made available to the general public. The corresponding infrastructure is owned by one organization acting as a CSP and offering services to its customers. In contrast, the private cloud is exclusively operated for an organization but may not provide the scalability and agility of public offers. The additional notions of community and hybrid cloud are not exclusively covered within this work. However, independently from the specific model used, the movement of applications and data to the cloud comes along with limited control for the customer about the application itself, the data pushed into the applications and also about the underlying technical infrastructure.C. Fault ModelBe it an account for a SaaS application, a development environment (PaaS) or a virtual image of an IaaS environment, systems in the cloud can be affected by inconsistencies. Hence, for both customer and CSP it is crucial to have the ability to assign faults to the causing party, even in the presence of Byzantine behavior [33]. Generally, inconsistencies can be caused by the following two reasons:1) Maliciously Intended FaultsInternal or external adversaries with specific malicious intentions can cause faults on cloud instances or applications. Economic rivals as well as former employees can be the reason for these faults and state a constant threat to customers and CSP. In this model, also a malicious CSP is included albeit he isassumed to be rare in real world scenarios. Additionally, from the technical point of view, the movement of computing power to a virtualized, multi-tenant environment can pose further threads and risks to the systems. One reason for this is that if a single system or service in the cloud is compromised, all other guest systems and even the host system are at risk. Hence, besides the need for further security measures, precautions for potential forensic investigations have to be taken into consideration.2) Unintentional FaultsInconsistencies in technical systems or processes in the cloud do not have implicitly to be caused by malicious intent. Internal communication errors or human failures can lead to issues in the services offered to the costumer(i.e. loss or modification of data). Although these failures are not caused intentionally, both the CSP and the customer have a strong intention to discover the reasons and deploy corresponding fixes.IV. TECHNICAL ISSUESDigital investigations are about control of forensic evidence data. From the technical standpoint, this data can be available in three different states: at rest, in motion or in execution. Data at rest is represented by allocated disk space. Whether the data is stored in a database or in a specific file format, it allocates disk space. Furthermore, if a file is deleted, the disk space is de-allocated for the operating system but the data is still accessible since the disk space has not been re-allocated and overwritten. This fact is often exploited by investigators which explore these de-allocated disk space on harddisks. In case the data is in motion, data is transferred from one entity to another e.g. a typical file transfer over a network can be seen as a data in motion scenario. Several encapsulated protocols contain the data each leaving specific traces on systems and network devices which can in return be used by investigators. Data can be loaded into memory and executed as a process. In this case, the data is neither at rest or in motion but in execution. On the executing system, process information, machine instruction and allocated/de-allocated data can be analyzed by creating a snapshot of the current system state. In the following sections, we point out the potential sources for evidential data in cloud environments and discuss the technical issues of digital investigations in XaaS environmentsas well as suggest several solutions to these problems.A. Sources and Nature of EvidenceConcerning the technical aspects of forensic investigations, the amount of potential evidence available to the investigator strongly diverges between thedifferent cloud service and deployment models. The virtual machine (VM), hosting in most of the cases the server application, provides several pieces of information that could be used by investigators. On the network level, network components can provide information about possible communication channels between different parties involved. The browser on the client, acting often as the user agent for communicating with the cloud, also contains a lot of information that could be used as evidence in a forensic investigation. Independently from the used model, the following three components could act as sources for potential evidential data.1) Virtual Cloud Instance: The VM within the cloud, where i.e. data is stored or processes are handled, contains potential evidence [2], [3]. In most of the cases, it is the place where an incident happened and hence provides a good starting point for a forensic investigation. The VM instance can be accessed by both, the CSP and the customer who is running the instance. Furthermore, virtual introspection techniques [25] provide access to the runtime state of the VM via the hypervisor and snapshot technology supplies a powerful technique for the customer to freeze specific states of the VM. Therefore, virtual instances can be still running during analysis which leads to the case of live investigations [41] or can be turned off leading to static image analysis. In SaaS and PaaS scenarios, the ability to access the virtual instance for gathering evidential information is highly limited or simply not possible.2) Network Layer: Traditional network forensics is knownas the analysis of network traffic logs for tracing events that have occurred in the past. Since the different ISO/OSI network layers provide several information on protocols and communication between instances within as well as with instances outside the cloud [4], [5], [6], network forensics is theoretically also feasible in cloud environments. However in practice, ordinary CSP currently do not provide any log data from the network components used by the customer’s instances or applications. For instance, in case of a malware infection of an IaaS VM, it will be difficult for the investigator to get any form of routing information and network log datain general which is crucial for further investigative steps. This situation gets even more complicated in case of PaaS or SaaS. So again, the situation of gathering forensic evidence is strongly affected by the support the investigator receives from the customer and the CSP.3) Client System: On the system layer of the client, it completely depends on the used model (IaaS, PaaS, SaaS) if and where potential evidence could beextracted. In most of the scenarios, the user agent (e.g. the web browser) on the client system is the only application that communicates with the service in the cloud. This especially holds for SaaS applications which are used and controlled by the web browser. But also in IaaS scenarios, the administration interface is often controlled via the browser. Hence, in an exhaustive forensic investigation, the evidence data gathered from the browser environment [7] should not be omitted.a) Browser Forensics: Generally, the circumstances leading to an investigation have to be differentiated: In ordinary scenarios, the main goal of an investigation of the web browser is to determine if a user has been victim of a crime. In complex SaaS scenarios with high client-server interaction, this constitutes a difficult task. Additionally, customers strongly make use of third-party extensions [17] which can be abused for malicious purposes. Hence, the investigator might want to look for malicious extensions, searches performed, websites visited, files downloaded, information entered in forms or stored in local HTML5 stores, web-based email contents and persistent browser cookies for gathering potential evidence data. Within this context, it is inevitable to investigate the appearance of malicious JavaScript [18] leading to e.g. unintended AJAX requests and hence modified usage of administration interfaces. Generally, the web browser contains a lot of electronic evidence data that could be used to give an answer to both of the above questions - even if the private mode is switched on [19].B. Investigations in XaaS EnvironmentsTraditional digital forensic methodologies permit investigators to seize equipment and perform detailed analysis on the media and data recovered [11]. In a distributed infrastructure organization like the cloud computing environment, investigators are confronted with an entirely different situation. They have no longer the option of seizing physical data storage. Data and processes of the customer are dispensed over an undisclosed amount of virtual instances, applications and network elements. Hence, it is in question whether preliminary findings of the computer forensic community in the field of digital forensics apparently have to be revised and adapted to the new environment. Within this section, specific issues of investigations in SaaS, PaaS and IaaS environments will be discussed. In addition, cross-disciplinary issues which affect several environments uniformly, will be taken into consideration. We also suggest potential solutions to the mentioned problems.1) SaaS Environments: Especially in the SaaS model, the customer does notobtain any control of the underlying operating infrastructure such as network, servers, operating systems or the application that is used. This means that no deeper view into the system and its underlying infrastructure is provided to the customer. Only limited userspecific application configuration settings can be controlled contributing to the evidences which can be extracted fromthe client (see section IV-A3). In a lot of cases this urges the investigator to rely on high-level logs which are eventually provided by the CSP. Given the case that the CSP does not run any logging application, the customer has no opportunity to create any useful evidence through the installation of any toolkit or logging tool. These circumstances do not allow a valid forensic investigation and lead to the assumption that customers of SaaS offers do not have any chance to analyze potential incidences.a) Data Provenance: The notion of Digital Provenance is known as meta-data that describes the ancestry or history of digital objects. Secure provenance that records ownership and process history of data objects is vital to the success of data forensics in cloud environments, yet it is still a challenging issue today [8]. Albeit data provenance is of high significance also for IaaS and PaaS, it states a huge problem specifically for SaaS-based applications: Current global acting public SaaS CSP offer Single Sign-On (SSO) access control to the set of their services. Unfortunately in case of an account compromise, most of the CSP do not offer any possibility for the customer to figure out which data and information has been accessed by the adversary. For the victim, this situation can have tremendous impact: If sensitive data has been compromised, it is unclear which data has been leaked and which has not been accessed by the adversary. Additionally, data could be modified or deleted by an external adversary or even by the CSP e.g. due to storage reasons. The customer has no ability to proof otherwise. Secure provenance mechanisms for distributed environments can improve this situation but have not been practically implemented by CSP [10]. Suggested Solution: In private SaaS scenarios this situation is improved by the fact that the customer and the CSP are probably under the same authority. Hence, logging and provenance mechanisms could be implemented which contribute to potential investigations. Additionally, the exact location of the servers and the data is known at any time. Public SaaS CSP should offer additional interfaces for the purpose of compliance, forensics, operations and security matters to their customers. Through an API, the customers should have the ability to receive specific information suchas access, error and event logs that could improve their situation in case of aninvestigation. Furthermore, due to the limited ability of receiving forensic information from the server and proofing integrity of stored data in SaaS scenarios, the client has to contribute to this process. This could be achieved by implementing Proofs of Retrievability (POR) in which a verifier (client) is enabled to determine that a prover (server) possesses a file or data object and it can be retrieved unmodified [24]. Provable Data Possession (PDP) techniques [37] could be used to verify that an untrusted server possesses the original data without the need for the client to retrieve it. Although these cryptographic proofs have not been implemented by any CSP, the authors of [23] introduced a new data integrity verification mechanism for SaaS scenarios which could also be used for forensic purposes.2) PaaS Environments: One of the main advantages of the PaaS model is that the developed software application is under the control of the customer and except for some CSP, the source code of the application does not have to leave the local development environment. Given these circumstances, the customer obtains theoretically the power to dictate how the application interacts with other dependencies such as databases, storage entities etc. CSP normally claim this transfer is encrypted but this statement can hardly be verified by the customer. Since the customer has the ability to interact with the platform over a prepared API, system states and specific application logs can be extracted. However potential adversaries, which can compromise the application during runtime, should not be able to alter these log files afterwards. Suggested Solution:Depending on the runtime environment, logging mechanisms could be implemented which automatically sign and encrypt the log information before its transfer to a central logging server under the control of the customer. Additional signing and encrypting could prevent potential eavesdroppers from being able to view and alter log data information on the way to the logging server. Runtime compromise of an PaaS application by adversaries could be monitored by push-only mechanisms for log data presupposing that the needed information to detect such an attack are logged. Increasingly, CSP offering PaaS solutions give developers the ability to collect and store a variety of diagnostics data in a highly configurable way with the help of runtime feature sets [38].3) IaaS Environments: As expected, even virtual instances in the cloud get compromised by adversaries. Hence, the ability to determine how defenses in the virtual environment failed and to what extent the affected systems havebeen compromised is crucial not only for recovering from an incident. Also forensic investigations gain leverage from such information and contribute to resilience against future attacks on the systems. From the forensic point of view, IaaS instances do provide much more evidence data usable for potential forensics than PaaS and SaaS models do. This fact is caused throughthe ability of the customer to install and set up the image for forensic purposes before an incident occurs. Hence, as proposed for PaaS environments, log data and other forensic evidence information could be signed and encrypted before itis transferred to third-party hosts mitigating the chance that a maliciously motivated shutdown process destroys the volatile data. Although, IaaS environments provide plenty of potential evidence, it has to be emphasized that the customer VM is in the end still under the control of the CSP. He controls the hypervisor which is e.g. responsible for enforcing hardware boundaries and routing hardware requests among different VM. Hence, besides the security responsibilities of the hypervisor, he exerts tremendous control over how customer’s VM communicate with the hardware and theoretically can intervene executed processes on the hosted virtual instance through virtual introspection [25]. This could also affect encryption or signing processes executed on the VM and therefore leading to the leakage of the secret key. Although this risk can be disregarded in most of the cases, the impact on the security of high security environments is tremendous.a) Snapshot Analysis: Traditional forensics expect target machines to be powered down to collect an image (dead virtual instance). This situation completely changed with the advent of the snapshot technology which is supported by all popular hypervisors such as Xen, VMware ESX and Hyper-V.A snapshot, also referred to as the forensic image of a VM, providesa powerful tool with which a virtual instance can be clonedby one click including also the running system’s mem ory. Due to the invention of the snapshot technology, systems hosting crucial business processes do not have to be powered down for forensic investigation purposes. The investigator simply creates and loads a snapshot of the target VM for analysis(live virtual instance). This behavior is especially important for scenarios in which a downtime of a system is not feasible or practical due to existing SLA. However the information whether the machine is running or has been properly powered down is crucial [3] for the investigation. Live investigations of running virtual instances become more common providing evidence data that。

云计算外文翻译参考文献

云计算外文翻译参考文献

云计算外文翻译参考文献(文档含中英文对照即英文原文和中文翻译)原文:Technical Issues of Forensic Investigations in Cloud Computing EnvironmentsDominik BirkRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyAbstract—Cloud Computing is arguably one of the most discussedinformation technologies today. It presents many promising technological and economical opportunities. However, many customers remain reluctant to move their business IT infrastructure completely to the cloud. One of their main concerns is Cloud Security and the threat of the unknown. Cloud Service Providers(CSP) encourage this perception by not letting their customers see what is behind their virtual curtain. A seldomly discussed, but in this regard highly relevant open issue is the ability to perform digital investigations. This continues to fuel insecurity on the sides of both providers and customers. Cloud Forensics constitutes a new and disruptive challenge for investigators. Due to the decentralized nature of data processing in the cloud, traditional approaches to evidence collection and recovery are no longer practical. This paper focuses on the technical aspects of digital forensics in distributed cloud environments. We contribute by assessing whether it is possible for the customer of cloud computing services to perform a traditional digital investigation from a technical point of view. Furthermore we discuss possible solutions and possible new methodologies helping customers to perform such investigations.I. INTRODUCTIONAlthough the cloud might appear attractive to small as well as to large companies, it does not come along without its own unique problems. Outsourcing sensitive corporate data into the cloud raises concerns regarding the privacy and security of data. Security policies, companies main pillar concerning security, cannot be easily deployed into distributed, virtualized cloud environments. This situation is further complicated by the unknown physical location of the companie’s assets. Normally,if a security incident occurs, the corporate security team wants to be able to perform their own investigation without dependency on third parties. In the cloud, this is not possible anymore: The CSP obtains all the power over the environmentand thus controls the sources of evidence. In the best case, a trusted third party acts as a trustee and guarantees for the trustworthiness of the CSP. Furthermore, the implementation of the technical architecture and circumstances within cloud computing environments bias the way an investigation may be processed. In detail, evidence data has to be interpreted by an investigator in a We would like to thank the reviewers for the helpful comments and Dennis Heinson (Center for Advanced Security Research Darmstadt - CASED) for the profound discussions regarding the legal aspects of cloud forensics. proper manner which is hardly be possible due to the lackof circumstantial information. For auditors, this situation does not change: Questions who accessed specific data and information cannot be answered by the customers, if no corresponding logs are available. With the increasing demand for using the power of the cloud for processing also sensible information and data, enterprises face the issue of Data and Process Provenance in the cloud [10]. Digital provenance, meaning meta-data that describes the ancestry or history of a digital object, is a crucial feature for forensic investigations. In combination with a suitable authentication scheme, it provides information about who created and who modified what kind of data in the cloud. These are crucial aspects for digital investigations in distributed environments such as the cloud. Unfortunately, the aspects of forensic investigations in distributed environment have so far been mostly neglected by the research community. Current discussion centers mostly around security, privacy and data protection issues [35], [9], [12]. The impact of forensic investigations on cloud environments was little noticed albeit mentioned by the authors of [1] in 2009: ”[...] to our knowledge, no research has been published on how cloud computing environments affect digital artifacts,and on acquisition logistics and legal issues related to cloud computing env ironments.” This statement is also confirmed by other authors [34], [36], [40] stressing that further research on incident handling, evidence tracking and accountability in cloud environments has to be done. At the same time, massive investments are being made in cloud technology. Combined with the fact that information technology increasingly transcendents peoples’ private and professional life, thus mirroring more and more of peoples’actions, it becomes apparent that evidence gathered from cloud environments will be of high significance to litigation or criminal proceedings in the future. Within this work, we focus the notion of cloud forensics by addressing the technical issues of forensics in all three major cloud service models and consider cross-disciplinary aspects. Moreover, we address the usability of various sources of evidence for investigative purposes and propose potential solutions to the issues from a practical standpoint. This work should be considered as a surveying discussion of an almost unexplored research area. The paper is organized as follows: We discuss the related work and the fundamental technical background information of digital forensics, cloud computing and the fault model in section II and III. In section IV, we focus on the technical issues of cloud forensics and discuss the potential sources and nature of digital evidence as well as investigations in XaaS environments including thecross-disciplinary aspects. We conclude in section V.II. RELATED WORKVarious works have been published in the field of cloud security and privacy [9], [35], [30] focussing on aspects for protecting data in multi-tenant, virtualized environments. Desired security characteristics for current cloud infrastructures mainly revolve around isolation of multi-tenant platforms [12], security of hypervisors in order to protect virtualized guest systems and secure network infrastructures [32]. Albeit digital provenance, describing the ancestry of digital objects, still remains a challenging issue for cloud environments, several works have already been published in this field [8], [10] contributing to the issues of cloud forensis. Within this context, cryptographic proofs for verifying data integrity mainly in cloud storage offers have been proposed,yet lacking of practical implementations [24], [37], [23]. Traditional computer forensics has already well researched methods for various fields of application [4], [5], [6], [11], [13]. Also the aspects of forensics in virtual systems have been addressed by several works [2], [3], [20] including the notionof virtual introspection [25]. In addition, the NIST already addressed Web Service Forensics [22] which has a huge impact on investigation processes in cloud computing environments. In contrast, the aspects of forensic investigations in cloud environments have mostly been neglected by both the industry and the research community. One of the first papers focusing on this topic was published by Wolthusen [40] after Bebee et al already introduced problems within cloud environments [1]. Wolthusen stressed that there is an inherent strong need for interdisciplinary work linking the requirements and concepts of evidence arising from the legal field to what can be feasibly reconstructed and inferred algorithmically or in an exploratory manner. In 2010, Grobauer et al [36] published a paper discussing the issues of incident response in cloud environments - unfortunately no specific issues and solutions of cloud forensics have been proposed which will be done within this work.III. TECHNICAL BACKGROUNDA. Traditional Digital ForensicsThe notion of Digital Forensics is widely known as the practice of identifying, extracting and considering evidence from digital media. Unfortunately, digital evidence is both fragile and volatile and therefore requires the attention of special personnel and methods in order to ensure that evidence data can be proper isolated and evaluated. Normally, the process of a digital investigation can be separated into three different steps each having its own specificpurpose:1) In the Securing Phase, the major intention is the preservation of evidence for analysis. The data has to be collected in a manner that maximizes its integrity. This is normally done by a bitwise copy of the original media. As can be imagined, this represents a huge problem in the field of cloud computing where you never know exactly where your data is and additionallydo not have access to any physical hardware. However, the snapshot technology, discussed in section IV-B3, provides a powerful tool to freeze system states and thus makes digital investigations, at least in IaaS scenarios, theoretically possible.2) We refer to the Analyzing Phase as the stage in which the data is sifted and combined. It is in this phase that the data from multiple systems or sources is pulled together to create as complete a picture and event reconstruction as possible. Especially in distributed system infrastructures, this means that bits and pieces of data are pulled together for deciphering the real story of what happened and for providing a deeper look into the data.3) Finally, at the end of the examination and analysis of the data, the results of the previous phases will be reprocessed in the Presentation Phase. The report, created in this phase, is a compilation of all the documentation and evidence from the analysis stage. The main intention of such a report is that it contains all results, it is complete and clear to understand. Apparently, the success of these three steps strongly depends on the first stage. If it is not possible to secure the complete set of evidence data, no exhaustive analysis will be possible. However, in real world scenarios often only a subset of the evidence data can be secured by the investigator. In addition, an important definition in the general context of forensics is the notion of a Chain of Custody. This chain clarifies how and where evidence is stored and who takes possession of it. Especially for cases which are brought to court it is crucial that the chain of custody is preserved.B. Cloud ComputingAccording to the NIST [16], cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications and services) that can be rapidly provisioned and released with minimal CSP interaction. The new raw definition of cloud computing brought several new characteristics such as multi-tenancy, elasticity, pay-as-you-go and reliability. Within this work, the following three models are used: In the Infrastructure asa Service (IaaS) model, the customer is using the virtual machine provided by the CSP for installing his own system on it. The system can be used like any other physical computer with a few limitations. However, the additive customer power over the system comes along with additional security obligations. Platform as a Service (PaaS) offerings provide the capability to deploy application packages created using the virtual development environment supported by the CSP. For the efficiency of software development process this service model can be propellent. In the Software as a Service (SaaS) model, the customer makes use of a service run by the CSP on a cloud infrastructure. In most of the cases this service can be accessed through an API for a thin client interface such as a web browser. Closed-source public SaaS offers such as Amazon S3 and GoogleMail can only be used in the public deployment model leading to further issues concerning security, privacy and the gathering of suitable evidences. Furthermore, two main deployment models, private and public cloud have to be distinguished. Common public clouds are made available to the general public. The corresponding infrastructure is owned by one organization acting as a CSP and offering services to its customers. In contrast, the private cloud is exclusively operated for an organization but may not provide the scalability and agility of public offers. The additional notions of community and hybrid cloud are not exclusively covered within this work. However, independently from the specific model used, the movement of applications and data to the cloud comes along with limited control for the customer about the application itself, the data pushed into the applications and also about the underlying technical infrastructure.C. Fault ModelBe it an account for a SaaS application, a development environment (PaaS) or a virtual image of an IaaS environment, systems in the cloud can be affected by inconsistencies. Hence, for both customer and CSP it is crucial to have the ability to assign faults to the causing party, even in the presence of Byzantine behavior [33]. Generally, inconsistencies can be caused by the following two reasons:1) Maliciously Intended FaultsInternal or external adversaries with specific malicious intentions can cause faults on cloud instances or applications. Economic rivals as well as former employees can be the reason for these faults and state a constant threat to customers and CSP. In this model, also a malicious CSP is included albeit he isassumed to be rare in real world scenarios. Additionally, from the technical point of view, the movement of computing power to a virtualized, multi-tenant environment can pose further threads and risks to the systems. One reason for this is that if a single system or service in the cloud is compromised, all other guest systems and even the host system are at risk. Hence, besides the need for further security measures, precautions for potential forensic investigations have to be taken into consideration.2) Unintentional FaultsInconsistencies in technical systems or processes in the cloud do not have implicitly to be caused by malicious intent. Internal communication errors or human failures can lead to issues in the services offered to the costumer(i.e. loss or modification of data). Although these failures are not caused intentionally, both the CSP and the customer have a strong intention to discover the reasons and deploy corresponding fixes.IV. TECHNICAL ISSUESDigital investigations are about control of forensic evidence data. From the technical standpoint, this data can be available in three different states: at rest, in motion or in execution. Data at rest is represented by allocated disk space. Whether the data is stored in a database or in a specific file format, it allocates disk space. Furthermore, if a file is deleted, the disk space is de-allocated for the operating system but the data is still accessible since the disk space has not been re-allocated and overwritten. This fact is often exploited by investigators which explore these de-allocated disk space on harddisks. In case the data is in motion, data is transferred from one entity to another e.g. a typical file transfer over a network can be seen as a data in motion scenario. Several encapsulated protocols contain the data each leaving specific traces on systems and network devices which can in return be used by investigators. Data can be loaded into memory and executed as a process. In this case, the data is neither at rest or in motion but in execution. On the executing system, process information, machine instruction and allocated/de-allocated data can be analyzed by creating a snapshot of the current system state. In the following sections, we point out the potential sources for evidential data in cloud environments and discuss the technical issues of digital investigations in XaaS environmentsas well as suggest several solutions to these problems.A. Sources and Nature of EvidenceConcerning the technical aspects of forensic investigations, the amount of potential evidence available to the investigator strongly diverges between thedifferent cloud service and deployment models. The virtual machine (VM), hosting in most of the cases the server application, provides several pieces of information that could be used by investigators. On the network level, network components can provide information about possible communication channels between different parties involved. The browser on the client, acting often as the user agent for communicating with the cloud, also contains a lot of information that could be used as evidence in a forensic investigation. Independently from the used model, the following three components could act as sources for potential evidential data.1) Virtual Cloud Instance: The VM within the cloud, where i.e. data is stored or processes are handled, contains potential evidence [2], [3]. In most of the cases, it is the place where an incident happened and hence provides a good starting point for a forensic investigation. The VM instance can be accessed by both, the CSP and the customer who is running the instance. Furthermore, virtual introspection techniques [25] provide access to the runtime state of the VM via the hypervisor and snapshot technology supplies a powerful technique for the customer to freeze specific states of the VM. Therefore, virtual instances can be still running during analysis which leads to the case of live investigations [41] or can be turned off leading to static image analysis. In SaaS and PaaS scenarios, the ability to access the virtual instance for gathering evidential information is highly limited or simply not possible.2) Network Layer: Traditional network forensics is knownas the analysis of network traffic logs for tracing events that have occurred in the past. Since the different ISO/OSI network layers provide several information on protocols and communication between instances within as well as with instances outside the cloud [4], [5], [6], network forensics is theoretically also feasible in cloud environments. However in practice, ordinary CSP currently do not provide any log data from the network components used by the customer’s instances or applications. For instance, in case of a malware infection of an IaaS VM, it will be difficult for the investigator to get any form of routing information and network log datain general which is crucial for further investigative steps. This situation gets even more complicated in case of PaaS or SaaS. So again, the situation of gathering forensic evidence is strongly affected by the support the investigator receives from the customer and the CSP.3) Client System: On the system layer of the client, it completely depends on the used model (IaaS, PaaS, SaaS) if and where potential evidence could beextracted. In most of the scenarios, the user agent (e.g. the web browser) on the client system is the only application that communicates with the service in the cloud. This especially holds for SaaS applications which are used and controlled by the web browser. But also in IaaS scenarios, the administration interface is often controlled via the browser. Hence, in an exhaustive forensic investigation, the evidence data gathered from the browser environment [7] should not be omitted.a) Browser Forensics: Generally, the circumstances leading to an investigation have to be differentiated: In ordinary scenarios, the main goal of an investigation of the web browser is to determine if a user has been victim of a crime. In complex SaaS scenarios with high client-server interaction, this constitutes a difficult task. Additionally, customers strongly make use of third-party extensions [17] which can be abused for malicious purposes. Hence, the investigator might want to look for malicious extensions, searches performed, websites visited, files downloaded, information entered in forms or stored in local HTML5 stores, web-based email contents and persistent browser cookies for gathering potential evidence data. Within this context, it is inevitable to investigate the appearance of malicious JavaScript [18] leading to e.g. unintended AJAX requests and hence modified usage of administration interfaces. Generally, the web browser contains a lot of electronic evidence data that could be used to give an answer to both of the above questions - even if the private mode is switched on [19].B. Investigations in XaaS EnvironmentsTraditional digital forensic methodologies permit investigators to seize equipment and perform detailed analysis on the media and data recovered [11]. In a distributed infrastructure organization like the cloud computing environment, investigators are confronted with an entirely different situation. They have no longer the option of seizing physical data storage. Data and processes of the customer are dispensed over an undisclosed amount of virtual instances, applications and network elements. Hence, it is in question whether preliminary findings of the computer forensic community in the field of digital forensics apparently have to be revised and adapted to the new environment. Within this section, specific issues of investigations in SaaS, PaaS and IaaS environments will be discussed. In addition, cross-disciplinary issues which affect several environments uniformly, will be taken into consideration. We also suggest potential solutions to the mentioned problems.1) SaaS Environments: Especially in the SaaS model, the customer does notobtain any control of the underlying operating infrastructure such as network, servers, operating systems or the application that is used. This means that no deeper view into the system and its underlying infrastructure is provided to the customer. Only limited userspecific application configuration settings can be controlled contributing to the evidences which can be extracted fromthe client (see section IV-A3). In a lot of cases this urges the investigator to rely on high-level logs which are eventually provided by the CSP. Given the case that the CSP does not run any logging application, the customer has no opportunity to create any useful evidence through the installation of any toolkit or logging tool. These circumstances do not allow a valid forensic investigation and lead to the assumption that customers of SaaS offers do not have any chance to analyze potential incidences.a) Data Provenance: The notion of Digital Provenance is known as meta-data that describes the ancestry or history of digital objects. Secure provenance that records ownership and process history of data objects is vital to the success of data forensics in cloud environments, yet it is still a challenging issue today [8]. Albeit data provenance is of high significance also for IaaS and PaaS, it states a huge problem specifically for SaaS-based applications: Current global acting public SaaS CSP offer Single Sign-On (SSO) access control to the set of their services. Unfortunately in case of an account compromise, most of the CSP do not offer any possibility for the customer to figure out which data and information has been accessed by the adversary. For the victim, this situation can have tremendous impact: If sensitive data has been compromised, it is unclear which data has been leaked and which has not been accessed by the adversary. Additionally, data could be modified or deleted by an external adversary or even by the CSP e.g. due to storage reasons. The customer has no ability to proof otherwise. Secure provenance mechanisms for distributed environments can improve this situation but have not been practically implemented by CSP [10]. Suggested Solution: In private SaaS scenarios this situation is improved by the fact that the customer and the CSP are probably under the same authority. Hence, logging and provenance mechanisms could be implemented which contribute to potential investigations. Additionally, the exact location of the servers and the data is known at any time. Public SaaS CSP should offer additional interfaces for the purpose of compliance, forensics, operations and security matters to their customers. Through an API, the customers should have the ability to receive specific information suchas access, error and event logs that could improve their situation in case of aninvestigation. Furthermore, due to the limited ability of receiving forensic information from the server and proofing integrity of stored data in SaaS scenarios, the client has to contribute to this process. This could be achieved by implementing Proofs of Retrievability (POR) in which a verifier (client) is enabled to determine that a prover (server) possesses a file or data object and it can be retrieved unmodified [24]. Provable Data Possession (PDP) techniques [37] could be used to verify that an untrusted server possesses the original data without the need for the client to retrieve it. Although these cryptographic proofs have not been implemented by any CSP, the authors of [23] introduced a new data integrity verification mechanism for SaaS scenarios which could also be used for forensic purposes.2) PaaS Environments: One of the main advantages of the PaaS model is that the developed software application is under the control of the customer and except for some CSP, the source code of the application does not have to leave the local development environment. Given these circumstances, the customer obtains theoretically the power to dictate how the application interacts with other dependencies such as databases, storage entities etc. CSP normally claim this transfer is encrypted but this statement can hardly be verified by the customer. Since the customer has the ability to interact with the platform over a prepared API, system states and specific application logs can be extracted. However potential adversaries, which can compromise the application during runtime, should not be able to alter these log files afterwards. Suggested Solution:Depending on the runtime environment, logging mechanisms could be implemented which automatically sign and encrypt the log information before its transfer to a central logging server under the control of the customer. Additional signing and encrypting could prevent potential eavesdroppers from being able to view and alter log data information on the way to the logging server. Runtime compromise of an PaaS application by adversaries could be monitored by push-only mechanisms for log data presupposing that the needed information to detect such an attack are logged. Increasingly, CSP offering PaaS solutions give developers the ability to collect and store a variety of diagnostics data in a highly configurable way with the help of runtime feature sets [38].3) IaaS Environments: As expected, even virtual instances in the cloud get compromised by adversaries. Hence, the ability to determine how defenses in the virtual environment failed and to what extent the affected systems havebeen compromised is crucial not only for recovering from an incident. Also forensic investigations gain leverage from such information and contribute to resilience against future attacks on the systems. From the forensic point of view, IaaS instances do provide much more evidence data usable for potential forensics than PaaS and SaaS models do. This fact is caused throughthe ability of the customer to install and set up the image for forensic purposes before an incident occurs. Hence, as proposed for PaaS environments, log data and other forensic evidence information could be signed and encrypted before itis transferred to third-party hosts mitigating the chance that a maliciously motivated shutdown process destroys the volatile data. Although, IaaS environments provide plenty of potential evidence, it has to be emphasized that the customer VM is in the end still under the control of the CSP. He controls the hypervisor which is e.g. responsible for enforcing hardware boundaries and routing hardware requests among different VM. Hence, besides the security responsibilities of the hypervisor, he exerts tremendous control over how customer’s VM communicate with the hardware and theoretically can intervene executed processes on the hosted virtual instance through virtual introspection [25]. This could also affect encryption or signing processes executed on the VM and therefore leading to the leakage of the secret key. Although this risk can be disregarded in most of the cases, the impact on the security of high security environments is tremendous.a) Snapshot Analysis: Traditional forensics expect target machines to be powered down to collect an image (dead virtual instance). This situation completely changed with the advent of the snapshot technology which is supported by all popular hypervisors such as Xen, VMware ESX and Hyper-V.A snapshot, also referred to as the forensic image of a VM, providesa powerful tool with which a virtual instance can be clonedby one click including also the running system’s mem ory. Due to the invention of the snapshot technology, systems hosting crucial business processes do not have to be powered down for forensic investigation purposes. The investigator simply creates and loads a snapshot of the target VM for analysis(live virtual instance). This behavior is especially important for scenarios in which a downtime of a system is not feasible or practical due to existing SLA. However the information whether the machine is running or has been properly powered down is crucial [3] for the investigation. Live investigations of running virtual instances become more common providing evidence data that。

云计算英文论文

云计算英文论文

云计算英文论文Cloud computing1. IntroductionCloud computing is known as the most anticipated technological revolution over the entire world, and the reason why cloud computing has brought widespread attention is that it not only represent a new technology appeared, but also lead to the entire industry to change. Therefore, the ranking of national competitiveness will change accordingly.2. Cloud computing definition and applicationThe concept of cloud computing is put forward by Google, which is a beautiful network applications mode. Based on increasing in Internet-related services and delivery models, it provides dynamic, easy to expand and virtualized resources usually through the Internet. This service can be related to IT, software and Internet; other services can be supported too. It means that computing power can be circulation as a commodity. As we expect, more and more enterprise use this new technology into the practical application, such as webmail, Gmail and apple store use it for computing and storage. With the technology development, cloud computing has given full play to it roles, Industry, Newcomers and Media buy computer technology to make network cheaper, faster, convenient and better control.Cloud computing is used in the distributed computers by computing, rather than the local computer or remote server, the operation of the enterprise data center and the Internet are more similar. This process allows companies to be able to switch resources to the urgent application according to computer and storage systems demand access. This means that the computingpower can be negotiable, like water and electricity, access convenient andlow-cost. The biggest difference is that it is transmitted via the Internet.3. AdvantageBy providing a large and secure data storage centers for cloud computing, users do not have to worry about their own data which be lost for some reason or destroyed by computer virus invasion, such as some data stored in the hard disk data will be lost by the computer damaged and virus invasion, so that users cannot access the data and recover data, another disadvantage may also appear when other people use your computer steal the user's computer, such as personal confidential information and business data loss. “Gate of blue photos " is a typical insecure example. If user uploads the photos through the Interne into data storage center, it may have less chance to access personal confidential information. And with cloud computing development, for the user, these services are free of charge, for the future, advanced technology will certainly become a cash cow for merchants, and for the country, the computing power is also able to reflect a country technological level and the level of overall national strength.As the user data stored in the cloud storage center, can reduce customer demand for thehardware level, coupled with cloud computing extremely powerful computing capability, it can enable users to call data more convenient and quick if can add the high-speed networks. Users no longer need replace the computer caused by having no enough hard disk space and CPU computing power. Otherwise, users only surf the Internet to access cloud storage center andthen easily access the data.As powerful data storage centers, the most wonderful features is sharing data. No matter computer data or a variety of communication devices such as mobile phones and PDA. When some damaged appeared in your phone, lost or replaced in order to chase the trend of the times and phone, data copy is a tedious thing. However, in another way, you can solve the entire problem through cloud storage center.Cloud computing services will create more millions of jobs. On March 6, a news show that show that Microsoft statement a commissioned research conducted by theworld-renowned market analyst firm IDC. The research show that cloud computing will create nearly 14 million new jobs worldwide in 2015. And also predicted that it can stimulate IT innovation and brought the new revenue which can reach about $ 1.1 trillion, coupled with cloud computing efficiency substantially promote, the organization will increasere-investment and work opportunity. IDC chief research officer , as senior vice president, John F. Gantz said: "For most organizations, there is no doubt that cloud computing will significantly enhance the return on investment and flexibility, reduce investment costs, and bring revenue growth multiplied as well, we usually believe that cloud computing will reduce jobs by mistake, the opposite, it can create a lot of jobs around the world, no matter emerging market, small cities and small businesses will increase employment opportunities and get benefit from cloud computing. "14. DisadvantageWith large storage center, cloud computing strongly urges powerful hardware conditions and safety measures, whichincluding a large storage space and powerful computing capabilities. As today's scientific and technological rapid develop, user increased, hardware capabilities become one of the necessary condition for development, moreover, the cloud should be kept to enhance their computing power and storage space, and endless. Security measures also include two aspects; firstly, it is necessary to ensure that user data not damaged, lost, and stolen by others, which requires a strong IT team to full range of maintenance and strict access management strategies. Some information has been confirmed, the existing cloud computing service provider still cannot guarantee no similar problem occurs. For users, this is a very serious problem. Another security issue is a natural or unnatural disaster, also can cause the storage center damaged, can cause users cannot access.Cloud computing services are available via the Internet, therefore, user usually have high demands for network access speed, although domestic access speeds improve fast, but compared with LAN, it will appear delays, incomparable, In addition, no network, the user will not be able to access cloud services.The rapid rise of Cloud technology, in another perspective, also restricts the speed of development, which also lead to lower demand for high-end devices. It is the basic reason1Reference[1]why can restrict terminal development. Customer needs, determine the business requirements for products, if customers reduce the terminal's hardware and software requirements, the business will be greatly reduced the degree of product development and restricted the pace of the terminal technology development.5. Cloud computing devel opment in chinaIn China, Cloud computing services have been in a state of rapid development. Looking at the cloud computing industry, government accounted for a major position, and in the market they have also been given the affirmation of the cloud computing. The cloud computing services major support for government agencies and enterprises, therefore, the personal cloud computing services in China is a potential big cake and wait for people enjoying. However because of the domestic personal network speed upgrade too slow, and relevant government network supervision, the development of personal cloud computing services will have more difficult. Due to excessive speculation in the international aspects, in our country, a lot of companies want to get some benefit from the rapid development of technology, exploit cloud computing services projects and promote enterprise development. My personal view is that enterprises should do some preparation, judge it’s good or bad and prospects, as well as estimate the strength of the company and do their best. Through access to information discovery, cloud computing as a technology trend has been got the world's attention, but look at the forefront of foreign enterprise closely, we can easily find almost no companies start the business rely on cloud computing; cloud computing applications and related research are also mostly implement by those large enterprise with more funds. As the prospects for development, for the enterprise, profit is the most critical issues.Although cloud computing has been integrated into our lives, but the domestic development of cloud computing is still confronted with many problems, such as lack of user awareness, migration risk, lack of standards, user locks, security and dataprivacy, safety, standards, data privacy has become a topic of which we are most concerned.Nowadays, people still lack of comprehensive, systematic cognitive of Cloud computing, part of them just imitate completely others and foreign experience, ignoring the specific conditions in our country, resulting in spending much money, but failed to alleviate the complex IT issues. In the original data center, hardware is relatively independent, but migrates to cloud computing data center, systematic assessment and scientific analysis must be essential.Otherwise it may lead to the hardware platform not achieve the proper effect and even the collapse of the application system. Cloud computing products is quite diversity, but the birth of the cloud standard is extremely difficult, the major manufacturers just have their own standards, the Government is also involved, both of them are striving to be able to dominate the major position, so more efforts are still needed. The other faced challenge is how to provide users legitimate services is also very important, except the risks in the systems. Compared with traditional data centers, it cloud provided more services diversity, which also bring more difficult to control. Therefore, analyze the needs of users, reasonable, enforceable service level agreements (SLAs) provided will help users establish confidence in cloud computing services in a large extent. Security problems can be said that a key factor in cloud computing landing. Amazon failure made people deep in thought, which is just one factor of the security.The other security risks include user privacy protection, data sovereignty, disaster recovery, and even illegal interests caused by some hackers. In the era of cloud computing, there will be more user privacy data placed in the clouds, more data, and morevalue. Most importantly, lack of user’s trust will lead to cloud computing landing.Although the development road is very far away, but the Government has been giving support and encouragement. In 2012 Cloud Computing Industry Development Forum was successfully held, which is strong evidence. At the meeting, as a CCID Consulting, cloud industry observers, as well as excellence forerunner of the cloud computing industry, Microsoft, with their co-operation, complete the first "China Cloud economic development Report”, which shows that in-depth interpretation the content of cloud formation and specify the key factor of economic development in the cloud. The result obtained based on thelong-term observation and in-depth research on cloud computing industry. In addition, the participants share their own point about industry environment, market situation, technology development, application service elements and so on; they also can come up with any question about the cloud computing, expects will do their best to give analysis and authoritative view.26. Future prospectsAbout cloud computing services future prospects, in my opinion, it will have great potential, we can find out the answer from the action of the giants of the major IT companies, but now it still in a state of excessive speculation. Many small companies have no any knowledge and experience about it, and still want to exploit the services project. I don’t think it is a sagacious decision. in fact, for current state, I maintained a cautious attitude, as a corporate without strong financial strength, I think they should focus on business development and growth, in order to reduce unnecessary waste of resources, they can develop and useit when the technology get mature stage.Finally, have to say that the future of the cloud computing, it not only can provide unlimited possibilities for storage, read, and manage data using the network. But also provides us infinite computing power because of its huge storage space, there is no powerful computing capabilities, it cannot provide users with easy and fast. In one word, the development potential of cloud computing is endless.7. Reference[1] Microsoft news, Cloud computing services will create more millions of jobs.2012-03-06 2Reference[2]。

云计算外文文献+翻译

云计算外文文献+翻译

云计算外文文献+翻译1. 引言云计算是一种基于互联网的计算方式,它通过共享的计算资源提供各种服务。

随着云计算的普及和应用,许多研究者对该领域进行了深入的研究。

本文将介绍一篇外文文献,探讨云计算的相关内容,并提供相应的翻译。

2. 外文文献概述作者:Antonio Fernández Anta, Chryssis Georgiou, Evangelos Kranakis出版年份:2019年该外文文献主要综述了云计算的发展和应用。

文中介绍了云计算的基本概念,包括云计算的特点、架构、服务模型以及云计算的挑战和前景。

3. 研究内容该研究综述了云计算技术的基本概念和相关技术。

文中首先介绍了云计算的定义和其与传统计算的比较,深入探讨了云计算的优势和不足之处。

随后,文中介绍了云计算的架构,包括云服务提供商、云服务消费者和云服务的基本组件。

在架构介绍之后,文中提供了云计算的三种服务模型:基础设施即服务(IaaS)、平台即服务(PaaS)和软件即服务(SaaS)。

每种服务模型都从定义、特点和应用案例方面进行了介绍,并为读者提供了更深入的了解。

此外,文中还讨论了云计算的挑战,包括安全性、隐私保护、性能和可靠性等方面的问题。

同时,文中也探讨了云计算的前景和未来发展方向。

4. 文献翻译《云计算:一项调查》是一篇全面介绍云计算的文献。

它详细解释了云计算的定义、架构和服务模型,并探讨了其优势、不足和挑战。

此外,该文献还对云计算的未来发展进行了预测。

对于研究云计算和相关领域的读者来说,该文献提供了一个很好的参考资源。

它可以帮助读者了解云计算的基本概念、架构和服务模型,也可以引导读者思考云计算面临的挑战和应对方法。

5. 结论。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

文献信息:文献标题:Integration of Cloud Computing with Internet of Things: Challenges and Open Issues(云计算与物联网的集成:挑战与开放问题)国外作者:HF Atlam等人文献出处:《IEEE International Conference on Internet of Things》, 2017字数统计:英文4176单词,23870字符;中文7457汉字外文文献:Integration of Cloud Computing with Internet of Things:Challenges and Open IssuesAbstract The Internet of Things (IoT) is becoming the next Internet-related revolution. It allows billions of devices to be connected and communicate with each other to share information that improves the quality of our daily lives. On the other hand, Cloud Computing provides on-demand, convenient and scalable network access which makes it possible to share computing resources; indeed, this, in turn, enables dynamic data integration from various data sources. There are many issues standing in the way of the successful implementation of both Cloud and IoT. The integration of Cloud Computing with the IoT is the most effective way on which to overcome these issues. The vast number of resources available on the Cloud can be extremely beneficial for the IoT, while the Cloud can gain more publicity to improve its limitations with real world objects in a more dynamic and distributed manner. This paper provides an overview of the integration of the Cloud into the IoT by highlighting the integration benefits and implementation challenges. Discussion will also focus on the architecture of the resultant Cloud-based IoT paradigm and its new applications scenarios. Finally, open issues and future research directions are also suggested.Keywords: Cloud Computing, Internet of Things, Cloud based IoT, Integration.I.INTRODUCTIONIt is important to explore the common features of the technologies involved in the field of computing. Indeed, this is certainly the case with Cloud Computing and the Internet of Things (IoT) – two paradigms which share many common features. The integration of these numerous concepts may facilitate and improve these technologies. Cloud computing has altered the way in which technologies can be accessed, managed and delivered. It is widely agreed that Cloud computing can be used for utility services in the future. Although many consider Cloud computing to be a new technology, it has, in actual fact, been involved in and encompassed various technologies such as grid, utility computing virtualisation, networking and software services. Cloud computing provides services which make it possible to share computing resources across the Internet. As such, it is not surprising that the origins of Cloud technologies lie in grid, utility computing virtualisation, networking and software services, as well as distributed computing, and parallel computing. On the other hand, the IoT can be considered both a dynamic and global networked infrastructure that manages self-configuring objects in a highly intelligent way. The IoT is moving towards a phase where all items around us will be connected to the Internet and will have the ability to interact with minimum human effort. The IoT normally includes a number of objects with limited storage and computing capacity. It could well be said that Cloud computing and the IoT will be the future of the Internet and next-generation technologies. However, Cloud services are dependent on service providers which are extremely interoperable, while IoT technologies are based on diversity rather than interoperability.This paper provides an overview of the integration of Cloud Computing into the IoT; this involves an examination of the benefits resulting from the integration process and the implementation challenges encountered. Open issues and research directions are also discussed. The remainder of the paper is organised as follows: Section II provides the basic concepts of Cloud computing, IoT, and Cloud-based IoT; SectionIII discusses the benefits of integrating the IoT into the Cloud; Could-based IoT Architecture is presented in section IV; Section V illustrates different Cloud-based IoT applications scenarios. Following this, the challenges facing Cloud-based IoT integration and open research directions are discussed in Section VI and Section VII respectively, before Section VIII concludes the paper.II.BASIC CONCEPTSThis section reviews the basic concepts of Cloud Computing, the IoT, and Cloud-based IoT.1.Cloud ComputingThere exist a number of proposed definitions for Cloud computing, although the most widely agreed upon seems be that put forth by the National Institute of Standards and Technology (NIST). Indeed, the NIST has defined Cloud computing as "a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction".As stated in this definition, Cloud computing comprises four types of deployment models, three different service models, and five essential characteristics.Cloud computing deployment models are most commonly classified as belonging to the public Cloud, where resources are made available to consumers over the Internet. Public Clouds are generally owned by a profitable organisation (e.g. Amazon EC2). Conversely, the infrastructure of a private Cloud is commonly provided by a single organisation to serve the particular purposes of its users. The private Cloud offers a secure environment and a higher level of control (Microsoft Private Cloud). Hybrid Clouds are a mixture of private and public Clouds. This choice is provided for consumers as it makes it possible to overcome some of the limitations of each model. In contrast, a community Cloud is a Cloud infrastructure which is delivered to a group of users by a number of organisations which share the same need.In order to allow consumers to choose the service that suits them, services inCloud computing are provided at three different levels, namely: the Software as a Service (SaaS) model, where software is delivered through the Internet to users (e.g. GoogleApps); the Platform as a Service (PaaS) model, which offers a higher level of integrated environment that can build, test, and deploy specific software (e.g. Microsoft Azure); and finally, with the Infrastructure as a Service (IaaS) model, infrastructure such as storage, hardware and servers are delivered as a service (e.g. Amazon Web Services).2.Internet of ThingsThe IoT represents a modern approach where boundaries between real and digital domains are progressively eliminated by consistently changing every physical device to a smart alternative ready to provide smart services. All things in the IoT (smart devices, sensors, etc.) have their own identity. They are combined to form the communication network and will become actively participating objects. These objects include not only daily usable electronic devices, but also things like food, clothing, materials, parts, and subassemblies; commodities and luxury items; monuments and landmarks; and various forms of commerce and culture. In addition, these objects are able to create requests and alter their states. Thus, all IoT devices can be monitored, tracked and counted, which significantly decreases waste, loss, and cost.The concept of the IoT was first mentioned by Kevin Ashton in 1999, when he stated that “The Internet of Things has the potential to change the world, just as the Internet did. Maybe even more so”. Later, the IoT was formally presented by the International Telecommunication Union (ITU) in 2005. A great many definitions of the IoT have been put forth by numerous organisations and researchers. According to the ITU (2012), the IoT is “a global infrastructure for the Information Society, enabling advanced services by interconnecting (physical and virtual) things based on, existing and evolving, interoperable information and communication technologies”. The IoT introduces a variety of opportunities and applications. However, it faces many challenges which could potentially hinder its successful implementation, such as data storage, heterogeneous resource-constrained, scalability, Things, variable geospatial deployment, and energy efficiency.3.Cloud-Based Internet of ThingsThe IoT and Cloud computing are both rapidly developing services, and have their own unique characteristics. On the one hand, the IoT approach is based on smart devices which intercommunicate in a global network and dynamic infrastructure. It enables ubiquitous computing scenarios. The IoT is typically characterised by widely-distributed devices with limited processing capabilities and storage. These devices encounter issues regarding performance, reliability, privacy, and security. On the other hand, Cloud computing comprises a massive network with unlimited storage capabilities and computation power. Furthermore, it provides a flexible, robust environment which allows for dynamic data integration from various data sources. Cloud computing has partially resolved most of the IoT issues. Indeed, the IoT and Cloud are two comparatively challenging technologies, and are being combined in order to change the current and future environment of internetworking services.The Cloud-based Internet of Things is a platform which allows for the smart usage of applications, information, and infrastructure in a cost-effective way. While the IoT and Cloud computing are different from each other, their features are almost complementary, as shown in TABLE 1. This complementarity is the primary reason why many researchers have proposed their integration.TABLE 1. COMPARISON OF THE IOT WITH CLOUD COMPUTINGIII.BENEFITS OF INTEGRATING IOT WITH CLOUDSince the IoT suffers from limited capabilities in terms of processing power and storage, it must also contend with issues such as performance, security, privacy, reliability. The integration of the IoT into the Cloud is certainly the best way toovercome most of these issues. The Cloud can even benefit from the IoT by expanding its limits with real world objects in a more dynamic and distributed way, and providing new services for billions of devices in different real life scenarios. In addition, the Cloud provides simplicity of use and reduces the cost of the usage of applications and services for end-users. The Cloud also simplifies the flow of the IoT data gathering and processing, and provides quick, low-cost installation and integration for complex data processing and deployment. The benefits of integrating IoT into Cloud are discussed in this section as follows.municationApplication and data sharing are two significant features of the Cloud-based IoT paradigm. Ubiquitous applications can be transmitted through the IoT, whilst automation can be utilised to facilitate low-cost data distribution and collection. The Cloud is an effective and economical solution which can be used to connect, manage, and track anything by using built-in apps and customised portals. The availability of fast systems facilitates dynamic monitoring and remote objects control, as well as data real-time access. It is worth declaring that, although the Cloud can greatly develop and facilitate the IoT interconnection, it still has weaknesses in certain areas. Thus, practical restrictions can appear when an enormous amount of data needs to be transferred from the Internet to the Cloud.2.StorageAs the IoT can be used on billions of devices, it comprises a huge number of information sources, which generate an enormous amount of semi-structured or non-structured data. This is known as Big Data, and has three characteristics: variety (e.g. data types), velocity (e.g. data generation frequency), and volume (e.g. data size). The Cloud is considered to be one of the most cost-effective and suitable solutions when it comes to dealing with the enormous amount of data created by the IoT. Moreover, it produces new chances for data integration, aggregation, and sharing with third parties.3.Processing capabilitiesIoT devices are characterised by limited processing capabilities which preventon-site and complex data processing. Instead, gathered data is transferred to nodes that have high capabilities; indeed, it is here that aggregation and processing are accomplished. However, achieving scalability remains a challenge without an appropriate underlying infrastructure. Offering a solution, the Cloud provides unlimited virtual processing capabilities and an on-demand usage model. Predictive algorithms and data-driven decisions making can be integrated into the IoT in order to increase revenue and reduce risks at a lower cost.4.ScopeWith billions of users communicating with one another together and a variety of information being collected, the world is quickly moving towards the Internet of Everything (IoE) realm - a network of networks with billions of things that generate new chances and risks. The Cloud-based IoT approach provides new applications and services based on the expansion of the Cloud through the IoT objects, which in turn allows the Cloud to work with a number of new real world scenarios, and leads to the emergence of new services.5.New abilitiesThe IoT is characterised by the heterogeneity of its devices, protocols, and technologies. Hence, reliability, scalability, interoperability, security, availability and efficiency can be very hard to achieve. Integrating IoT into the Cloud resolves most of these issues. It provides other features such as ease- of-use and ease-of-access, with low deployment costs.6.New ModelsCloud-based IoT integration empowers new scenarios for smart objects, applications, and services. Some of the new models are listed as follows:•SaaS (Sensing as a Service), which allows access to sensor data;•EaaS (Ethernet as a Service), the main role of which is to provide ubiquitous connectivity to control remote devices;•SAaaS (Sensing and Actuation as a Service), which provides control logics automatically;•IPMaaS (Identity and Policy Management as a Service), which provides access to policy and identity management;•DBaaS (Database as a Service), which provides ubiquitous database management;•SEaaS (Sensor Event as a Service), which dispatches messaging services that are generated by sensor events;•SenaaS (Sensor as a Service), which provides management for remote sensors;•DaaS (Data as a Service), which provides ubiquitous access to any type of data.IV.CLOUD-BASED IOT ARCHITECTUREAccording to a number of previous studies, the well-known IoT architecture is typically divided into three different layers: application, perception and network layer. Most assume that the network layer is the Cloud layer, which realises the Cloud-based IoT architecture, as depicted in Fig. 1.Fig. 1. Cloud-based IoT architectureThe perception layer is used to identify objects and gather data, which is collected from the surrounding environment. In contrast, the main objective of the network layer is to transfer the collected data to the Internet/Cloud. Finally, the application layer provides the interface to different services.V.CLOUD-BASED IOT APPLICATIONSThe Cloud-based IoT approach has introduced a number of applications and smart services, which have affected end users’ daily lives. TABLE 2 presents a brief discussion of certain applications which have been improved by the Cloud-based IoT paradigm.TABLE 2. CLOUD-BASED IOT APPLICATIONSVI.CHALLENGES FACING CLOUD-BASED IOT INTEGRATION There are many challenges which could potentially prevent the successful integration of the Cloud-based IoT paradigm. These challenges include:1.Security and privacyCloud-based IoT makes it possible to transport data from the real world to theCloud. Indeed, one particularly important issues which has not yet been resolved is how to provide appropriate authorisation rules and policies while ensuring that only authorised users have access to the sensitive data; this is crucial when it comes to preserving users’ privacy, and particularly when data integrity must be guaranteed. In addition, when critical IoT applications move into the Cloud, issues arise because of the lack of trust in the service provider, information regarding service level agreements (SLAs), and the physical location of data. Sensitive information leakage can also occur due to the multi-tenancy. Moreover, public key cryptography cannot be applied to all layers because of the processing power constraints imposed by IoT objects. New challenges also require specific attention; for example, the distributed system is exposed to number of possible attacks, such as SQL injection, session riding, cross- site scripting, and side-channel. Moreover, important vulnerabilities, including session hijacking and virtual machine escape are also problematic.2.HeterogeneityOne particularly important challenge faced by the Cloud- based IoT approach is related to the extensive heterogeneity of devices, platforms, operating systems, and services that exist and might be used for new or developed applications. Cloud platforms suffer from heterogeneity issues; for instance, Cloud services generally come with proprietary interfaces, thus allowing for resource integration based on specific providers. In addition, the heterogeneity challenge can be exacerbated when end-users adopt multi-Cloud approaches, and thus services will depend on multiple providers to improve application performance and resilience.3.Big dataWith many predicting that Big Data will reach 50 billion IoT devices by 2020, it is important to pay more attention to the transportation, access, storage and processing of the enormous amount of data which will be produced. Indeed, given recent technological developments, it is clear that the IoT will be one of the core sources of big data, and that the Cloud can facilitate the storage of this data for a long period of time, in addition to subjecting it to complex analysis. Handling the huge amount of data produced is a significant issue, as the application’s whole performance is heavilyreliant on the properties of this data management service. Finding a perfect data management solution which will allow the Cloud to manage massive amounts of data is still a big issue. Furthermore, data integrity is a vital element, not only because of its effect on the service’s quality, but also because of security and privacy issues, the majority of which relate to outsourced data.4.PerformanceTransferring the huge amount of data created from IoT devices to the Cloud requires high bandwidth. As a result, the key issue is obtaining adequate network performance in order to transfer data to Cloud environments; indeed, this is because broadband growth is not keeping pace with storage and computation evolution. In a number of scenarios, services and data provision should be achieved with high reactivity. This is because timeliness might be affected by unpredictable matters and real-time applications are very sensitive to performance efficiency.5.Legal aspectsLegal aspects have been very significant in recent research concerning certain applications. For instance, service providers must adapt to various international regulations. On the other hand, users should give donations in order to contribute to data collection.6.MonitoringMonitoring is a primary action in Cloud Computing when it comes to performance, managing resources, capacity planning, security, SLAs, and for troubleshooting. As a result, the Cloud-based IoT approach inherits the same monitoring demands from the Cloud, although there are still some related challenges that are impacted by velocity, volume, and variety characteristics of the IoT.rge scaleThe Cloud-based IoT paradigm makes it possible to design new applications that aim to integrate and analyse data coming from the real world into IoT objects. This requires interacting with billions of devices which are distributed throughout many areas. The large scale of the resulting systems raises many new issues that are difficult to overcome. For instance, achieving computational capability and storage capacityrequirements is becoming difficult. Moreover, the monitoring process has made the distribution of the IoT devices more difficult, as IoT devices have to face connectivity issues and latency dynamics.VII.OPEN ISSUES AND RESEARCH DIRECTIONSThis section will address some of the open issues and future research directions related to Cloud-based IoT, and which still require more research efforts. These issues include:1.StandardisationMany studies have highlighted the issues of lack of standards, which is considered critical in relation to the Cloud- based IoT paradigm. Although a number of proposed standardisations have been put forth by the scientific society for the deployment of IoT and Cloud approaches, it is obvious that architectures, standard protocols, and APIs are required to allow for interconnection between heterogeneous smart things and the generation of new services, which make up the Cloud- based IoT paradigm.2.Fog ComputingFog computing is a model which extends Cloud computing services to the edge of the network. Similar to the Cloud, Fog supply communicates application services to users. Fog can essentially be considered an extension of Cloud Computing which acts as an intermediate between the edge of the network and the Cloud; indeed, it works with latency-sensitive applications that require other nodes to satisfy their delay requirements. Although storage, computing, and networking are the main resources of both Fog and the Cloud, the Fog has certain features, such as location awareness and edge location, that provide geographical distribution, and low latency; moreover, there are a large nodes; this is in contrast with the Cloud, which is supported for real-time interaction and mobility.3.Cloud CapabilitiesAs in any networked environment, security is considered to be one of the main issues of the Cloud-based IoT paradigm. There are more chances of attacks on boththe IoT and the Cloud side. In the IoT context, data integrity, confidentiality and authenticity can be guaranteed by encryption. However, insider attacks cannot be resolved and it is also hard to use the IoT on devices with limited capabilities.4.SLA enforcementCloud-based IoT users need created data to be conveyed and processed based on application-dependent limitations, which can be tough in some cases. Ensuring a specific Quality of Service (QoS) level regarding Cloud resources by depending on a single provider raises many issues. Thus, multiple Cloud providers may be required to avoid SLA violations. However, dynamically choosing the most appropriate mixture of Cloud providers still represents an open issue due to time, costs, and heterogeneity of QoS management support.5.Big dataIn the previous section, we discussed Big Data as a critical challenge that is tightly coupled with the Cloud-based IoT paradigm. Although a number of contributions have been proposed, Big Data is still considered a critical open issue, and one in need of more research. The Cloud-based IoT approach involves the management and processing of huge amounts of data stemming from various locations and from heterogeneous sources; indeed, in the Cloud-based IoT, many applications need complicated tasks to be performed in real-time.6.Energy efficiencyRecent Cloud-based IoT applications include frequent data that is transmitted from IoT objects to the Cloud, which quickly consumes the node energy. Thus, generating efficient energy when it comes to data processing and transmission remains a significant open issue. Several directions have been suggested to overcome this issue, such as compression technologies, efficient data transmission; and data caching techniques for reusing collected data with time-insensitive application.7.Security and privacyAlthough security and privacy are both critical research issues which have received a great deal of attention, they are still open issues which require more efforts. Indeed, adapting to different threats from hackers is still an issue. Moreover, anotherproblem is providing the appropriate authorisation rules and policies while ensuring that only authorised users have access to sensitive data; this is crucial for preserving users’ privacy, specifically when data integrity must be guaranteed.VIII.CONCLUSIONThe IoT is becoming an increasingly ubiquitous computing service which requires huge volumes of data storage and processing capabilities. The IoT has limited capabilities in terms of processing power and storage, while there also exist consequential issues such as security, privacy, performance, and reliability; As such, the integration of the Cloud into the IoT is very beneficial in terms of overcoming these challenges. In this paper, we presented the need for the creation of the Cloud-based IoT approach. Discussion also focused on the Cloud-based IoT architecture, different applications scenarios, challenges facing the successful integration, and open research directions. In future work, a number of case studies will be carried out to test the effectiveness of the Cloud-based IoT approach in healthcare applications.中文译文:云计算与物联网的集成:挑战与开放问题摘要物联网(IoT)正在成为下一次与互联网相关的革命。

相关文档
最新文档