Data-intensive applications, challenges, techniques(A)(1)
Microsoft

Access
vs. Univeaker Microsoft Corporation One Microsoft Way Redmond, WA 98052-6399
billbak@
Abstract
Modern data-intensive applications require the integration of information stored not only in traditional database management systems, but also in file systems, indexed-sequential files, desktop databases, spread sheets, project management tools, electronic mail, directory services, multimedia data stores, spatial data stores, and more. Several database companies are predictably pursuing a traditional database-centric approach generally called universal storage. In this approach, the database vendor extends the database engine and programming interface to support. new data types, including text, spatial, video and audio. They require their customer to move all interesting data inside the corporation into their database implementation. Universal access is an interesting and exciting alternative to the universal storage approach. The key to universal access is to allow applications to access data where it lives without replication, transformation, conversion or delay. Industry standard interfaces allow any-to-any data Independent engines provide for cross-data-source transaction control connectivity. (heterogeneous two-phase commit) and content indexing. The universal access approach attempts to be data source and tool agnostic. This presentation will cover the wide range of corporate data, the data access requirements of fast-moving, competitive corporations, the advantages and disadvantages of the universal storage strategy, the advantages and disadvantages of the universal access approach, key technical requirements on data clients and data providers to participate in universal access and details of an existing universal access implementation. The presentation builds heavily on the work of Vaskevjtch [I] and Blakeley [2].
未来人工智能的发展趋势英语作文

未来人工智能的发展趋势英语作文全文共3篇示例,供读者参考篇1The Future of AI: Exciting Possibilities and Potential PitfallsArtificial Intelligence (AI) has already transformed our world in countless ways, from the smart assistants on our phones to the recommendation algorithms that power our favorite streaming services. However, the AI revolution is still in its early stages, and the future of this technology promises to be even more profound and disruptive. As a student fascinated by the rapid advancements in AI, I can't help but wonder what the future might hold for this powerful tool.One of the most exciting prospects of future AI development is the potential for significant breakthroughs in fields like healthcare and scientific research. AI systems are already being used to analyze vast amounts of data, identify patterns and make predictions that would be impossible for human minds alone. In the medical field, AI could revolutionize disease diagnosis, drug discovery, and personalized treatment plans. By processing millions of patient records, genome sequences, and scientificstudies, AI could uncover hidden correlations and insights that lead to new cures and better patient outcomes.Furthermore, AI's ability to process and analyze data at an unprecedented scale could accelerate scientific progress across numerous disciplines, from astrophysics to climate science. Researchers could leverage AI to sift through vast datasets, test hypotheses, and uncover new theories and models that explain the complexities of the natural world. The potential for AI to aid in tackling global challenges such as climate change, energy sustainability, and food security is truly remarkable.Another area where AI is poised to have a profound impact is in the realm of automation and robotics. As AI systems become more advanced and capable, they could take on an ever-increasing range of tasks, from manufacturing and logistics to service industries and even creative endeavors like writing and art. While this raises concerns about job displacement and the future of work, it also presents opportunities for increased productivity, efficiency, and potentially higher living standards.However, the widespread adoption of AI also raises significant ethical and societal concerns that must be carefully considered. One of the most pressing issues is the potential for AI systems to perpetuate or amplify existing biases anddiscrimination, particularly if the training data used to develop these systems reflects human prejudices. There is a risk that AI could reinforce societal inequalities and marginalize certain groups if proper safeguards and checks are not put in place.Additionally, as AI becomes more sophisticated and autonomous, there are valid concerns about the potential for these systems to be misused or cause unintended harm. The development of advanced AI systems capable of making independent decisions raises questions about accountability, transparency, and the need for robust ethical frameworks to govern their use.篇2The Future of AI: Accelerating Progress and Profound ImpactsArtificial intelligence (AI) is one of the most transformative and rapidly evolving technologies of our time. As a student witnessing the breathtaking pace of innovation in this field, I am both awed and somewhat apprehensive about the future trajectory of AI and its potential implications for humanity.In recent years, we have seen remarkable breakthroughs in AI, ranging from natural language processing and computervision to game-playing systems that can outperform humans in complex strategy games like chess and Go. The rise of deep learning and neural networks has been a game-changer, enabling machines to learn and adapt in ways that were previously unimaginable.Looking ahead, the development of AI is likely to accelerate even further, driven by several key trends and advancements:Increasing computational power: As we continue to make strides in hardware development, particularly in areas like quantum computing and specialized AI chips, machines will gain unprecedented computational capabilities. This will enable more complex and data-intensive AI models to be trained and deployed, unlocking new frontiers in performance and functionality.Availability of massive datasets: The exponential growth of data generated by humans and machines alike is fueling the development of AI systems. With access to vast repositories of information across various domains, AI algorithms can continue to learn and refine their abilities, becoming increasingly accurate and versatile.Advancements in algorithmic techniques: Researchers and engineers are constantly pushing the boundaries of AIalgorithms, exploring new architectures and methods for training and optimizing models. Techniques like reinforcement learning, generative adversarial networks (GANs), and transfer learning are just a few examples of the cutting-edge approaches that are driving AI forward.Democratization of AI: As AI technologies become more accessible and user-friendly, we are likely to see a proliferation of AI applications across various sectors and industries. This democratization will empower individuals, small businesses, and organizations to leverage the power of AI, fostering innovation and creating new opportunities.While these trends paint an exciting picture of AI's future potential, they also raise important ethical and societal concerns that must be carefully addressed. Some key areas of concern include:Job displacement and economic disruption: As AI systems become more capable and pervasive, there is a risk of widespread job displacement, particularly in industries and roles that are susceptible to automation. This could lead to significant economic disruption and exacerbate existing inequalities if not managed properly.Privacy and security risks: The vast amounts of data required to train AI systems and the potential for AI to be used for surveillance and monitoring purposes raise serious privacy and security concerns. Robust governance frameworks and ethical guidelines will be crucial to mitigate these risks.Algorithmic bias and fairness: AI systems can perpetuate and amplify existing biases present in the data they are trained on, leading to unfair and discriminatory outcomes. Ensuring algorithmic fairness and accountability will be a crucial challenge as AI becomes more pervasive.Existential risk: While perhaps a more distant and speculative concern, some experts have warned about the potential existential risks posed by advanced AI systems that surpass human intelligence and capabilities, potentially leading to unintended consequences or even direct threats to humanity.As a student passionate about the potential of AI, I believe that addressing these ethical and societal concerns should be a top priority alongside technical advancements. We must foster interdisciplinary collaboration between AI researchers, ethicists, policymakers, and stakeholders from various sectors to develop responsible and inclusive AI governance frameworks.Moreover, education and public awareness about AI will be crucial in preparing society for the transformative impacts of this technology. As students, we must strive to develop awell-rounded understanding of AI, its capabilities, limitations, and ethical implications, to ensure that we can navigate this rapidly evolving landscape responsibly and effectively.In conclusion, the future of AI is poised for remarkable progress, driven by advancements in computational power, data availability, algorithmic techniques, and democratization. However, this progress must be accompanied by a thoughtful and proactive approach to addressing the ethical and societal challenges that AI presents. By fostering responsible innovation, interdisciplinary collaboration, and public education, we can harness the immense potential of AI while mitigating its risks and ensuring that this technology serves the greater good of humanity.篇3The Future of Artificial Intelligence: Trends and ImplicationsAs a student living in an era where technological advancements are rapidly reshaping our world, the topic of artificial intelligence (AI) has captured my imagination andpiqued my curiosity. AI, a broad field encompassing machine learning, deep learning, and neural networks, has already made its mark across various sectors, from healthcare and finance to entertainment and transportation. However, the future holds even more profound implications as AI continues to evolve and infiltrate every aspect of our lives.One of the most exciting trends in AI development is the pursuit of artificial general intelligence (AGI), also known as strong AI. While current AI systems excel at specific tasks, AGI aims to create machines with the ability to reason, learn, and adapt like humans, across a wide range of domains. Achieving AGI would represent a monumental leap forward, potentially leading to machines that can match or even surpass human intelligence. Researchers are exploring various approaches, including neural networks that mimic the human brain, symbolic logic systems, and hybrid models that combine multiple techniques.Another area of significant progress is the integration of AI into the realm of robotics. Advanced robots equipped with AI systems are already being employed in manufacturing, healthcare, and exploration. As AI capabilities continue to improve, we can expect to see more sophisticated robotscapable of performing complex tasks, navigating unstructured environments, and interacting seamlessly with humans. Robotic assistants, autonomous vehicles, and even robotic companions could become commonplace in the not-too-distant future.The field of natural language processing (NLP) is also poised for remarkable advancements. NLP aims to enable machines to understand, interpret, and generate human language with increasing accuracy and fluency. As NLP technologies mature, we can anticipate more natural and intuitive interactions between humans and machines, potentially revolutionizing industries such as customer service, education, and content creation.Moreover, the convergence of AI with other cutting-edge technologies, such as the Internet of Things (IoT), blockchain, and quantum computing, holds immense potential. AI-powered IoT systems could enable seamless communication and coordination between countless devices, optimizing efficiency and resource utilization. Blockchain technology, combined with AI, could lead to more secure and transparent systems for various applications, ranging from financial transactions to supply chain management. Quantum computing, which harnesses the principles of quantum mechanics, could provide the computational power necessary to tackle complex problemsthat are intractable for classical computers, unlocking new frontiers in AI research and development.While the prospects of AI are undoubtedly exciting, it is crucial to address the ethical and societal implications of this transformative technology. As AI systems become more capable and autonomous, concerns around privacy, security, and accountability arise. Responsible development and governance frameworks are essential to ensure AI is deployed in a manner that aligns with human values and prioritizes the well-being of society.Additionally, the impact of AI on the job market and workforce cannot be overlooked. As AI automates certain tasks and displaces certain roles, there is a pressing need to reskill and adapt the workforce to thrive in an AI-driven economy. Education systems must evolve to equip students with the necessary skills and mindsets to collaborate effectively with AI systems and leverage their capabilities effectively.In conclusion, the future of AI is brimming with both remarkable opportunities and significant challenges. As a student, I am in awe of the potential breakthroughs that lie ahead, from artificial general intelligence and advanced robotics to natural language processing and the convergence of AI withother emerging technologies. However, I also recognize the importance of addressing the ethical, social, and economic implications of AI to ensure its responsible development and deployment. It is our collective responsibility to shape the trajectory of AI in a manner that benefits humanity while mitigating potential risks and unintended consequences. By embracing a holistic and forward-thinking approach, we can harness the transformative power of AI to create a better, more sustainable, and more equitable world for generations to come.。
Wireless-Networks(5)

Wireless NetworksWireless networks have become an integral part of our daily lives, providing us with the convenience and flexibility to stay connected from anywhere. However, they also come with their own set of challenges and problems that can impact their performance and reliability. In this response, I will address some of the common problems associated with wireless networks, including interference, security concerns, and connectivity issues, while also discussing potential solutions and best practices to mitigate these issues. One of the most prevalent issues with wireless networks is interference, which can result from various sources such as other electronic devices, neighboring networks, or physical obstructions. This interference can cause disruptions in the wireless signal, leading to slow or unreliable connections. To address this problem, it is essential to identify the source of interference and take steps to minimize its impact. This can involve repositioning the wireless router, using devices that operate on different frequencies, or investing in signal boosters to improve coverage. Another significant concern with wireless networks is security. With the increasing prevalence of cyber threats and attacks, ensuring the security of a wireless network is crucial. Without proper security measures in place, wireless networks are vulnerable to unauthorized access, data breaches, and other malicious activities. To mitigate these risks, it is essential to implement strong encryption protocols, such as WPA2 or WPA3, and regularly update passwords to prevent unauthorized access. Additionally, enabling features such as network segmentation and guest networks can help enhance the overall security posture of the wireless network. In addition to interference and security issues, connectivity problems are also common in wireless networks. Dead zones, where the wireless signal is weak or nonexistent, can be a significant challenge, especially in larger or multi-story buildings. To address this issue, strategically placing wireless access points and repeaters can help extend the coverage area and improve connectivity in dead zones. Furthermore, utilizing mesh networking technology can create a more robust and seamless wireless network by enabling devices to connect to the nearest access point, thereby reducing connectivity issues. Moreover, the increasing number of connected devices in modern households and workplaces canstrain wireless networks, leading to performance issues and slow speeds. This problem, known as network congestion, can be alleviated by upgrading to a higher bandwidth or utilizing advanced technologies such as MU-MIMO (multi-user, multiple input, multiple output) to enable the router to communicate with multiple devices simultaneously. Additionally, optimizing the placement of wireless devices and minimizing the use of bandwidth-intensive applications can help alleviate network congestion and improve overall performance. Furthermore, the rapid advancement of wireless technology and the proliferation of IoT (Internet of Things) devices have introduced new complexities and challenges for wireless networks. The sheer volume of connected devices, each with its unique requirements and demands, can place a strain on the network infrastructure and impact its performance. To address this issue, implementing network management tools and solutions, such as Quality of Service (QoS) settings and device prioritization, can help optimize the network to accommodate the diverse needs of connected devices. In conclusion, wireless networks offer unparalleled convenience and flexibility, but they also come with their fair share of challenges and problems. Interference, security concerns, connectivity issues, network congestion, and the complexities introduced by IoT devices are all significant issues that can impact the performance and reliability of wireless networks. However, by implementing best practices, such as optimizing the placement of wireless devices, securing the network with robust encryption, and utilizing advanced technologies like mesh networking and MU-MIMO, many of these problems can be mitigated. As wireless technology continues to evolve, it is essential for users and organizations to stay informed about emerging challenges and adopt proactive measures to ensure the optimal performance and security of their wireless networks.。
大数据技术与工程英文介绍范文

大数据技术与工程英文介绍范文In today's digital era, the concept of big data has emerged as a pivotal factor influencing various fields, including business, healthcare, education, and scientific research. Big data refers to the vast volumes of structured and unstructured data that are generated at an unprecedented rate. The ability to process and analyze this data is crucial for organizations aiming to gain insights, drive decision-making, and improve operations.Big data technology encompasses a range of tools and frameworks designed to manage and analyze large datasets. Technologies such as Hadoop, Spark, and NoSQL databases like MongoDB and Cassandra are integral in enabling organizations to store, process, and analyze data efficiently. These technologies provide the infrastructure necessary to handle the three Vs of big data: volume, velocity, and variety. Byleveraging distributed computing and storage, big data technologies allow organizations to scale their data processing capabilities while reducing costs.Moreover, big data engineering is concerned with the creation of architectures and systems that facilitate the processing of data. This includes the design and implementation of data pipelines, data lakes, and data warehouses. Data engineers play a vital role in ensuring that data is collected, processed, and made accessible for analytical purposes. Their expertise in programming, database management, and data modeling is essential for developing solutions that meet the specific needs of an organization.The application of big data analysis is vast and varied. For instance, in healthcare, it helps in predicting disease outbreaks, personalizing treatment plans, and optimizing operational efficiency. In retail, businesses can analyze consumer behavior to enhance customer experiences and drivesales. Furthermore, in the realm of finance, big data analytics detects fraud and assesses risk in real-time.In conclusion, the intersection of big data technology and engineering is transforming the way organizations operate and make decisions. By harnessing the power of big data, businesses can unlock valuable insights that drive innovation and improve their competitive edge in the market. As we continue to generate more data, understanding and implementing effective big data solutions will become increasingly important for future advancements across all sectors.。
scientific data 级别

scientific data 级别Scientific Data: The Key to Unlocking the Secrets of the UniverseIntroduction:Scientific data plays a crucial role in advancing our understanding of the universe. From the depths of the oceans to the vastness of outer space, scientists rely on data to unravel mysteries, make discoveries, and formulate theories. In this article, we will explore the significance of scientific data and how it has revolutionized various fields of research.1. The Importance of Data Collection:Data collection is the foundation of scientific research. It involves gathering information through systematic observation, experimentation, and measurement. By collecting data, scientists can identify patterns, detect trends, and draw meaningful conclusions. Without accurate and reliable data, scientific progress would be stunted.2. Data Analysis and Interpretation:Once data is collected, it needs to be analyzed and interpreted. This involves applying statistical methods,mathematical models, and other analytical tools to extract meaningful insights. Data analysis allows scientists to identify correlations, establish cause-and-effect relationships, and validate or disprove hypotheses. It is through this process that scientific theories are developed and refined.3. Big Data and Advancements in Technology:Advancements in technology have led to the generation of vast amounts of data, commonly referred to as "big data." This influx of data has been a game-changer in various scientific disciplines. Powerful computers and sophisticated algorithms can now process and analyze large datasets, enabling scientists to make breakthroughs in areas such as genomics, climate modeling, and astrophysics. 4. Data Sharing and Collaboration:In today's interconnected world, data sharing and collaboration have become essential for scientific progress. Open access policies and data repositories facilitate the sharing of research findings, allowing scientists from around the globe to build upon each other's work. Collaborative efforts not only enhance the reliability ofscientific data but also accelerate the pace of discoveries.5. Data-driven Decision Making:Scientific data has also revolutionized decision making in various sectors, including healthcare, environmental management, and policy formulation. By analyzing data related to disease patterns, for example, researchers can identify risk factors, develop preventive measures, and improve patient outcomes. Similarly, environmental data helps policymakers make informed decisions about conservation efforts and sustainable development.6. Challenges and Ethical Considerations:While scientific data offers immense opportunities, it also poses challenges and ethical considerations. Data privacy, security, and ownership rights are critical issues that need to be addressed. Additionally, biases in data collection, analysis, and interpretation can lead to skewed results, emphasizing the need for transparency and rigorous peer review.7. The Future of Scientific Data:As technology continues to advance, the future of scientific data looks promising. Artificial intelligence,machine learning, and data visualization techniques are transforming the way data is analyzed and presented. These advancements will further enhance our understanding of complex phenomena and enable scientists to tackle previously unsolvable problems.Conclusion:Scientific data is the lifeblood of research and discovery. It empowers scientists to explore the unknown, make evidence-based decisions, and push the boundaries of knowledge. The continued collection, analysis, and sharing of scientific data will undoubtedly lead to groundbreaking advancements in various fields, ultimately shaping the future of our world.。
5g的发明与应用的英语作文

5g的发明与应用的英语作文The Invention and Applications of 5GThe world of telecommunications has undergone a remarkable transformation in recent years, and the advent of 5G technology has been a game-changer. 5G, or the fifth generation of wireless technology, has revolutionized the way we communicate, interact, and access information. This cutting-edge technology has the potential to unlock a new era of connectivity, revolutionizing various industries and improving the quality of our lives.The journey of 5G began with the recognition of the limitations of the existing wireless technologies. As the demand for faster data speeds, lower latency, and increased connectivity grew, the need for a more advanced and efficient system became apparent. The development of 5G was a collaborative effort involving various stakeholders, including telecommunications companies, research institutions, and government agencies, all working towards a common goal of creating a network that would surpass the capabilities of its predecessors.One of the key features that sets 5G apart is its significantly higherdata transfer rates. The previous generation of wireless technology, 4G, had a maximum download speed of around 1 Gbps. In contrast, 5G boasts download speeds of up to 10 Gbps, with the potential to reach even higher rates in the future. This remarkable improvement in speed translates to a seamless and immersive user experience, enabling the smooth streaming of high-definition videos, the rapid download of large files, and the real-time transmission of data-intensive applications.Another critical aspect of 5G is its reduced latency. Latency refers to the time it takes for data to travel from one point to another. In the case of 4G, the latency was typically around 50 milliseconds, which was already a significant improvement over previous generations. However, 5G takes this even further, with latency as low as 1 millisecond. This near-instantaneous response time is crucial for applications that require real-time interaction, such as remote surgery, autonomous vehicles, and virtual reality experiences.The increased bandwidth and reduced latency of 5G also enable the development of the Internet of Things (IoT) ecosystem. IoT refers to the interconnected network of devices, sensors, and systems that can communicate and exchange data without human intervention. With 5G, the number of devices that can be connected simultaneously has increased exponentially, allowing for the seamless integration of smart home appliances, industrial automation systems, and city-wideinfrastructure.One of the most promising applications of 5G is in the realm of autonomous vehicles. The low latency and high reliability of the 5G network are essential for the smooth and safe operation of self-driving cars. These vehicles rely on the real-time exchange of data, such as road conditions, traffic updates, and sensor information, to navigate their surroundings effectively. With 5G, the communication between vehicles and infrastructure becomes more efficient, reducing the risk of accidents and enabling a more seamless transportation experience.Another area where 5G is making a significant impact is in the field of healthcare. The technology's ability to support remote medical services, such as telemedicine and remote patient monitoring, has become increasingly valuable, particularly during the COVID-19 pandemic. Doctors can now conduct virtual consultations, access medical records, and even perform remote surgeries with the help of 5G-enabled devices and applications. This has improved access to healthcare, especially in underserved or remote areas, and has the potential to revolutionize the way we deliver and receive medical services.The potential of 5G extends beyond just personal and medical applications. The technology also has a significant impact on variousindustries, including manufacturing, agriculture, and energy. In the manufacturing sector, 5G can enable the integration of advanced robotics, real-time monitoring, and predictive maintenance, leading to increased efficiency, reduced downtime, and improved product quality. In the agricultural industry, 5G-powered sensors and drones can gather data on soil conditions, crop health, and weather patterns, allowing farmers to make more informed decisions and optimize their operations. In the energy sector, 5G can facilitate the integration of renewable energy sources, smart grids, and efficient energy management systems, contributing to a more sustainable and resilient power infrastructure.However, the deployment of 5G technology is not without its challenges. One of the primary concerns is the issue of network security and privacy. The increased connectivity and data exchange enabled by 5G networks have raised concerns about the potential for cyber threats, data breaches, and unauthorized access to sensitive information. Addressing these security concerns is crucial to ensuring the widespread adoption and trust in 5G technology.Another challenge is the deployment of the necessary infrastructure to support 5G networks. The implementation of 5G requires significant investments in new cell towers, fiber-optic cables, and other supporting infrastructure. This can be particularly challenging in areas with limited resources or geographical barriers, such as ruraland remote regions. Governments and telecommunications companies must work collaboratively to ensure that the benefits of 5G are accessible to all communities, bridging the digital divide and promoting inclusive development.Despite these challenges, the potential of 5G technology is undeniable. As the world continues to evolve and embrace new technologies, the impact of 5G will only become more profound. From transforming the way we communicate and access information to revolutionizing entire industries, 5G has the power to shape the future of our connected world. As we continue to explore and harness the capabilities of this groundbreaking technology, we can look forward to a future where the boundaries of what is possible are continuously expanded, leading to a more efficient, interconnected, and innovative world.。
Data Integration - 数据集成

1.create the source descriptions 2. writing the semantic mappings This was the main bottleneck.
Techniques for Schema Mapping
semi-automatically generating schema mappings Goal: create tools that speed up the creation of the mappings and reduce the amount of human effort involved.
Query Answering in LAV =
Answering queries using views (AQUV)
a problem which was earlier considered in the context of query optimization Given a set of views V1,…,Vn, And a query Q,
informationmanifold?buildingonthefoundation?dataintegrationindustry?futurechallenges?conclusionbuildingonthefoundation?generatingschemamappings?adaptivequeryprocessing?xml?modelmanagement?peertopeerdatamanagement?theroleofartificialintelligencegeneratingschemamappings?lookatthatobservation
Geometric Modeling

Geometric ModelingGeometric modeling is a crucial aspect of computer graphics and design, playing a significant role in various fields such as engineering, architecture, animation, and gaming. It involves the creation and manipulation of geometric shapes and structures in a digital environment, allowing for the visualization and representation of complex objects and scenes. However, despite its importance, geometric modeling presents several challenges and limitations that need to be addressed in order to improve its efficiency and effectiveness. One of the primary issues in geometric modeling is the complexity of representing real-world objects and environments in a digital format. The process of converting physical objects into digital models involves capturing and processing a vast amount of data, which can be time-consuming and resource-intensive. This is particularly challenging when dealing with intricate and irregular shapes, as it requires advanced techniques such as surface reconstruction and mesh generation to accurately capture the details of the object. As a result, geometric modeling often requires a balance between precision and efficiency, as the level of detail in the model directly impacts its computational cost and performance. Another challenge in geometric modeling is the need for seamless integration with other design and simulation tools. In many applications, geometric models are used as a basis for further analysis and manipulation, such as finite element analysis in engineering or physics-based simulations in animation. Therefore, it is essential for geometric modeling software to be compatible with other software and data formats, allowing for the transfer and utilization of geometric models across different platforms. This interoperability is crucial for streamlining the design and production process, as it enables seamless collaboration and data exchange between different teams and disciplines. Furthermore, geometric modeling also faces challenges related to the representation and manipulation of geometric data. Traditional modeling techniques, such as boundary representation (B-rep) and constructive solid geometry (CSG), have limitations in representing complex and organic shapes, often leading to issues such as geometric inaccuracies and topological errors. To address this, advanced modeling techniques such as non-uniform rational B-splines (NURBS) and subdivision surfaces have been developed toprovide more flexible and accurate representations of geometric shapes. However, these techniques also come with their own set of challenges, such as increased computational complexity and difficulty in controlling the shape of the model. In addition to technical challenges, geometric modeling also raises ethical and societal considerations, particularly in the context of digital representation and manipulation. As the boundary between physical and digital reality becomes increasingly blurred, issues such as intellectual property rights, privacy, and authenticity of digital models have become more prominent. For example, the unauthorized use and reproduction of digital models can lead to copyright infringement and legal disputes, highlighting the need for robust mechanisms to protect the intellectual property of digital content creators. Similarly, the rise of deepfakes and digital forgeries has raised concerns about the potential misuse of geometric modeling technology for malicious purposes, such as misinformation and identity theft. It is crucial for the industry to address these ethical concerns and develop standards and regulations to ensure the responsible use of geometric modeling technology. Despite these challenges, the field of geometric modeling continues to evolve and advance, driven by the growing demand forrealistic and interactive digital experiences. Recent developments in machine learning and artificial intelligence have shown promise in addressing some of the technical limitations of geometric modeling, such as automated feature recognition and shape optimization. Furthermore, the increasing availability of powerful hardware and software tools has enabled more efficient and accessible geometric modeling workflows, empowering designers and artists to create intricate and immersive digital content. With ongoing research and innovation, it is likely that many of the current challenges in geometric modeling will be overcome, leading to more sophisticated and versatile tools for digital design and visualization. In conclusion, geometric modeling is a critical component of modern digital design and visualization, enabling the creation and manipulation of complex geometric shapes and structures. However, the field faces several challenges related to the representation, integration, and ethical implications of geometric models. By addressing these challenges through technological innovation and ethical considerations, the industry can continue to push the boundaries of what ispossible in digital design and create more immersive and impactful experiences for users.。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Data-intensive applications,challenges,techniquesand technologies:A survey on BigDataC.L.Philip Chen ⇑,Chun-Yang ZhangDepartment of Computer and Information Science,Faculty of Science and Technology,University of Macau,Macau,Chinaa r t i c l e i n f o Article history:Received 28March 2013Received in revised form 3January 2014Accepted 10January 2014Available online 21January 2014Keywords:Big Data Data-intensive computing e-Science Parallel and distributed computing Cloud computinga b s t r a c tIt is already true that Big Data has drawn huge attention from researchers in informationsciences,policy and decision makers in governments and enterprises.As the speed of infor-mation growth exceeds Moore’s Law at the beginning of this new century,excessive data ismaking great troubles to human beings.However,there are so much potential and highlyuseful values hidden in the huge volume of data.A new scientific paradigm is born as data-intensive scientific discovery (DISD),also known as Big Data problems.A large number offields and sectors,ranging from economic and business activities to public administration,from national security to scientific researches in many areas,involve with Big Data prob-lems.On the one hand,Big Data is extremely valuable to produce productivity in busi-nesses and evolutionary breakthroughs in scientific disciplines,which give us a lot ofopportunities to make great progresses in many fields.There is no doubt that the futurecompetitions in business productivity and technologies will surely converge into the BigData explorations.On the other hand,Big Data also arises with many challenges,such asdifficulties in data capture,data storage,data analysis and data visualization.This paperis aimed to demonstrate a close-up view about Big Data,including Big Data applications,Big Data opportunities and challenges,as well as the state-of-the-art techniques and tech-nologies we currently adopt to deal with the Big Data problems.We also discuss severalunderlying methodologies to handle the data deluge,for example,granular computing,cloud computing,bio-inspired computing,and quantum computing.Ó2014Elsevier Inc.All rights reserved.1.IntroductionBig Data has been one of the current and future research frontiers.In this year,Gartner listed the ‘‘Top 10Strategic Tech-nology Trends For 2013’’[158]and ‘‘Top 10Critical Tech Trends For The Next Five Years’’[157],and Big Data is listed in the both two.It is right to say that Big Data will revolutionize many fields,including business,the scientific research,public administration,and so on.For the definition of the Big Data,there are various different explanations from 3Vs to 4Vs.Doug Laney used volume ,velocity and variety ,known as 3Vs [96],to characterize the concept of Big Data.The term volume is the size of the data set,velocity indicates the speed of data in and out,and variety describes the range of data types and sources.Sometimes,people extend another V according to their special requirements.The fourth V can be value ,variability ,or virtual[207].More commonly,Big Data is a collection of very huge data sets with a great diversity of types so that it becomes dif-ficult to process by using state-of-the-art data processing approaches or traditional data processing platforms.In 2012,Gart-ner retrieved and gave a more detailed definition as:‘‘Big Data are high-volume,high-velocity,and/or high-variety 0020-0255/$-see front matter Ó2014Elsevier Inc.All rights reserved./10.1016/j.ins.2014.01.015⇑Corresponding author.E-mail addresses:Philip.Chen@ (C.L.Philip Chen),cyzhangfst@ (C.-Y.Zhang).information assets that require new forms of processing to enable enhanced decision making,insight discovery and process optimization’’.More generally,a data set can be called Big Data if it is formidable to perform capture,curation,analysis and visualization on it at the current technologies.With diversified data provisions,such as sensor networks,telescopes,scientific experiments,and high throughput instru-ments,the datasets increase at exponential rate [178,110]as demonstrated in Fig.1(source from [67]).The off-the-shelf techniques and technologies that we ready used to store and analyse data cannot work efficiently and satisfactorily.The challenges arise from data capture and data curation to data analysis and data visualization.In many instances,science is legging behind the real world in the capabilities of discovering the valuable knowledge from massive volume of data.Based on precious knowledge,we need to develop and create new techniques and technologies to excavate Big Data and benefit our specified purposes.Big Data has changed the way that we adopt in doing businesses,managements and researches.Data-intensive science especially in data-intensive computing is coming into the world that aims to provide the tools that we need to handle the Big Data problems.Data-intensive science [18]is emerging as the fourth scientific paradigm in terms of the previous three,namely empirical science,theoretical science and computational science.Thousand years ago,scientists describing the nat-ural phenomenon only based on human empirical evidences,so we call the science at that time as empirical science.It is also the beginning of science and classified as the first paradigm.Then,theoretical science emerged hundreds years ago as the second paradigm,such as Newton’s Motion Laws and Kepler’s Laws.However,in terms of many complex phenomenon and problems,scientists have to turn to scientific simulations,since theoretical analysis is highly complicated and some-times unavailable and infeasible.Afterwards,the third science paradigm was born as computational branch.Simulations in large of fields generate a huge volume of data from the experimental science,at the same time,more and more large data sets are generated in many pipelines.There is no doubt that the world of science has changed just because of the increasing data-intensive applications.The techniques and technologies for this kind of data-intensive science are totally distinct with the previous three.Therefore,data-intensive science is viewed as a new and fourth science paradigm for scientific discov-eries [65].In Section 2,we will discuss several transparent Big Data applications around three fields.The opportunities and chal-lenges aroused from Big Data problems will be introduced in Section 3.Then,we give a detailed demonstration of state-of-the-art techniques and technologies to handle data-intensive applications in Section 4,where Big Data tools discussed there will give a helpful guide for expertise users.In Section 5,a number of principles for designing effective Big Data sys-tems are listed.One of the most important parts of this paper,which provides several underlying techniques to settle Big Data problems,is ranged in Section 6.In the last section,we draw a conclusion.2.Big Data problemsAs more and more fields involve Big Data problems,ranging from global economy to society administration,and from scientific researches to national security,we have entered the era of Big Data.Recently,a report [114]from McKinsey insti-tute gives transformative potentials of Big Data in five domains:health care of the United States,public sector administration of European Union,retail of the United States,global manufacturing and personal location data.Their research claimsthatC.L.Philip Chen,C.-Y.Zhang /Information Sciences 275(2014)314–347315316 C.L.Philip Chen,C.-Y.Zhang/Information Sciences275(2014)314–347Big Data can make prominent growth of the world economy by enhancing the productivity and competitiveness of enter-prises and also the public administrations.Big Data has a deep relationship with e-Science[66],which is computationally intensive science which usually is imple-mented in distributed computing systems.Many issues on Big Data applications can be resolved by e-Science which require grid computing[80].e-Sciences include particle physics,bio-informatics,earth sciences and social simulations.It also pro-vides technologies that enable distributed collaboration,such as the Access Grid.Particle physics has a well-developed e-Sci-ence infrastructure in particular because of its need for adequate computing facilities for the analysis of results and storage of data originating from the European Organization for Nuclear Research(CERN)Large Hadron Collider,which started taking data in2009.e-Science is a big concept with many sub-fields,such as e-Social Science which can be regarded as a higher development in e-Science.It plays a role as a part of social science to collect,process,and analyse the social and behavioral data.Other Big Data applications lies in many scientific disciplines like astronomy,atmospheric science,medicine,genomics, biologic,biogeochemistry and other complex and interdisciplinary scientific researches.Web-based applications encounter Big Data frequently,such as recent hot spots social computing(including social network analysis,online communities,rec-ommender systems,reputation systems,and prediction markets),Internet text and documents,Internet search indexing. Alternatively,There are countless sensor around us,they generate sumless sensor data that need to be utilized,for instance, intelligent transportation systems(ITS)[203]are based on the analysis of large volumes of complex sensor rge-scale e-commerce[183]are particularly data-intensive as it involves large number of customers and transactions.In the following subsections,we will briefly introduce several applications of the Big Data problems in commerce and business,society administration and scientific researchfields.2.1.Big Data in commerce and businessAccording to estimates,the volume of business data worldwide,across almost companies,doubles every1.2years[114]. Taking retail industry as an example,we try to give a brief demonstration for the functionalities of Big Data in commercial activities.There are around267million transactions per day in Wal-Mart’s6000stores worldwide.For seeking for higher competitiveness in retail,Wal-Mart recently collaborated with Hewlett Packard to establish a data warehouse which has a capability to store4petabytes(see the size of data unit in Appendix A)of data,i.e.,4000trillion bytes,tracing every pur-chase record from their point-of-sale terminals.Taking advantage of sophisticated machine learning techniques to exploit the knowledge hidden in this huge volume of data,they successfully improve efficiency of their pricing strategies and adver-tising campaigns.The management of their inventory and supply chains also significantly benefits from the large-scale warehouse.In the era of information,almost every big company encounters Big Data problems,especially for multinational corpora-tions.On the one hand,those companies mostly have a large number of customers around the world.On the other hand, there are very large volume and velocity of their transaction data.For instance,FICO’s falcon credit card fraud detection sys-tem manages over2.1billion valid accounts around the world.There are above3billion pieces of content generated on Face-book every day.The same problem happens in every Internet companies.The list could go on and on,as we witness the future businesses battlefields focusing on Big Data.2.2.Big Data in society administrationPublic administration also involves Big Data problems[30].On one side,the population of one country usually is very large.For another,people in each age level need different public services.For examples,kids and teenagers need more edu-cation,the elders require higher level of health care.Every person in one society generates a lot of data in each public section, so the total number of data about public administration in one nation is extremely huge.For instance,there are almost3 terabytes of data collected by the US Library of Congress by2011.The Obama administration announced the Big Data re-search and development initiative in2012,which investigate addressing important problems facing the government by make use of Big Data.The initiative was constitutive of84different Big Data programs involving six departments.1The sim-ilar thing also happened in ernments around the world are facing adverse conditions to improve their productivity. Namely,they are required to be more effective in public administration.Particularly in the recent global recession,many gov-ernments have to provide a higher level of public services with significant budgetary constraints.Therefore,they should take Big Data as a potential budget resource and develop tools to get alternative solutions to decrease big budget deficits and reduce national debt levels.According to McKinsey’s report[114],Big Data functionalities,such as reserving informative patterns and knowledge, provide the public sector a chance to improve productivity and higher levels of efficiency and effectiveness.European’s pub-lic sector could potentially reduce expenditure of administrative activities by15–20percent,increasing223billion to446 billion values,or even more.This estimate is under efficiency gains and a reduction in the difference between actual and 1/blog/2012/03/29/big-data-big-deal.potential aggregate of tax revenue.These functionalities could speed up year productivity growth by up to 0.5percentage points over the next decade.2.3.Big Data in scientific researchMany scientific fields have already become highly data-driven [179,31]with the development of computer sciences.For instance,astronomy,meteorology,social computing [187],bioinformatics [100]and computational biology [117]are greatly based on data-intensive scientific discovery as large volume of data with various types generated or produced in these sci-ence fields [45].How to probe knowledge from the data produced by large-scale scientific simulation?It is a certain Big Data problem which the answer is still unsatisfiable or unknown.For instances,a sophisticated telescope is regarded as a very large digital camera which generate huge number of uni-versal images.For example,the Large Synoptic Survey Telescope (LSST)will record 30trillion bytes of image data in a single day.The size of the data equals to two entire Sloan Digital Sky Surveys daily.Astronomers will utilize computing facilities and advanced analysis methods to this data to investigate the origins of the universe.The Large Hadron Collider (LHC)is a particle accelerator that can generate 60terabytes of data per day [29].The patterns in those data can give us an unprecedented understanding the nature of the universe.32petabytes of climate observations and simulations were con-served on the discovery supercomputing cluster in the NASA Center for Climate Simulation (NCCS).The volume of human genome information is also so large that decoding them originally took a decade to process.Otherwise,a lot of other e-Science projects [66]are proposed or underway in a wide variety of other research fields,range from environmental sci-ence,oceanography and geology to biology and sociology.One common point exists in these disciplines is that they gen-erate enormous data sets that automated analysis is highly required.Additionally,centralized repository is necessary as it is impractical to replicate copies for remote individual research groups.Therefore,centralized storage and analysis ap-proaches drive the whole system designs.3.Big Data opportunities and challenges3.1.OpportunitiesRecently,several US government agencies,such as the National Institutes of Health (NIH)and the National Science Foundation (NSF),ascertain that the utilities of Big Data to data-intensive decision-making have profound influences in their future developments [1].Consequently,they are trying to developing Big Data technologies and techniques to facil-itate their missions after US government passed a large-scale Big Data initiative.This initiative is very helpful for building new capabilities for exploiting informative knowledge and facilitate decision-makers.From the Networking Information Technology Research and Development (NITRD)program which is recently recognized by President’s Council of Advisors on Science and Technology (PCAST),we know that the bridges between Big Data and knowledge hidden in it are highly crucial in all areas of national priority.This initiative will also lay the groundwork for com-plementary Big Data activities,such as Big Data infrastructure projects,platforms development,and techniques in settling complex,data-driven problems in sciences and engineering.Finally,they will be put into practice and benefit society.According to the report from McKinsey institute [114],the effective use of Big Data has the underlying benefits to transform economies,and delivering a new wave of productive growth.Taking advantages of valuable knowledge beyond Big Data will become the basic competition for today’s enterprises and will create new competitors who are able to attract employees that have the critical skills on Big Data.Researchers,policy and decision makers have to recognize the potential of harnessing Big Data to uncover the next wave of growth in their fields.There are many advantages in business section that can be obtained through harnessing Big Data as illustrated in Fig.2,including increasing operational efficiency,informing strategic direction,developing better customer service,identifying and developing new products andservices,C.L.Philip Chen,C.-Y.Zhang /Information Sciences 275(2014)314–347317318 C.L.Philip Chen,C.-Y.Zhang/Information Sciences275(2014)314–347identifying new customers and markets,etc.The vertical axis denotes the percentages that how many enterprises think Big Data can help them with respect to specific purposes.By liberal estimates[114],Big Data could produce$300billion potential annual value to US health care,and€250bil-lion to European public administration.There will be$600billion potential annual consumer surplus from using personal location data globally,and give a potential increase with60%.Only in United States,Big Data produce140,000to190,000 deep analytical talent positions and1.5million data-savvy managers.Undoubtedly,Big Data is usually juicy and lucrative if explored correctly.3.2.ChallengesOpportunities are always followed by challenges.On the one hand,Big Data bring many attractive opportunities.On the other hand,we are also facing a lot of challenges[137]when handle Big Data problems,difficulties lie in data capture, storage,searching,sharing,analysis,and visualization.If we cannot surmount those challenges,Big Data will become a gold ore but we do not have the capabilities to explore it,especially when information surpass our capability to harness. One challenge is existing in computer architecture for several decades,that is,CPU-heavy but I/O-poor[65].This system imbalance still restraint the development of the discovery from Big Data.The CPU performance is doubling each18months following the Moore’s Law,and the performance of disk drives is also doubling at the same rate.However,the disks’rotational speed has slightly improved over the last decade.The conse-quence of this imbalance is that random I/O speeds have improved moderately while sequential I/O speeds increase with density slowly.Moreover,information is increasing at exponential rate simultaneously,but the improvement of informa-tion processing methods is also relatively slower.In a lot of important Big Data applications,the state-of-the-art tech-niques and technologies cannot ideally solve the real problems,especially for real-time analysis.So partially speaking, until now,we do not have the proper tools to exploit the gold ores completely.Typically,the analysis process is shown In Fig.3,where the knowledge is discovered in data mining[59].Challenges in Big Data analysis include data inconsistence and incompleteness,scalability,timeliness and data security[8,92].As the prior step to data analysis,data must be well-constructed.However,considering variety of data sets in Big Data problems, it is still a big challenge for us to purpose efficient representation,access,and analysis of unstructured or semi-structured data in the further researches.How can the data be preprocessed in order to improve the quality data and the analysis results before we begin data analysis?As the sizes of data set are often very huge,sometimes several gigabytes or more, and their origin from heterogeneous sources,current real-world databases are severely susceptible to inconsistent,incom-Knowledge discoveryC.L.Philip Chen,C.-Y.Zhang/Information Sciences275(2014)314–3473193.2.1.Data capture and storageData sets grow in size because they are increasingly being gathered by ubiquitous information-sensing mobile devices, aerial sensory technologies,remote sensing,software logs,cameras,microphones,radio-frequency identification readers, wireless sensor networks,and so on.There are2:5quintillion bytes of data created every day,and this number keeps increas-ing exponentially[67].The world’s technological capacity to store information has roughly doubled about every3years since the1980s.In manyfields,likefinancial and medical data often be deleted just because there is no enough space to store these data.These valuable data are created and captured at high cost,but ignoredfinally.The bulk storage requirements for experimental data bases,array storage for large-scale scientific computations,and large outputfiles are reviewed in [194].Big Data has changed the way we capture and store data[133],including data storage device,data storage architecture, data access mechanism.As we require more storage mediums and higher I/O speed to meet the challenges,there is no doubt that we need great innovations.Firstly,the accessibility of Big Data is on the top priority of the knowledge discovery process. Big Data should be accessed easily and promptly for further analysis,fully or partially break the restraint:CPU-heavy but I/O-poor.In addition,the under-developing storage technologies,such as solid-state drive(SSD)[73]and phase-change memory(PCM)[144],may help us alleviate the difficulties,but they are far from enough.One significant shift is also under-way,that is the transformative change of the traditional I/O subsystems.In the past decades,the persistent data were stored by using hard disk drives(HDDs)[87].As we known,HDDs had much slower random I/O performance than sequential I/O performance,and data processing engines formatted their data and designed their query processing methods to work around this limitation.But,HDDs are increasingly being replaced by SSDs today,and other technologies such as PCM are also around the corner[8].These current storage technologies cannot possess the same high performance for both the sequential and random I/O simultaneously,which requires us to rethink how to design storage subsystems for Big Data processing systems.Direct-attached storage(DAS),network-attached storage(NAS),and storage area network(SAN)are the enterprise stor-age architectures that were commonly used[99].However,all these existing storage architectures have severe drawbacks and limitations when it comes to large-scale distributed systems.Aggressive concurrency and per server throughput are the essential requirements for the applications on highly scalable computing clusters,and today’s storage systems lack the both.Optimizing data access is a popular way to improve the performance of data-intensive computing[78,77,79],these techniques include data replication,migration,distribution,and access parallelism.In[19],the performance,reliability and scalability in data-access platforms were discussed.Data-access platforms,such as CASTOR,dCache,GPFS and Scalla/Xrootd, are employed to demonstrate the large scale validation and performance measurement.Data storage and search schemes also lead to high overhead and latency[162],distributed data-centric storage is a good approach in large-scale wireless sen-sor networks(WSNs).Shen,Zhao and Li proposed a distributed spatial–temporal similarity data storage scheme to provide efficient spatial–temporal and similarity data searching service in WSNs.The collective behavior of individuals that cooper-ate in a swarm provide approach to achieve self-organization in distributed systems[124,184].3.2.2.Data transmissionCloud data storage is popularly used as the development of cloud technologies.We know that the network bandwidth capacity is the bottleneck in cloud and distributed systems,especially when the volume of communication is large.On the other side,cloud storage also lead to data security problems[190]as the requirements of data integrity checking.Many schemes were proposed under different systems and security models[189,134].3.2.3.Data curationData curation is aimed at data discovery and retrieval,data quality assurance,value addition,reuse and preservation over time.Thisfield specifically involves a number of sub-fields including authentication,archiving,management,preservation, retrieval,and representation.The existing database management tools are unable to process Big Data that grow so large and complex.This situation will continue as the benefits of exploiting Big Data allowing researchers to analyse business trends, prevent diseases,and combat crime.Though the size of Big Data keeps increasing exponentially,current capability to work with is only in the relatively lower levels of petabytes,exabytes and zettabytes of data.The classical approach of managing structured data includes two parts,one is a schema to storage the data set,and another is a relational database for data re-trieval.For managing large-scale datasets in a structured way,data warehouses and data marts are two popular approaches.A data warehouse is a relational database system that is used to store and analyze data,also report the results to users.The data mart is based on a data warehouse and facilitate the access and analysis of the data warehouse.A data warehouse is mainly responsible to store data that is sourced from the operational systems.The preprocessing of the data is necessary before it is stored,such as data cleaning,transformation and cataloguing.After these preprocessing,the data is available for higher level online data mining functions.The data warehouse and marts are Standard Query Language(SQL)based dat-abases systems.NoSQL database[60],also called‘‘Not Only SQL’’,is a current approach for large and distributed data management and database design.Its name easily leads to misunderstanding that NoSQL means‘‘not SQL’’.On the contrary,NoSQL does not avoid SQL.While it is true that some NoSQL systems are entirely non-relational,others simply avoid selected relational functionality such asfixed table schemas and join operations.The mainstream Big Data platforms adopt NoSQL to break and transcend the rigidity of normalized RDBMS schemas.For instance,Hbase is one of the most famous used NoSQL databases(see Fig.4).However,many Big Data analytic platforms,like SQLstream and Cloudera Impala,series still use SQL in its data-base systems,because SQL is more reliable and simpler query language with high performance in stream Big Data real-time analytics.To store and manage unstructured data or non-relational data,NoSQL employs a number of specific approaches.Firstly,data storage and management are separated into two independent parts.This is contrary to relational databases which try to meet the concerns in the two sides simultaneously.This design gives NoSQL databases systems a lot of advantages.In the storage part which is also called key-value storage,NoSQL focuses on the scalability of data storage with high-performance.In the management part,NoSQL provides low-level access mechanism in which data management tasks can be implemented in the application layer rather than having data management logic spread across in SQL or DB-specific stored procedure lan-guages [37].Therefore,NoSQL systems are very flexible for data modeling,and easy to update application developments and deployments [60].Most NoSQL databases have an important ly,they are commonly schema-free.Indeed,the biggest advan-tage of schema-free databases is that it enables applications to quickly modify the structure of data and does not need to rewrite tables.Additionally,it possesses greater flexibility when the structured data is heterogeneously stored.In the data management layer,the data is enforced to be integrated and valid.The most popular NoSQL database is Apache Cassandra.Cassandra,which was once Facebook proprietary database,was released as open source in 2008.Other NoSQL implementa-tions include SimpleDB,Google BigTable,Apache Hadoop,MapReduce,MemcacheDB,and panies that use NoSQL include Twitter,LinkedIn and NetFlix.3.2.4.Data analysisThe first impression of Big Data is its volume,so the biggest and most important challenge is scalability when we deal with the Big Data analysis tasks.In the last few decades,researchers paid more attentions to accelerate analysis algorithms to cope with increasing volumes of data and speed up processors following the Moore’s Law.For the former,it is necessary to develop sampling,on-line,and multiresolution analysis methods [59].In the aspect of Big Data analytical techniques,incre-ment algorithms have good scalability property,not for all machine learning algorithms.Some researchers devote into this area [180,72,62].As the data size is scaling much faster than CPU speeds,there is a natural dramatic shift [8]in processor technology—although the clock cycle frequency of processors is doubling following Moore’s Law,the clock speeds still highly lag behind.Alternatively,processors are being embedded with increasing numbers of cores.This shift in processors leads to the development of parallel computing [130,168,52].For those real-time Big Data applications,like navigation,social networks,finance,biomedicine,astronomy,intelligent transport systems,and internet of thing,timeliness is at the top priority.How can we grantee the timeliness of response when the volume of data will be processed is very large?It is still a big challenge for stream processing involved by Big Data.It is right to say that Big Data not only have produced many challenge and changed the directions of the development of the hardware,but also in software architectures.That is the swerve to cloud computing [50,186,7,48],which aggregates multiple disparate workloads into a large cluster of processors.In this direction,distributed computing is being developed at high speed recently.We will give a more detail discussion about it in next section.Fig.4.Hbase NoSQL database system architecture.Source:from Apache Hadoop.。