云存储服务系统研究外文文献翻译
- 格式:doc
- 大小:358.00 KB
- 文档页数:9
云存储服务端海量数据安全存储的加密解决方案朱荣;周彩兰;高瑞【摘要】云存储是利用计算机网络技术发展起来的一种为使用者提供数据存储和访问的服务,是在云计算的基础上发展而来的。
对云存储的主要概念及相关结构进行了具体介绍,对于现阶段安全方面云存储的问题进行研究,提出了一种适合的数据加密解决方案,能够对使用者的数据隐私有效保护,为云存储的发展及应用起到了重要作用。
%The cloud storage developed by computer network technology is a service to provide the data storage and access for users,which is developed based on the cloud computing. The key concept and relevance structure of the cloud storage are in?troduced in detail,and the cloud storage security problem at present stage is studied. A suitable data encryption solution is put forward,which can protect the data privacy effective for users,and play a main significance for the development and application of the cloud storage.【期刊名称】《现代电子技术》【年(卷),期】2017(040)003【总页数】3页(P79-81)【关键词】云存储服务端;海量数据;安全存储;数据加密解决方案【作者】朱荣;周彩兰;高瑞【作者单位】汉江师范学院,湖北十堰 442000;武汉理工大学,湖北武汉430070;汉江师范学院,湖北十堰 442000【正文语种】中文【中图分类】TN915.08-341.1 云存储的定义云存储(Cloud Storage)通过集成合作软件技术,以计算机网络技术为基础,分布式存储技术、海量数据存储技术为核心,使接入网络的各类型计算机存储设备将各种信息传输至外界,同时提供业务访问、信息共享等服务的系统。
基于云服务翻译平台的翻译质量评估思考语言是人类交流思想和文化传播的载体。
翻译作为一种双语转换活动,承载着促进不同民族、不同国家之间互相融合、共同发展的重任。
近年来,计算机技术和网络环境的飞速发展引起信息的快速更新和传播,翻译任务的成倍增长要求翻译方式在应对大批量、多语种翻译任务时必须有更快的翻译速度和更高的翻译质量。
为了应对市场要求,一些翻译公司开始将云计算技术应用于翻译行业,构建能满足用户需求日渐增长的云服务翻译平台,将分散在各地的翻译资源整合到一起,降低成本、方便管理,为各地商业贸易发展和文化交流提供广泛全面、安全可信的翻译服务,推动社会经济和文化快速发展。
一、云平台下的翻译模式构建(一)云平台下的翻译技术框架云翻译是一个高效的翻译生态系统。
云服务翻译平台的基础在于语料库系统的建设,通过翻译的云存储,实现翻译资源的合理分布,因此各种语料库资源、各种存储介质及服务器可以共存于云端,从而实现翻译资源的共享,将硬件(服务器、存储器及网络)和相关软件(语料库、在线词典系统)作为服务交付给用户(客户、翻译工作者)使用,形成一个相互促进、共同发展的翻译生态系统。
云翻译平台的核心是构建翻译管理平台。
云服务翻译平台的管理系统主要包括四大模块:客户管理、众包管理、翻译管理、质量管理。
通过管理系统实现工作人员与客户的交流,管理客户翻译的需求、过程以及关系维护,客户则可以提出翻译定制要求,实现在线支付等功能。
云翻译平台的主要管理模式技术是实现机器翻译和众包翻译的结合。
因此众包管理模块是云翻译平台最为重要的组成部分,通过众包模块按专业或语种实现翻译任务的分解、发包、过程控制等流程,实时呈现项目成员翻译过程中的工作记录与沟通协调,实现译者筛选。
云翻译平台下,每一项翻译任务均利用多语言翻译人才储备库形成专业项目翻译团队。
一个专业翻译团队是翻译项目顺利实施的基础。
首先根据翻译项目内容确定一位具有多年本语种、本领域实践经验且具有良好职业道德的项目组组长,组长再根据工作任务挑选具备必要的专业知识且职业素养高的翻译成员,形成了团队共同完成翻译任务一种格局。
2.信息技术的发展,使得内容需求的翻译大幅增长随着信息技术,特别是w eb2.o的应用,使内容产业需求迅速增长,同时也带来了翻译需求的增长。
内容需求导致的翻译特点在于内容更新及时,尽管许多对于信息及内容的需求是暂时的,并不构成一个完整的传统意义上的正式的翻译任务,但是内容的快速更新和信息的快速传播,却意味着更快的翻译速度和翻译任务的几何级增长。
因此新的翻译模式必须具备能应对频繁、快速翻译所导致的译员数量增长和水平需要提高的特点。
总之,信息时代新的竞争环境变化,迫使翻译产业朝更低廉、更快速和更高质量方向发展。
二、翻译产业环境下的翻译模式比较在探讨了翻译产业的环境需求后,我们再来审视一下目前翻译产业的主要供应模式。
在W eb2.0的今天,机器翻译和众包翻译可以看成两种主要的信息技术类翻译模式,尽管有人提出了云计算翻译也可视为一种主要的翻译模式,但我们认为云计算翻译仅仅是一种翻译技术和翻译环境,并没有改变翻译产业的主要流程与各方参与者,因此我们仅对机器翻译和众包翻译进行分析。
1.机器翻译机器翻译(M T)使用计算机软件将文本或谈话从一种自然语言(源语言)翻译到另一种语言(目标语言),根据不同的架构主要包括基于规则的机器翻译(R B M T),基于统计的机器翻译(SM T),基于实例的机器翻译(E B M T)。
经过几十年的发展,如今的机器翻译进入到了开放式的翻译阶段,如s Y S.,I R A N,L a nguage w eaver以及A ppTek TraJl Sphere等(A nast asi ou&G upt a,201l:637),可以实现在线的同步翻译,是目前翻译产业一个重要的应用模式。
2.众包翻译2007年,著名的社交网站Fac ebook发动双语用户志愿为网站进行翻译,并一举成功(M eer,2010),开启了众包翻译的时代。
众包翻译模式迅速在社交媒体、非盈利性组织、文化传播、政府机构等方面得到了许多应用,并迅速成为翻译产业的一个新兴的应用模式。
软件工程2班代兄2011020339Low Power Mode in Cloud Storage Systems低功率模式在云存储系统Danny Harnik, Dalit Naor and Itai SegallIBM Haifa Research Labs, Haifa, Israelfdannyh, dalit, itaisg@.AbstractWe consider large scale, distributed storage systems with a redundancy mechanism; cloud storage being a prime example. We investigate how such systems can reduce their power consumption during low-utilization time intervals by operating in a low-power mode. In a low power mode, a subset of the disks or nodes are powered down, yet we ask that each data item remains accessible in the system; this is called full coverage. The objective is to incorporate this option into an existing system rather than redesign the system. When doing so, it is crucial that the low power option should not affect the performance or other important characteristics of the system during full-power (normal) operation. This work is a comprehensive study of what can or cannot be achieved with respect to full coverage low power modes.我们考虑包含冗余机制的大规模、分布式的存储系统,云存储是一个最典型的例子。
Database Systems1. Fundamental Concepts of DatabaseDatabase and database technology are having a major impact on the growing use of computers. It is fair to say that database will play a critical role in almost all areas where computers are used, including business, engineering, medicine, law, education, and library science, to name a few. The word "database" is in such common use that we must begin by defining what a database is. Our initial definition is quit general.A database is a collection of related data. By data, we mean known facts that can be recorded and that have implicit meaning. For example, consider the names, telephone numbers, and addresses of all the people you know. Y ou may have recorded this data in an indexed address book, or you may have stored it on a diskette using a personal computer and software such as DBASE III or Lotus 1-2-3. This is a collection of related data with an implic it meaning and hence is a database.The above definition of database is quite general; for example, we may consider the collection of words that make up thispage of text to be related data and hence a database. However, the common use of the term database is usually more restricted.A database has the following implicit properties:.A database is a logically coherent collection of data with some inherent meaning. A random assortment of data cannot bereferred to as a database..A database is designed, built, and populated with data for a specific purpose. It has an intended group of users and somepreconceived applications in which these users are interested..A database represents some aspect of the real world, sometimes called the mini world. Changes to the mini world are reflected in the database.In other words, a database has some source from which data are derived, some degree of interaction with events in the real world, and an audience that is actively interested in the contents of the database.A database can be of any size and of varying complexity. For example, the list of names and addresses referred to earlier may have only a couple of hundred records in it, each with asimple structure. On the other hand, the card catalog of a large library may contain half a million cards stored under different categories-by primary author’s last name, by subject, by book title, and the like-with each category organized in alphabetic order. A database of even greater size and complexity may be that maintained by the Internal Revenue Service to keep track of the tax forms filed by taxpayers of the United States. If we assume that there are 100million taxpayers and each taxpayer files an average of five forms with approximately 200 characters of information per form, we would get a database of 100*(106)*200*5 characters(bytes) of information. Assuming the IRS keeps the past three returns for each taxpayer in addition to the current return, we would get a database of 4*(1011) bytes. This huge amount of information must somehow be organized and managed so that users can search for, retrieve, and update the data as needed.A database may be generated and maintained manually or by machine. Of course, in this we are mainly interested in computerized database. The library card catalog is an example of a database that may be manually created and maintained. A computerized database may be created and maintained either by a group of application programs written specifically for that task or by a database management system.A data base management system (DBMS) is a collection of programs that enables users to create and maintain a database. The DBMS is hence a general-purpose software system that facilitates the processes of defining, constructing, and manipulating databases for various applications. Defining a database involves specifying the types of data to be stored in the database, along with a detailed description of each type of data. Constructing the database is the process of storing the data itself on some storage medium that is controlled by the DBMS. Manipulating a database includes such functions as querying the database to retrieve specific data, updating the database to reflect changes in the mini world, and generating reports from the data.Note that it is not necessary to use general-purpose DBMS software for implementing a computerized database. We could write our own set of programs to create and maintain the database, in effect creating our own special-purpose DBMS software. In either case-whether we use a general-purpose DBMS or not-we usually have a considerable amount of software to manipulate the database in addition to the database itself. The database and software are together called a database system.2. Data ModelsOne of the fundamental characteristics of the database approach is that it provides some level of data abstraction by hiding details of data storage that are not needed by most database users. A data model is the main tool for providing this abstraction. A data is a set of concepts that can beused to describe the structure of a database. By structure of a database, we mean the data types, relationships, and constraints that should hold on the data. Most data models also include a set of operations for specifying retrievals and updates on the database.Categories of Data ModelsMany data models have been proposed. We can categorize data models based on the types of concepts they provide to describe the database structure. High-level or conceptual data models provide concepts that are close to the way many users perceive data, whereas low-level or physical data models provide concepts that describe the details of how data is stored in the computer. Concepts provided by low-level data models are generally meant for computer specialists, not for typical end users. Between these two extremes is a class of implementation data models, which provide concepts that may be understood by end users but that are not too far removed from the way data is organized within the computer. Implementation data models hide some details of data storage but can be implemented on a computer system in a direct way.High-level data models use concepts such as entities, attributes, and relationships. An entity is an object that is represented in the database. An attribute is a property that describes some aspect of an object. Relationships among objects are easily represented in high-level data models, which are sometimes called object-based models because they mainly describe objects and their interrelationships.Implementation data models are the ones used most frequently in current commerc ial DBMSs and include the three most widely used data models-relational, network, and hierarchical. They represent data using record structures and hence are sometimes called record-based data modes.Physical data models describe how data is stored in the computer by representing information such as record formats, record orderings, and access paths. An access path is a structure that makes the search for particular database records much faster.3. Classification of Database Management SystemsThe main criterion used to classify DBMSs is the data model on which the DBMS is based. The data models used most often in current commercial DBMSs are the relational, network, and hierarchical models. Some recent DBMSs are based on conceptual or object-oriented models. We will categorize DBMSs as relational, hierarchical, and others.Another criterion used to classify DBMSs is the number of users supported by the DBMS. Single-user systems support only one user at a time and are mostly used with personal computer. Multiuser systems include the majority of DBMSs and support many users concurrently.A third criterion is the number of sites over which the database is distributed. Most DBMSs are centralized, meaning that their data is stored at a single computer site. A centralized DBMS can support multiple users, but the DBMS and database themselves reside totally at a single computer site. A distributed DBMS (DDBMS) can have the actual database and DBMS software distributed over many sites connected by a computer network. Homogeneous DDBMSs use the same DBMS software at multiple sites. A recent trend is to develop software to access several autonomous preexisting database stored under heterogeneous DBMSs. This leads to a federated DBMS (or multidatabase system),, where the participating DBMSs are loosely coupled and have a degree of local autonomy.We can also classify a DBMS on the basis of the types of access paty options available for storing files. One well-known family of DBMSs is based on inverted file structures. Finally, a DBMS can be general purpose of special purpose. When performance is a prime consideration, a special-purpose DBMS can be designed and built for a specific application and cannot be used for other applications, Many airline reservations and telephone directory systems are special-purpose DBMSs.Let us briefly discuss the main criterion for classifying DBMSs: the data mode. The relational data model represents a database as a collection of tables, which look like files. Mos t relational databases have high-level query languages and support a limited form of user views.The network model represents data as record types and also represents a limited type of 1:N relationship, called a set type. The network model, also known as the CODASYL DBTG model, has an associated record-at-a-time language that must be embedded in a host programming language.The hierarchical model represents data as hierarchical tree structures. Each hierarchy represents a number of related records. There is no standard language for the hierarchical model, although most hierarchical DBMSs have record-at-a-time languages.4. Client-Server ArchitectureMany varieties of modern software use a client-server architecture, in which requests by one process (the client) are sent to another process (the server) for execution. Database systems are no exception. In the simplest client/server architecture, the entire DBMS is a server, except for the query interfaces that interact with the user and send queries or other commands across to the server. For example, relational systems generally use the SQL language for representing requests from the client to the server. The database server then sends the answer, in the form of a table or relation, back to the client. The relationship between client and server can get more work in theclient, since the server will e a bottleneck if there are many simultaneous database users.。
云计算外文翻译参考文献(文档含中英文对照即英文原文和中文翻译)原文:Technical Issues of Forensic Investigations in Cloud Computing EnvironmentsDominik BirkRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyRuhr-University BochumHorst Goertz Institute for IT SecurityBochum, GermanyAbstract—Cloud Computing is arguably one of the most discussedinformation technologies today. It presents many promising technological and economical opportunities. However, many customers remain reluctant to move their business IT infrastructure completely to the cloud. One of their main concerns is Cloud Security and the threat of the unknown. Cloud Service Providers(CSP) encourage this perception by not letting their customers see what is behind their virtual curtain. A seldomly discussed, but in this regard highly relevant open issue is the ability to perform digital investigations. This continues to fuel insecurity on the sides of both providers and customers. Cloud Forensics constitutes a new and disruptive challenge for investigators. Due to the decentralized nature of data processing in the cloud, traditional approaches to evidence collection and recovery are no longer practical. This paper focuses on the technical aspects of digital forensics in distributed cloud environments. We contribute by assessing whether it is possible for the customer of cloud computing services to perform a traditional digital investigation from a technical point of view. Furthermore we discuss possible solutions and possible new methodologies helping customers to perform such investigations.I. INTRODUCTIONAlthough the cloud might appear attractive to small as well as to large companies, it does not come along without its own unique problems. Outsourcing sensitive corporate data into the cloud raises concerns regarding the privacy and security of data. Security policies, companies main pillar concerning security, cannot be easily deployed into distributed, virtualized cloud environments. This situation is further complicated by the unknown physical location of the companie’s assets. Normally,if a security incident occurs, the corporate security team wants to be able to perform their own investigation without dependency on third parties. In the cloud, this is not possible anymore: The CSP obtains all the power over the environmentand thus controls the sources of evidence. In the best case, a trusted third party acts as a trustee and guarantees for the trustworthiness of the CSP. Furthermore, the implementation of the technical architecture and circumstances within cloud computing environments bias the way an investigation may be processed. In detail, evidence data has to be interpreted by an investigator in a We would like to thank the reviewers for the helpful comments and Dennis Heinson (Center for Advanced Security Research Darmstadt - CASED) for the profound discussions regarding the legal aspects of cloud forensics. proper manner which is hardly be possible due to the lackof circumstantial information. For auditors, this situation does not change: Questions who accessed specific data and information cannot be answered by the customers, if no corresponding logs are available. With the increasing demand for using the power of the cloud for processing also sensible information and data, enterprises face the issue of Data and Process Provenance in the cloud [10]. Digital provenance, meaning meta-data that describes the ancestry or history of a digital object, is a crucial feature for forensic investigations. In combination with a suitable authentication scheme, it provides information about who created and who modified what kind of data in the cloud. These are crucial aspects for digital investigations in distributed environments such as the cloud. Unfortunately, the aspects of forensic investigations in distributed environment have so far been mostly neglected by the research community. Current discussion centers mostly around security, privacy and data protection issues [35], [9], [12]. The impact of forensic investigations on cloud environments was little noticed albeit mentioned by the authors of [1] in 2009: ”[...] to our knowledge, no research has been published on how cloud computing environments affect digital artifacts,and on acquisition logistics and legal issues related to cloud computing env ironments.” This statement is also confirmed by other authors [34], [36], [40] stressing that further research on incident handling, evidence tracking and accountability in cloud environments has to be done. At the same time, massive investments are being made in cloud technology. Combined with the fact that information technology increasingly transcendents peoples’ private and professional life, thus mirroring more and more of peoples’actions, it becomes apparent that evidence gathered from cloud environments will be of high significance to litigation or criminal proceedings in the future. Within this work, we focus the notion of cloud forensics by addressing the technical issues of forensics in all three major cloud service models and consider cross-disciplinary aspects. Moreover, we address the usability of various sources of evidence for investigative purposes and propose potential solutions to the issues from a practical standpoint. This work should be considered as a surveying discussion of an almost unexplored research area. The paper is organized as follows: We discuss the related work and the fundamental technical background information of digital forensics, cloud computing and the fault model in section II and III. In section IV, we focus on the technical issues of cloud forensics and discuss the potential sources and nature of digital evidence as well as investigations in XaaS environments including thecross-disciplinary aspects. We conclude in section V.II. RELATED WORKVarious works have been published in the field of cloud security and privacy [9], [35], [30] focussing on aspects for protecting data in multi-tenant, virtualized environments. Desired security characteristics for current cloud infrastructures mainly revolve around isolation of multi-tenant platforms [12], security of hypervisors in order to protect virtualized guest systems and secure network infrastructures [32]. Albeit digital provenance, describing the ancestry of digital objects, still remains a challenging issue for cloud environments, several works have already been published in this field [8], [10] contributing to the issues of cloud forensis. Within this context, cryptographic proofs for verifying data integrity mainly in cloud storage offers have been proposed,yet lacking of practical implementations [24], [37], [23]. Traditional computer forensics has already well researched methods for various fields of application [4], [5], [6], [11], [13]. Also the aspects of forensics in virtual systems have been addressed by several works [2], [3], [20] including the notionof virtual introspection [25]. In addition, the NIST already addressed Web Service Forensics [22] which has a huge impact on investigation processes in cloud computing environments. In contrast, the aspects of forensic investigations in cloud environments have mostly been neglected by both the industry and the research community. One of the first papers focusing on this topic was published by Wolthusen [40] after Bebee et al already introduced problems within cloud environments [1]. Wolthusen stressed that there is an inherent strong need for interdisciplinary work linking the requirements and concepts of evidence arising from the legal field to what can be feasibly reconstructed and inferred algorithmically or in an exploratory manner. In 2010, Grobauer et al [36] published a paper discussing the issues of incident response in cloud environments - unfortunately no specific issues and solutions of cloud forensics have been proposed which will be done within this work.III. TECHNICAL BACKGROUNDA. Traditional Digital ForensicsThe notion of Digital Forensics is widely known as the practice of identifying, extracting and considering evidence from digital media. Unfortunately, digital evidence is both fragile and volatile and therefore requires the attention of special personnel and methods in order to ensure that evidence data can be proper isolated and evaluated. Normally, the process of a digital investigation can be separated into three different steps each having its own specificpurpose:1) In the Securing Phase, the major intention is the preservation of evidence for analysis. The data has to be collected in a manner that maximizes its integrity. This is normally done by a bitwise copy of the original media. As can be imagined, this represents a huge problem in the field of cloud computing where you never know exactly where your data is and additionallydo not have access to any physical hardware. However, the snapshot technology, discussed in section IV-B3, provides a powerful tool to freeze system states and thus makes digital investigations, at least in IaaS scenarios, theoretically possible.2) We refer to the Analyzing Phase as the stage in which the data is sifted and combined. It is in this phase that the data from multiple systems or sources is pulled together to create as complete a picture and event reconstruction as possible. Especially in distributed system infrastructures, this means that bits and pieces of data are pulled together for deciphering the real story of what happened and for providing a deeper look into the data.3) Finally, at the end of the examination and analysis of the data, the results of the previous phases will be reprocessed in the Presentation Phase. The report, created in this phase, is a compilation of all the documentation and evidence from the analysis stage. The main intention of such a report is that it contains all results, it is complete and clear to understand. Apparently, the success of these three steps strongly depends on the first stage. If it is not possible to secure the complete set of evidence data, no exhaustive analysis will be possible. However, in real world scenarios often only a subset of the evidence data can be secured by the investigator. In addition, an important definition in the general context of forensics is the notion of a Chain of Custody. This chain clarifies how and where evidence is stored and who takes possession of it. Especially for cases which are brought to court it is crucial that the chain of custody is preserved.B. Cloud ComputingAccording to the NIST [16], cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications and services) that can be rapidly provisioned and released with minimal CSP interaction. The new raw definition of cloud computing brought several new characteristics such as multi-tenancy, elasticity, pay-as-you-go and reliability. Within this work, the following three models are used: In the Infrastructure asa Service (IaaS) model, the customer is using the virtual machine provided by the CSP for installing his own system on it. The system can be used like any other physical computer with a few limitations. However, the additive customer power over the system comes along with additional security obligations. Platform as a Service (PaaS) offerings provide the capability to deploy application packages created using the virtual development environment supported by the CSP. For the efficiency of software development process this service model can be propellent. In the Software as a Service (SaaS) model, the customer makes use of a service run by the CSP on a cloud infrastructure. In most of the cases this service can be accessed through an API for a thin client interface such as a web browser. Closed-source public SaaS offers such as Amazon S3 and GoogleMail can only be used in the public deployment model leading to further issues concerning security, privacy and the gathering of suitable evidences. Furthermore, two main deployment models, private and public cloud have to be distinguished. Common public clouds are made available to the general public. The corresponding infrastructure is owned by one organization acting as a CSP and offering services to its customers. In contrast, the private cloud is exclusively operated for an organization but may not provide the scalability and agility of public offers. The additional notions of community and hybrid cloud are not exclusively covered within this work. However, independently from the specific model used, the movement of applications and data to the cloud comes along with limited control for the customer about the application itself, the data pushed into the applications and also about the underlying technical infrastructure.C. Fault ModelBe it an account for a SaaS application, a development environment (PaaS) or a virtual image of an IaaS environment, systems in the cloud can be affected by inconsistencies. Hence, for both customer and CSP it is crucial to have the ability to assign faults to the causing party, even in the presence of Byzantine behavior [33]. Generally, inconsistencies can be caused by the following two reasons:1) Maliciously Intended FaultsInternal or external adversaries with specific malicious intentions can cause faults on cloud instances or applications. Economic rivals as well as former employees can be the reason for these faults and state a constant threat to customers and CSP. In this model, also a malicious CSP is included albeit he isassumed to be rare in real world scenarios. Additionally, from the technical point of view, the movement of computing power to a virtualized, multi-tenant environment can pose further threads and risks to the systems. One reason for this is that if a single system or service in the cloud is compromised, all other guest systems and even the host system are at risk. Hence, besides the need for further security measures, precautions for potential forensic investigations have to be taken into consideration.2) Unintentional FaultsInconsistencies in technical systems or processes in the cloud do not have implicitly to be caused by malicious intent. Internal communication errors or human failures can lead to issues in the services offered to the costumer(i.e. loss or modification of data). Although these failures are not caused intentionally, both the CSP and the customer have a strong intention to discover the reasons and deploy corresponding fixes.IV. TECHNICAL ISSUESDigital investigations are about control of forensic evidence data. From the technical standpoint, this data can be available in three different states: at rest, in motion or in execution. Data at rest is represented by allocated disk space. Whether the data is stored in a database or in a specific file format, it allocates disk space. Furthermore, if a file is deleted, the disk space is de-allocated for the operating system but the data is still accessible since the disk space has not been re-allocated and overwritten. This fact is often exploited by investigators which explore these de-allocated disk space on harddisks. In case the data is in motion, data is transferred from one entity to another e.g. a typical file transfer over a network can be seen as a data in motion scenario. Several encapsulated protocols contain the data each leaving specific traces on systems and network devices which can in return be used by investigators. Data can be loaded into memory and executed as a process. In this case, the data is neither at rest or in motion but in execution. On the executing system, process information, machine instruction and allocated/de-allocated data can be analyzed by creating a snapshot of the current system state. In the following sections, we point out the potential sources for evidential data in cloud environments and discuss the technical issues of digital investigations in XaaS environmentsas well as suggest several solutions to these problems.A. Sources and Nature of EvidenceConcerning the technical aspects of forensic investigations, the amount of potential evidence available to the investigator strongly diverges between thedifferent cloud service and deployment models. The virtual machine (VM), hosting in most of the cases the server application, provides several pieces of information that could be used by investigators. On the network level, network components can provide information about possible communication channels between different parties involved. The browser on the client, acting often as the user agent for communicating with the cloud, also contains a lot of information that could be used as evidence in a forensic investigation. Independently from the used model, the following three components could act as sources for potential evidential data.1) Virtual Cloud Instance: The VM within the cloud, where i.e. data is stored or processes are handled, contains potential evidence [2], [3]. In most of the cases, it is the place where an incident happened and hence provides a good starting point for a forensic investigation. The VM instance can be accessed by both, the CSP and the customer who is running the instance. Furthermore, virtual introspection techniques [25] provide access to the runtime state of the VM via the hypervisor and snapshot technology supplies a powerful technique for the customer to freeze specific states of the VM. Therefore, virtual instances can be still running during analysis which leads to the case of live investigations [41] or can be turned off leading to static image analysis. In SaaS and PaaS scenarios, the ability to access the virtual instance for gathering evidential information is highly limited or simply not possible.2) Network Layer: Traditional network forensics is knownas the analysis of network traffic logs for tracing events that have occurred in the past. Since the different ISO/OSI network layers provide several information on protocols and communication between instances within as well as with instances outside the cloud [4], [5], [6], network forensics is theoretically also feasible in cloud environments. However in practice, ordinary CSP currently do not provide any log data from the network components used by the customer’s instances or applications. For instance, in case of a malware infection of an IaaS VM, it will be difficult for the investigator to get any form of routing information and network log datain general which is crucial for further investigative steps. This situation gets even more complicated in case of PaaS or SaaS. So again, the situation of gathering forensic evidence is strongly affected by the support the investigator receives from the customer and the CSP.3) Client System: On the system layer of the client, it completely depends on the used model (IaaS, PaaS, SaaS) if and where potential evidence could beextracted. In most of the scenarios, the user agent (e.g. the web browser) on the client system is the only application that communicates with the service in the cloud. This especially holds for SaaS applications which are used and controlled by the web browser. But also in IaaS scenarios, the administration interface is often controlled via the browser. Hence, in an exhaustive forensic investigation, the evidence data gathered from the browser environment [7] should not be omitted.a) Browser Forensics: Generally, the circumstances leading to an investigation have to be differentiated: In ordinary scenarios, the main goal of an investigation of the web browser is to determine if a user has been victim of a crime. In complex SaaS scenarios with high client-server interaction, this constitutes a difficult task. Additionally, customers strongly make use of third-party extensions [17] which can be abused for malicious purposes. Hence, the investigator might want to look for malicious extensions, searches performed, websites visited, files downloaded, information entered in forms or stored in local HTML5 stores, web-based email contents and persistent browser cookies for gathering potential evidence data. Within this context, it is inevitable to investigate the appearance of malicious JavaScript [18] leading to e.g. unintended AJAX requests and hence modified usage of administration interfaces. Generally, the web browser contains a lot of electronic evidence data that could be used to give an answer to both of the above questions - even if the private mode is switched on [19].B. Investigations in XaaS EnvironmentsTraditional digital forensic methodologies permit investigators to seize equipment and perform detailed analysis on the media and data recovered [11]. In a distributed infrastructure organization like the cloud computing environment, investigators are confronted with an entirely different situation. They have no longer the option of seizing physical data storage. Data and processes of the customer are dispensed over an undisclosed amount of virtual instances, applications and network elements. Hence, it is in question whether preliminary findings of the computer forensic community in the field of digital forensics apparently have to be revised and adapted to the new environment. Within this section, specific issues of investigations in SaaS, PaaS and IaaS environments will be discussed. In addition, cross-disciplinary issues which affect several environments uniformly, will be taken into consideration. We also suggest potential solutions to the mentioned problems.1) SaaS Environments: Especially in the SaaS model, the customer does notobtain any control of the underlying operating infrastructure such as network, servers, operating systems or the application that is used. This means that no deeper view into the system and its underlying infrastructure is provided to the customer. Only limited userspecific application configuration settings can be controlled contributing to the evidences which can be extracted fromthe client (see section IV-A3). In a lot of cases this urges the investigator to rely on high-level logs which are eventually provided by the CSP. Given the case that the CSP does not run any logging application, the customer has no opportunity to create any useful evidence through the installation of any toolkit or logging tool. These circumstances do not allow a valid forensic investigation and lead to the assumption that customers of SaaS offers do not have any chance to analyze potential incidences.a) Data Provenance: The notion of Digital Provenance is known as meta-data that describes the ancestry or history of digital objects. Secure provenance that records ownership and process history of data objects is vital to the success of data forensics in cloud environments, yet it is still a challenging issue today [8]. Albeit data provenance is of high significance also for IaaS and PaaS, it states a huge problem specifically for SaaS-based applications: Current global acting public SaaS CSP offer Single Sign-On (SSO) access control to the set of their services. Unfortunately in case of an account compromise, most of the CSP do not offer any possibility for the customer to figure out which data and information has been accessed by the adversary. For the victim, this situation can have tremendous impact: If sensitive data has been compromised, it is unclear which data has been leaked and which has not been accessed by the adversary. Additionally, data could be modified or deleted by an external adversary or even by the CSP e.g. due to storage reasons. The customer has no ability to proof otherwise. Secure provenance mechanisms for distributed environments can improve this situation but have not been practically implemented by CSP [10]. Suggested Solution: In private SaaS scenarios this situation is improved by the fact that the customer and the CSP are probably under the same authority. Hence, logging and provenance mechanisms could be implemented which contribute to potential investigations. Additionally, the exact location of the servers and the data is known at any time. Public SaaS CSP should offer additional interfaces for the purpose of compliance, forensics, operations and security matters to their customers. Through an API, the customers should have the ability to receive specific information suchas access, error and event logs that could improve their situation in case of aninvestigation. Furthermore, due to the limited ability of receiving forensic information from the server and proofing integrity of stored data in SaaS scenarios, the client has to contribute to this process. This could be achieved by implementing Proofs of Retrievability (POR) in which a verifier (client) is enabled to determine that a prover (server) possesses a file or data object and it can be retrieved unmodified [24]. Provable Data Possession (PDP) techniques [37] could be used to verify that an untrusted server possesses the original data without the need for the client to retrieve it. Although these cryptographic proofs have not been implemented by any CSP, the authors of [23] introduced a new data integrity verification mechanism for SaaS scenarios which could also be used for forensic purposes.2) PaaS Environments: One of the main advantages of the PaaS model is that the developed software application is under the control of the customer and except for some CSP, the source code of the application does not have to leave the local development environment. Given these circumstances, the customer obtains theoretically the power to dictate how the application interacts with other dependencies such as databases, storage entities etc. CSP normally claim this transfer is encrypted but this statement can hardly be verified by the customer. Since the customer has the ability to interact with the platform over a prepared API, system states and specific application logs can be extracted. However potential adversaries, which can compromise the application during runtime, should not be able to alter these log files afterwards. Suggested Solution:Depending on the runtime environment, logging mechanisms could be implemented which automatically sign and encrypt the log information before its transfer to a central logging server under the control of the customer. Additional signing and encrypting could prevent potential eavesdroppers from being able to view and alter log data information on the way to the logging server. Runtime compromise of an PaaS application by adversaries could be monitored by push-only mechanisms for log data presupposing that the needed information to detect such an attack are logged. Increasingly, CSP offering PaaS solutions give developers the ability to collect and store a variety of diagnostics data in a highly configurable way with the help of runtime feature sets [38].3) IaaS Environments: As expected, even virtual instances in the cloud get compromised by adversaries. Hence, the ability to determine how defenses in the virtual environment failed and to what extent the affected systems havebeen compromised is crucial not only for recovering from an incident. Also forensic investigations gain leverage from such information and contribute to resilience against future attacks on the systems. From the forensic point of view, IaaS instances do provide much more evidence data usable for potential forensics than PaaS and SaaS models do. This fact is caused throughthe ability of the customer to install and set up the image for forensic purposes before an incident occurs. Hence, as proposed for PaaS environments, log data and other forensic evidence information could be signed and encrypted before itis transferred to third-party hosts mitigating the chance that a maliciously motivated shutdown process destroys the volatile data. Although, IaaS environments provide plenty of potential evidence, it has to be emphasized that the customer VM is in the end still under the control of the CSP. He controls the hypervisor which is e.g. responsible for enforcing hardware boundaries and routing hardware requests among different VM. Hence, besides the security responsibilities of the hypervisor, he exerts tremendous control over how customer’s VM communicate with the hardware and theoretically can intervene executed processes on the hosted virtual instance through virtual introspection [25]. This could also affect encryption or signing processes executed on the VM and therefore leading to the leakage of the secret key. Although this risk can be disregarded in most of the cases, the impact on the security of high security environments is tremendous.a) Snapshot Analysis: Traditional forensics expect target machines to be powered down to collect an image (dead virtual instance). This situation completely changed with the advent of the snapshot technology which is supported by all popular hypervisors such as Xen, VMware ESX and Hyper-V.A snapshot, also referred to as the forensic image of a VM, providesa powerful tool with which a virtual instance can be clonedby one click including also the running system’s mem ory. Due to the invention of the snapshot technology, systems hosting crucial business processes do not have to be powered down for forensic investigation purposes. The investigator simply creates and loads a snapshot of the target VM for analysis(live virtual instance). This behavior is especially important for scenarios in which a downtime of a system is not feasible or practical due to existing SLA. However the information whether the machine is running or has been properly powered down is crucial [3] for the investigation. Live investigations of running virtual instances become more common providing evidence data that。
毕业设计说明书英文文献及中文翻译学生姓名:学号:计算机与控制工程学院:专指导教师:2017 年 6 月英文文献Cloud Computing1。
Cloud Computing at a Higher LevelIn many ways,cloud computing is simply a metaphor for the Internet, the increasing movement of compute and data resources onto the Web. But there's a difference: cloud computing represents a new tipping point for the value of network computing. It delivers higher efficiency, massive scalability, and faster,easier software development. It's about new programming models,new IT infrastructure, and the enabling of new business models。
For those developers and enterprises who want to embrace cloud computing, Sun is developing critical technologies to deliver enterprise scale and systemic qualities to this new paradigm:(1) Interoperability —while most current clouds offer closed platforms and vendor lock—in, developers clamor for interoperability。
附件1:外文资料翻译译文大容量存储器由于计算机主存储器的易失性和容量的限制, 大多数的计算机都有附加的称为大容量存储系统的存储设备, 包括有磁盘、CD 和磁带。
相对于主存储器,大的容量储存系统的优点是易失性小,容量大,低成本, 并且在许多情况下, 为了归档的需要可以把储存介质从计算机上移开。
术语联机和脱机通常分别用于描述连接于和没有连接于计算机的设备。
联机意味着,设备或信息已经与计算机连接,计算机不需要人的干预,脱机意味着设备或信息与机器相连前需要人的干预,或许需要将这个设备接通电源,或许包含有该信息的介质需要插到某机械装置里。
大量储存器系统的主要缺点是他们典型地需要机械的运动因此需要较多的时间,因为主存储器的所有工作都由电子器件实现。
1. 磁盘今天,我们使用得最多的一种大量存储器是磁盘,在那里有薄的可以旋转的盘片,盘片上有磁介质以储存数据。
盘片的上面和(或)下面安装有读/写磁头,当盘片旋转时,每个磁头都遍历一圈,它被叫作磁道,围绕着磁盘的上下两个表面。
通过重新定位的读/写磁头,不同的同心圆磁道可以被访问。
通常,一个磁盘存储系统由若干个安装在同一根轴上的盘片组成,盘片之间有足够的距离,使得磁头可以在盘片之间滑动。
在一个磁盘中,所有的磁头是一起移动的。
因此,当磁头移动到新的位置时,新的一组磁道可以存取了。
每一组磁道称为一个柱面。
因为一个磁道能包含的信息可能比我们一次操作所需要得多,所以每个磁道划分成若干个弧区,称为扇区,记录在每个扇区上的信息是连续的二进制位串。
传统的磁盘上每个磁道分为同样数目的扇区,而每个扇区也包含同样数目的二进制位。
(所以,盘片中心的储存的二进制位的密度要比靠近盘片边缘的大)。
因此,一个磁盘存储器系统有许多个别的磁区, 每个扇区都可以作为独立的二进制位串存取,盘片表面上的磁道数目和每个磁道上的扇区数目对于不同的磁盘系统可能都不相同。
磁区大小一般是不超过几个KB; 512 个字节或1024 个字节。
Text 3 云计算的优势1.介绍云计算在讨论云计算的优势之前,先看看云计算是什么,还有它的不同类型。
云计算有很多优势,它可以让你使用基础设施和应用程序的服务,并且(或者)为象征性的收费提供存储空间。
因为这些服务项目是由云服务供应商创造和提供的,你不必为基础设施的额外使用而付费(如服务器、应用程序、操作系统等)。
我们可以定义云计算为每次使用都付费的模式。
经过请求就能得到可靠、可配置的资源,这些资源可以很快被提供、被释放——客户参与的管理程度最小。
你只为你使用的资源付费,不需建立基础设施或购买软件,这只是云计算许多优势的一个抽象概念。
任何云都有以下特点,不管是私有的还是公有的,不管它提供的服务类型是什么:1). 无论何时客户请求它能很快分配和释放资源2). 它有实时的备份,为客户提供最大的正常运行时间3). 它能够迎合客户的需求,而不需要让客户参与服务的管理接下来看看云计算的优点,主要研究在他们提供的服务基础上的不同种类的云。
2.云服务的类型软件即服务模型:这是最常见的云服务的形式。
这种服务供应者提供软件支持服务,软件是服务供应者建立的,而终端用户可以装配以适应自己的需求。
但是客户不能改变或修改软件。
在线备份服务就是一个例子。
它基本上是一个备份服务,它提供软件以帮助人们备份自己的数据。
这样,你可以使用服务而不必编码或购买软件,你只需每月或每年付费以使用这种服务项目。
平台即服务模型:它提供一个平台给客户,以满足不同目的。
比如:微软云计算操作系统提供一个平台给开发者,让他们建立、测试和主持应用程序。
这些程序可以被终端使用者使用。
终端使用者也许知道、也许不知道应用程序是通过云计算来进行的。
前面提到过,用户数据的存储空间可能会增加,也可能会缩小。
根据应用程序的要求,使用作为服务的软件,你不必建立平台。
你只需为使用服务支付象征性的费用。
基础架构即服务模型:它根据需求提供基础设施。
基础设施可以是存储服务器、应用程序和操作系统。
云计算外文文献+翻译1. 引言云计算是一种基于互联网的计算方式,它通过共享的计算资源提供各种服务。
随着云计算的普及和应用,许多研究者对该领域进行了深入的研究。
本文将介绍一篇外文文献,探讨云计算的相关内容,并提供相应的翻译。
2. 外文文献概述作者:Antonio Fernández Anta, Chryssis Georgiou, Evangelos Kranakis出版年份:2019年该外文文献主要综述了云计算的发展和应用。
文中介绍了云计算的基本概念,包括云计算的特点、架构、服务模型以及云计算的挑战和前景。
3. 研究内容该研究综述了云计算技术的基本概念和相关技术。
文中首先介绍了云计算的定义和其与传统计算的比较,深入探讨了云计算的优势和不足之处。
随后,文中介绍了云计算的架构,包括云服务提供商、云服务消费者和云服务的基本组件。
在架构介绍之后,文中提供了云计算的三种服务模型:基础设施即服务(IaaS)、平台即服务(PaaS)和软件即服务(SaaS)。
每种服务模型都从定义、特点和应用案例方面进行了介绍,并为读者提供了更深入的了解。
此外,文中还讨论了云计算的挑战,包括安全性、隐私保护、性能和可靠性等方面的问题。
同时,文中也探讨了云计算的前景和未来发展方向。
4. 文献翻译《云计算:一项调查》是一篇全面介绍云计算的文献。
它详细解释了云计算的定义、架构和服务模型,并探讨了其优势、不足和挑战。
此外,该文献还对云计算的未来发展进行了预测。
对于研究云计算和相关领域的读者来说,该文献提供了一个很好的参考资源。
它可以帮助读者了解云计算的基本概念、架构和服务模型,也可以引导读者思考云计算面临的挑战和应对方法。
5. 结论。
毕业设计附件外文文献翻译:原文+译文文献出处:Mehra P. Cloud computing: Distributed internet computing for IT and scientific research[J]. Internet Computing, IEEE, 2016, 1(5): 10-19.原文The study of cloud storage service systemMehra PAbstractCloud storage is a new concept, which developments and extensions in a cloud computing, so to understand cloud storage is the first to know about cloud computing. Cloud computing is a kind of super computing model based on Internet, in a remote data center, tens of thousands of computer and server connected to a computer cloud. Therefore, cloud computing allows you to experience even 10 trillion times a second operation ability, have such a powerful computing ability to simulate nuclear explosion, forecasting and market development trend of climate change. User through a computer, laptop, mobile phone access to the data center, operation according to their needs. With the acceleration development of the concept of cloud computing, people began to looking for a new place for huge amounts of information in cloud storage. The cloud (cloud storage) emerged from a widely attention and support. Similar to the concept of cloud storage and cloud computing, it refers to the application through the cluster, grid technology or distributed file systems, and other functions, the network of a large number of various types of storage devices set up by applying the software to work together, common external provide access to data storage and business functions of a system.Keywords: cloud storage, cloud storage service system, the HDFS1 IntroductionThe rise of cloud makes the whole IT industry in a significant change, from equipment/application centered toward centered on information and this change will cause a series of changes, and affect the technical and business mode two levels. The biggest characteristic of the cloud is a mass, high performance/high traffic and lowcost, and the biggest change is that its bring providers from sales tools to gradually according to the actual use of tools to collect fees, from selling products to selling services. Therefore, it can be said that cloud storage is not stored, but service. Cloud storage but also has the following characteristics: strong extensibility, should not be limited by the specific geographic location, based on the business component, according to the use of fees, and across different applications. The research content of this article for the study of the cloud storage service system based on HDFS, aims to build a cloud storage service system based on HDFS, solve the enterprise mass data storage problem, reduce the cost of implementing the distributed file system, promote the Hadoop technology promotion. Cloud storage is widely discussed in the present on the cloud computing concept of extension and development, to a large number of different types of storage devices in the network integration, thereby providing access to data storage and business functions. Hadoop distributed file system (HDFS) is the underlying implementation of open source cloud computing software platform Hadoop framework part, has the characteristic such as high transmission rate, high fault tolerance, can be in the form of a flow to access the data in the file system, so as to solve the access speed and security issues, achieve huge amounts of data storage management.2 Each big cloud storage products of the company2.1 The Amazon’s strategyAmazon is among the first to launch the cloud storage service enterprises. Amazon first launch a service of cloud computing is Amazon web services (Amazon web services, the AWS), the cloud computing service is composed of four core components: simple arrangement, simple storage service, elastic computing cloud and is still in the backs of the test. In August 2008, Amazon in order to enhance its efforts on the cloud storage strategy, its Internet services add "persist" function to the elastic compute cloud (ECZ).The vendor launched Elastic Block Storage (Elastic Block Storage, EBS) products, and claims that the product can through the Internet service form at the same time provide Storage and computing functions.2.2 Nirvana and CDNetworks strategyFamous cloud storage platform providers Nirvanix and content delivery network service provider CDNetworks released a new cooperation, and strategic partnerships, to provide the industry the only cloud storage and content delivery service integration platform. Use it’s located in all parts of the world 63 content distribution node, the user can store unlimited data on the Internet, and get good data protection and data security guarantee. Cooperation will bring CDNetworks in cloud storage and Nirvanix the same capacity, not only can safely store huge amounts of media content, and can rely on CDNetworks data center to deliver data anywhere in the world in real time, the two companies, said based on this partnership of cooperation, make it have better overall media delivery ability, also helps users save 80% 90% of the construction of its own storage infrastructure costs.2.3 Google's strategyThe company in this year's FO developer technical conference announced called "Google storage cloud storage services, to challenge Amazon s3 cloud storage service. Look from the function design, Google storage will refer to the Amazon s3, for existing s3 user’s switch to Google storage service. Google storage services will include RESTAPI agreement, to allow developers to download via Google account provides authentication, data backup services. In addition, Google will also to outside developers to provide network user interface and data management tools.2.4 EMC’s strategyEMC's cloud storage infrastructure solution is a kind of management system based on strategy, the service provided can create different types of cloud storage ability, for example, it can be for not paying customers to create file two copies, and stored in different locations around the world, and for paying customers to create a backup storage on October 5, and provides its all over the world access to the file of higher reliability and faster access. In software systems, Atm0S including data services, such as copying, data compression, data reduplication, with cheap standard x86 server to hundreds of terabytes of hard disk storage space.EMC has promised that it automatically configure the new storage and adaptive ability of a hardware failure, also allows the user to use W b manage service agreement and read. At present thereare three versions, Atm0S system capacity is respectively 120 TB otb, 24, and 36 orb, All of them are based on x86 servers and support gigabit or 10 gb Ethernet connection.3 Cluster development of storage technologyThe rise of cloud storage is upending the existing network storage architecture. Facing the current pet bytes of mass storage requirements, the traditional SAN or NAS will exist in the expansion of the capacity and performance bottlenecks. Such as by its physical elements (such as the number of disk drives, the connected server performance and memory size and the number of controller), can cause a lot of functional limitations (such as: the number of file system support, snapshot or copy number, etc.).Once encountered the bottleneck of storage system, it will constantly encourage users to upgrade to a bigger storage system and add more management tools, thus increasing the cost. Cloud storage the service mode of the new storage architecture is demanded to keep very low cost, and some existing high-end storage devices are obviously cannot meet this need. From the perspective of the practice of Google company, they are not used in the existing cloud computing environment SAN architecture, but use, is a scalable distributed file system GFS) this is a highly efficient cluster storage technology.GFS is a scalable distributed file system, used in large, distributed, on a visit to a large amount of data applications. It runs on ordinary PC, but can provide strong fault tolerance, can give a large number of users with the overall performance of the service. Cloud storage 130] is not stored, but service. Wan and the Internet like a cloud, the cloud storage for users, not referring to a particular device, but is a by many a collection of storage devices and servers. Users use the cloud storage, not using a storage device, but use, is the entire cloud storage system with a data access service. So strictly speaking, the cloud is not stored, but a service. Cloud storage is the core of application software combined with a storage device, by applying the software to realize the change of the service to the storage device.4 Cloud storage system analysesCompared with the traditional storage devices, cloud storage is not only hardware, but a network devices, storage devices, servers, applications, public accessinterface, access, and the client program such as a complex system composed of multiple parts. Parts storage device as the core, through the application software to provide access to data storage and business services. The structure of cloud storage system model consists of four layers.(1)The storage layerStorage layer is the most basic part of the cloud storage. Storage devices can be a fiber channel storage devices, or other storage devices. Storage devices are often large number of cloud storage and distribution of many different areas, between each other through the wide area network, Internet or fiber channel network together. Storage devices is a unified storage management system, can realize the logic of storage virtualization management, more link redundancy management, as well as the hardware equipment condition monitoring and fault maintenance.(2)The basic managementBased management is the core part of the cloud storage, is also the most difficult part of the cloud storage. Based management through cluster and grid computing, distributed file system such as technology, realize the cloud storage between multiple storage devices in the work, make multiple storage devices can provide the same service, and to provide better data access performance, bigger and stronger content distribution system, 1391, data encryption technology to ensure the data in the cloud storage will not be access by unauthorized users, at the same time, through a variety of data for disaster and techniques and measures can ensure that data is not lost in the cloud storage, ensure the security and stability of the cloud storage itself.(3)The application of the interface layerApplication of the interface layer is the most flexible part of the cloud storage. Different cloud storage operation unit can be according to actual business types, different application service interface, with the application of different services. Such as video monitoring application platform, network hard disk reference platform, the remote data backup application platform, etc.(4) Any an authorized user can access layer through a standard utility application login interface to cloud storage system, the cloud storage service. Cloud storageoperation services, cloud storage provide different type of access and the access method.译文云存储服务系统研究Mehra P摘要云存储是在云计算(cloud computing)概念上延伸和发展出来的一个新的概念,因此,要了解云存储首先要了解云计算。