外文文献翻译译稿与原文
- 格式:doc
- 大小:329.00 KB
- 文档页数:17
乡村旅游外文翻译文献乡村旅游外文翻译文献(文档含英文原文和中文翻译)翻译:Factors for success in rural tourism tourism developmentSince the 1970s,economic restructuring and farm crisis have reduced rural communities' economic development options, making older development strategies less viable and forcing many to look for nontraditional ways to sustain themselves. One of the most popular nontraditional rural development strategies has been tourism and its associated entrepreneurship rural development strategies has been tourism and its associated entrepreneurship opportunities because oftourism's ability to bring in dollars and to generate jobs and support retail growth. The purpose of this study was to identify and examine those factors that have helped rural communities successfully develop tourism and its entrepreneurship opportunities. Several focus groups were conducted with local businesspersons and leaders in six rural Illinois communities. The results clearly demonstrate the importance of the community approach to tourism development and that rural tourism development and entrepreneurship cannot work without the participation and collaboration of businesspersons directly and indirectly involved in tourism.Since the 1970s, economic restructuring and the farm crisis have severely reduced rural communities’economic opportunities.Economic restructuring has caused a loss of rural manufacturing plants and many jobs. The 1980s farm crisis in the Midwest also led to a decline in the numbers of farmers and restructured farm ownership, forcing some farm families to augment their incomes with off-farm jobs, to depart farming, or to declare bankruptcy. The farm crisis and the loss of manufacturing jobs had substantial ripple effects in rural communities. As rural joblessness rates rose above urban levels, real income growth stagnated in rural areas (Sears and Reid 1992). Many stores and agribusinesses disappeared from small rural towns. Not surprisingly, a 1992 statewide survey in Illinois found that 39% of rural residents perceived their economic prospects as worsening (Walzer 1993).These changes limited rural communities’economic development options, making older development strategies such as manufacturing less viable and forcing many to look for nontraditional ways to sustain themselves . One of the most popular nontraditional rural developmentstrategies has been tourism and its associated entrepreneurship opportunities (Edgell and Harbaugh 1993; Luloff et al. 1994). Rural areas have a special appeal to tourists because of the mystique associated with rural areas and their distinct cultural, historic, ethnic, and geographic characteristics (Edgell and Harbaugh 1993). Rural tourism also is less costly and easier to establish than other rural economic development strategies such as manufacturing. Rural tourism can be development strategies such as manufacturing. Rural tourism can be developed locally with participation from local government and small businesses, and its development is not necessarily dependent on outside firms or companies. Although tourism can be expensive to develop in certain cases (e.g., large resort areas) or can involve large firms and chains, rural tourism can be developed with relatively little investment credit, training, and capital. Hence, rural tourism can be less costly to develop as compared to other economic development strategies; additionally, rural tourism need not involve dependency on outside firms and their decisions on whether they want to be in an area. Rural tourism provides a base for these small businesses that might not otherwise be in rural communities because of their small populations. Tourism particularly helps two types of small businesses in rural areas—those directly involved in tourism (e.g., attractions and hotels/motels) and those indirectly involved in tourism (e.g., gas stations and grocery stores). Additionally, rural tourism works well with existing rural enterprises such as farms (e.g., U-Pick farms) and can generate important secondary income for farm households (Oppermann 1996).Nonetheless, rural tourism remains one of the few viable economic options for rural communities .Like other economic development strategies, rural tourism requires several components to be successful.Tourism development involves(1) attractions: the natural and manmade features both within and adjacent to a community; (2) promotion:the marketing of a community and its tourism attractions to potential tourists;(3) tourism infrastructure: access facilities (roads, airports, trains, and buses),water and power services, parking, signs, and recreation facilities; (4) services: lodging, restaurants, and the various retail businesses needed to take care of tourists’ needs; (5) hospitality: how tourists are treated by both community residents and employees in tourism businesses and attractions (Gunn 1988). Left out of this list are tourism entrepreneurs and their role in fostering these components.While the above components and a community’s assets are clearly important to tourism development, only the widespread participation and contribution of rural tourism entrepreneurs can ensure a broad-based foundation for successful tourism development. A research literature has emerged on how to best facilitate the development of tourism. One view, drawing heavily on the economic literature, argues that tourism and its associated entrepreneurship opportunities are best developed by helping and creating individuals businesses and then letting them compete in the marketplace for a review and description of this view). This view, however, has been critiqued because (1) it views tourism and tourism-related businesses as isolated from the larger community and its issues;(2) it does not recognize the interdependence of the various sectors and actors involved in tourism; and (3)most small tourism business, especially those in rural areas, do not have the individual resources to promote either themselves or the community as a tourist product (Gunn 1988; Murphy 1985; Palmer and Bejou 1995).Opposing this view is the community approach to tourism development and entrepreneurship (Murphy 1985). As its name implies, the approach argues that tourism is a community product and that, along with entrepreneurial skills and the presence of tourist businesses, it is also necessary to have the community and local capabilities (e.g., local leadership and formal and informal networks)directly involved in tourism development and promotion effort (Murphy 1985). While the community approach may be an effective way to develop and promote tourism, creating the necessary intercommunity cooperation and collaboration is a complex and difficult process. Businesses are asked to share resources while simultaneously competing. Local governments may see collaborating to develop tourism as risky, or they may be worried about losing control over local decision making (Huang and Stewart 1996; Jamal and Getz 1995). Because of these problems, research on collaboration and those factors that allow for community development of tourism is needed . The purpose of the present study is to identify and examine those factors that help rural communities successfully develop tourism and its entrepreneurship opportunities. The present study makes an additional contribution to the research literature by including rural tourism entrepreneurs, an overlooked group in rural tourism research (Stokowski 1990).乡村旅游发展成功因素作者:苏珊娜威尔逊,丹尼尔·朱莉和约翰国籍:美国出处:SAGE 出版社中文译文:20世纪70年代以来,经济体制的改革和农业危机使得农村社区经济发展的选择减少,同时促使老年人发展战略可行性降低,迫使人们去寻找非传统的方式来维持生计。
外文文献原稿和译文原稿1. IntroductionOver the past two decades, organizations of all types have increasingly acknowledged the importance of customer satisfaction and loyalty. The marketing literature suggests that the long term success of a firm is clearly based on its ability to rapidly respond to changing customer needs and preferences (Narver &Slater, 1990; Webster, 1992). A key motivation for the increasing emphasis on customer satisfaction is that higher customer satisfaction can lead to have a stronger competitive position resulting in higher market share and profitability (Fornell, 1992), reduced price elasticity, lower business cost, reduced failure cost, and mitigated cost of attracting new customers (Chien, Chang, & Su, 2003).The principal focus of this study is on evaluating the efficiency of customer satisfaction and loyalty (CS&L) for existing mobile phone brands in Turkish mobile phone sector. Since the early1990s, with the launch of the mobile phones, there has been a remarkable development both in their product sophistication and their rapid and widespread adoption. With more than three billion subscribers around the world, the extent of mobile phone diffusion in emerging markets has been increasingly larger than that in developed countries (Kalba, 2008). Turkey, being one of the fastest emerging market economies in the world, adopted mobile phone technology in 1994. Since then, there has been a considerable increase in the level of mobile phone ownership, where the number of mobile phone users in the country is expected to reach around70 million by the end of 2013, representing a penetration rate of over 90% (RNCOS, 2010). The significant rise in mobile phone usage can partially be attributed to the fact that Turkey has the youngest population in Western Europe. Turkey currently has the 6th largest young mobile phone user base in the world, with more than 11million subscribers underthe age of 25, providing a very lucrative market for mobile phone companies (Euro monitor International,2010). It should however be noted that the penetration in this market at present is still below the EU average, indicating that the mobile phone sector is not saturated yet, and there is still space for new investors. Currently, there exist nearly more than 10 major mobile phone companies operating in the Turkish mobile phone sector, each having a relatively large product line. As of 2010, the top five mobile phone brands were Nokia, Samsung, LG, Motorola and Sony Ericsson and together they account for nearly 75% of overall market sales. As a new comer, phone is rapidly increasing its market share, but as of the start of this study, did not have a significantly large presence. In terms of market share, Nokia has been undisputedly the market leader (36.4% of sales) with Samsung featuring second (19.5%) and LG ranking third (10.1%) (Patron Turk,2010).Commensurate to its widespread diffusion globally, there has been a growing worldwide academic interest in mobile phone usage which focuses mainly on examining its contribution to social life, user preferences and its ergonomic features (Bag chi, Kirs, & Lopez, 2008). A number of empirical studies were also conducted within the context of Turkish mobile phone sector. The topics of these studies ranged from examining motivation of use (Dedeoglu,2004; Oscan & Kodak, 2003) to mobile phone selection (Isiklar &Buyukozkan, 2007), from customer satisfaction (Turkyilmaz &Ozkan, 2007) to brand loyalty (Simsek & Noyan, 2009).The methodology used in study to evaluate the relative CS&Lefficiency of mobile phone brands is based on data envelopment analysis (DEA). The traditional DEA technique has long been utilized as an invaluable tool in the field of operations research and management science to solve problems in wide range of industries(Hu, Lai, & Huang, 2009; Lee, 2009; Lin, Lee, & Chiu, 2009) as well as in not-for-profit organizations (Mahajan, 1991; Wu, Liang, &Chen, 2009; Zhang, Huang, Lin, & Yu, 2009); but its diffusion into the field of marketing and related disciplines has been relatively slow. For instance, in the marketing field, DEA has recently been employed as a powerful tool for data analysis in measuring efficiency in retailing sector (Charnes, Cooper, Learner, & Phillips,1985; Donthu & Yoo, 1998; Keh, 2000; Keh & Chu, 2003; Thomas,Barr, Cron, &Slocum, 1998), evaluating website marketing efficiency(Shuai & Wu, 2011), benchmarking marketing productivity(Donthu, Hershberger, & Osmonbekov, 2005; Kamakura, Ratchford,& Agrawal, 1988), and measuring relative market efficiency(Murthi, Srinivasan, & Kalyanaram, 1996) or service quality(Athanassopoulos, 1997; Soteriou & Staurinides, 1997). The assessment of CS&L has always been a major research item on the agenda of researchers in the marketing and related fields, because the issue of how efficiently a firm manages its marketing processes and their relationship with their customers is central to its ability to gain competitive edge vis-à-vis its rivals. The DEA approach adopted in this study illustrates how differences in CS&L efficiency between various mobile phone brands can be ascertained empirically, and thus helps management determine proper policies and courses of action.The rest of the paper is organized as follows. Section 2 reviews the recent literature on customer satisfaction and customer loyalty studies. Section 3 provides an in-depth description of our research methodology. Section 4 presents the results of our analysis. The last section (Section 5) summarizes our findings, describes managerial implications of the study and provides the concluding remarks.2. Background literatureWhile customer satisfaction has been defined in various ways, the high-level conceptualization that appears to have gained the widest acceptance states that satisfaction is a customer’s post purchase evaluation of a product or service (Cronin & Taylor,1992; Westbrook & Oliver, 1991). Customer satisfaction is also generally assumed to be a significant determinant of repeat sales, positive word-of-mouth, and customer loyalty. It has also long been considered as one of the key antecedents of creating brand loyalty (Cronin, Brady, & Hult, 2000; Dick & Basu, 1994; Fornell,Michael, Eugene, Jaesung, & Barbara, 1996; Syzmanski & Henard,2001). Satisfied customers return and buy more, and they tell other people about their experiences, both positive and negative (Fornellet al., 1996).Building on Hirschman’s (1970) exit-voice theory, weakly dissatisfied consumers would be of primary importance to a firm. While strongly dissatisfied consumers generally choose the exit option (i.e., they leave the firm), the weakly dissatisfied customers tendto stay loyal to the firm and rather employ the voice option, which implies overt complaints as an attempt to change the firm’practices or offerings (Fornell &Wernerfelt, 1988). Thereby, proper handling of customer complaints may ensure that weakly dissatisfied consumers remain loyal, and serve as an exit barrier (Fornell,1992; Halstead & Page, 1992). The impact of loyal customers is considerable; for many industries the profitability of a firm increases proportionally with the number of loyal customers and up to 60% of sales to new customers can be attributed to the word of mouth referrals (Reichheld & Sasser, 1990).Within the existing literature on customer satisfaction research, various customer satisfaction models were developed based on a cumulative view of satisfaction. To this end, a number of customer satisfaction indices (CSIs) were designed with most prominent of those being Swedish Customer Satisfaction Barometer (SCSB), the American Customer Satisfaction Index (ACSI) and European Customer Satisfaction Index (ECSI). Of these CSIs, we employed the ECSI model as the backbone of our CS&L efficiency model in this study due to its recent popularity in the literature and its comprehensiveness in CS&L coverage. The ECSI is a structural model based on the assumptions that customer satisfaction is derived by a number of factors such as perceived quality, perceived value, expectations of customers, and image of a firm. These factors are the antecedents of overall customer satisfaction (Turkyilmaz &Ozkan, 2007). The model also estimates the results when a customer is satisfied or not. The four antecedents of customer satisfaction may also have direct effects on customer loyalty(Johnson, Gustafson, Andreessen, Lervik, & Cha, 2001). Each construct in the ECSI model is a latent construct which is operational zed by multiple indicators (Chien et al., 2003; Fornell,1992). The underlying constructs of the ECSI model are explained as follows:The image construct evaluates the underlying image of the company. Image refers to the brand name and the kind of associations customers obtain from the product/company (Andreassen &Lindestad, 1998). Martensen, Kristiansen, and Rosholt (2000)argue that image is an important dimension of the customer satisfaction model. Image is a consequence of being reliable,professional and innovative, having contributions to society, and adding prestige to its user. It is anticipated that image has a positive effecton customer satisfaction, customer expectations and customer loyalty.Customer expectations are the consequences of prior experience with the company’s products (Rotondaro, 2002). This construct evaluates customer expectations for overall quality, for product and service quality, and for fulfillment of personal needs. The customer expectations construct is expected to have a direct and positive relationship with customer satisfaction (Anderson, Fornell, &Lehmann, 1994).Perceived quality is evaluation of recent consumption experience by the market served. This construct evaluates customization and reliability of a given product or service. Customization is the degree to which a product or service meets a customer’s requirements, and reliability is the degree to which firm’s offering is reliable, standardized, and free from deficiencies. Perceived quality is expected to have a positive effect on customer satisfaction (Fornellet al., 1996).Perceived value is the perceived level of product quality relative to the price paid by customers. Perceived value is the rating of the price paid for the quality perceived and a rating of the quality perceived for the price paid (Fornell et al., 1996). Perceived value structure provides an opportunity for comparison of the firms according their price-value ratio (Anderson et al., 1994). In the model, perceived value is expected to have a positive impact on satisfaction.Customer satisfaction construct indicates how much customers are satisfied, and how well their expectations are fulfilled. This construct evaluates overall satisfaction level of customers, fulfillment of their expectations, and company’s performance versus the ideal provider.Customer loyalty is the ultimate factor in the ECSI model. Loyalty is measured by repurchase intention, price tolerance and intention to recommend products or services to others. It is expected that better image and higher customer satisfaction should increase customer loyalty.3. MethodologyThis section presents the research methodology adopted in this study. The following subsections explain the survey instrument, the data collection procedure, and the DEA model.3.1. Survey instrumentThe DEA model of CS&L, which is shown in Fig. 1, consists of the aforementioned constructs which are based on previous research and prominent theories in the field of consumer behavior. The constructs of the CS&L model are unobservable (latent) variables indirectly described by a set of observable variables which are called manifest variables or indicators. The constructs and their constituent items are shown in Table 1. The use of multiple measures for each construct increases the precision of the estimate as compared to an approach of relying on a single measure. In our CS&L efficiency model, all four antecedents of customer satisfaction and loyalty which include image, customer expectations, perceived quality and perceived value were treated as input variables, while the two constructs, namely customer satisfaction and customer loyalty were considered as output variables.The survey questionnaire was designed using a three-step process. First, the consumer behavior literature was extensively reviewed for the manifest variables. Secondly, the questionnaire items were prepared in Turkish and refined through a series of discussions with two senior marketing managers of a prominent mobile phone company and a number of experienced academics in the field of consumer behavior. Finally, the survey questionnaire was subjected to extensive pre-testing and refinement based on a pilot study of 30 mobile phone users. Feedback from this pilot study indicated that some questions were ambiguous, difficult to understand,or irrelevant for mobile phone sector. This pilot study also served as a practical exercise for interviewers. The final questionnaire contained a total of 23 items pertaining to the CS&L. These23 items appeared to have face validity as to what should be measured. All the items were measured on 10-point scales, with anchors ranging from 1 denoting a very negative view and 10indicating a very positive view. Relying on 10-point scales enables customers to make better discriminations (Andrews, 1984).译文1.介绍在过去的二十年中,所有类型的组织都越来越多地承认了客户满意度和忠诚度的重要性。
外文文献原稿和译文原稿MechanicalandRegenerativeBrakingIntegrationforaHybridElectricVehicleAbstract,namelyanelectricmotorandinternalcombustionengine,whichallowtheelimina tionofidling,,theaddedcostofthehybridelectricsystemhashinderedthesalesofth esevehicles.JapanNorth America automotive companieshavedevelopedandreleasedforsaletheirownhybridelectric unpredictablegasprices,thesalesofhybridelectricvehicleshaveincreaseddramat ically inrecentyears.2.1.1HybridConfigurationsForthepast100yearstheobjectiveofthehybridhasbeentoextendtherangeofelec tricvehiclesandtoovercometheproblemoflongrechargingtimes35.Therearethree ,par allelhybridsandseries/parallelhybrids.Eachconfigurationhasitsadvantagesanddisadvantageswhichwillbediscussedi nthefollowingsections.SeriesHybridsInserieshybridsthemechanicaloutputfromtheinternalcombustionengineisuse dto driveageneratorwhichproduceselectricalpowerthatcanbestoredinthebatteriesor highpowersystemssuchaslargetrucksorlocomotivesbutcanalsobeusedforlowerpowe rpassengervehicles18.2.1.2‘degreeofhybridization’toquantifytheelectrica lpowerpotentialofthesevehicles.ThedegreeofhybridizationrangesfromDOH=0foraconventionalvehicletoDOH=1f oranallelectricvehicle25.Asthedegreeofhybridizationincreases,asmallerICEcanbeusedandoperatedclosertoitsoptimumefficiencyforagreaterproportionoftheti me,.MicroHybrid MicrohybridshavethesmallestdegreeofhybridizationandusuallyconsistofanintegratedstartergeneratorISG2.1.31500 kg100 km/h0 km/h0 km/h50 km/h2.1.42.1.5译文混合动力电动汽车机械和再生制动的整合摘要为了减少对环境的污染和车辆的燃油消耗,混合动力电动汽车已经成为汽车工业的首选方法;混合动力电动汽车通过使用由电动马达和内燃发动机组成的混合动力系统来达到减少环境污染和燃油消耗的目的;混合动力系统消除了怠速,使发动机以一种更有效的方式运行,增加了再生制动的使用;但是,混合动力的成本的增加阻碍了这些车辆的销售;在这里提出一个更具成本效益的电液制动系统的设计;该系统使用电控机械结合的控制方式控制制动助力器产生的推动力,并有足够的时间反应;这个系统使驾驶员清楚地了解机械和再生制动力矩的混合,使再生制动力系统得到有效的控制;一个系统化的设计过程是其次,重点在于展示概念设计方案的可行性和使用虚拟和实物模型的初步设计功用;虚拟和实物模型的结合使用成为验证和开发系统的强大工具,本文将介绍和讨论在设计过程中模型所起到的作用;因为在设计过程中设计者可以获得相关的经验,提倡学生设计实物模型,以提高学生的学习经验;很明显,这正是本文所要提出的;现代混合动力电动汽车随着油价的上涨和环境保护意识的提高,消费者和政府迫使汽车行业开始生产省油和对环境污染小的汽车;一个有前景的方法就是现在实行的混合动力电动汽车;混合动力汽车指的是有两个或两个以上动力来源的车辆;混合动力汽车动力的来源可能有很多的不同,但是现在混合动力汽车最常见的布局是由内燃发动机和电动马达,能量储存系统共同输出动力,这样的车辆就叫混合动力电动汽车;汽车可以同时使用发动机和电动马达输出的动力,从而可以提高汽车的使用性能和效率,进而又可以提高燃油经济性,减少废气的排放,同时还能满足消费者对汽车性能的要求;1997年,丰田成普瑞斯为了第一款混合动力电动汽车,在日本进行了批量生产;本田公司花费了三年的时间进行混合动力电动车的生产,然后进军北美市场;丰田普瑞斯在北美发行几个月后,本田Insight紧随其后也在北美进行发行;混合动力电动车具有再生制动系统的独特优势;在制动过程,通常用于动力输出的电动马达,可以起到发电机的功用,把汽车的动能转化为蓄电池的电能,而不会转化为热能浪费掉;转换的电能可以储存到蓄电池中,然后可以作为电动马达驱动汽车使用的能量;考虑到蓄电池能量密度时,动能转换为电能这个过程就更加重要了;能量密度是指单位体积或质量下能量储存系统所储存的能量;为了说明这一点,我们可以做个对比,4.5公升的汽油通常可以维持一辆汽车行驶50千米;而要把相同的能量储存到蓄电池中,则需要一个质量约为270千克的铅酸蓄电池;这就说明了在汽车行驶过程中能够有效地储存再生制动系统产生的能量的重要性,从而可以保证在提高混合动力电能车性能的前提下,不至使能量储存系统所占体积过大;再生制动系统研究范围本文所提出的再生系统的研究范围是研究再生制动系统和机械制动系统之间相互作用的关系,目的是设计开发出一个低成本的再生制动系统,从而可以应用到未来经济型的混合动力电动汽车上;这个系统可以根据驾驶员的需要进而控制再生制动系统和机械制动系统产生的制动力矩的结合,还应该保证这个过程的平顺性和安全性;再生制动力矩是通过使用的异步电动机的矢量控制算法进行控制的;但是,独立地控制制动踏板产生的机械制动力矩,同时又要保持机械制动系在再生制动系统失效后起到备用作用,这是一个很大的难题;为了解决这个问题,需要研究一个通过减少制动主缸内制动液压来来控制机械制动系统产生的制动力矩的制动系统;混合电动汽车概述混合动力电动车已经成为了可以在短时间内减少汽车污染排放和提高燃油经济型的解决方法之一;在过去的10年几乎所有的主要汽车公司都已经向公众发行销售自己的混合动力电动汽车,混合动力电动汽车的普及和销售在这个世纪有了很明显的增长,随着不可预测的汽油价格的增长和对环境保护的关注,混合动力电动汽车的销售将在最近几年内急剧增长;2.1.1混合动力装置在过去100年来混合动力的研究目标是延长电动汽车的使用寿命,解决蓄电池的长期充电问题;在目前市场,现在主要有三种混合动力装置,这些混合动力装置为串联混合动力,并联混合动力,串并联混合动力;每一种动力装置都有其优点和缺点,这将在以后的章节进行讨论;串联混合动力串联混合动力汽车使用发动机输出的动力来驱动发电机产生电能,这些电能可以储存在蓄电池中,也可以用来驱动电动马达来驱动汽车;在串联混合动力汽车上,发动机和驱动轮之间没有直接的机械连接,串联混合动力往往在高功率系统中使用,如大型货车或火车,也可以应用到低功率的客运车辆上;发动机输出的机械能和蓄电池输出的电能可以通过电子控制器进行控制接合,然后这个电子控制器通过比较驾驶员所需的动力和汽车车速,电动马达输出的转矩,从而决定每个动力源驱动汽车行驶所要输出的能量;在制动过程中,这个电子控制装置可以使电能输出模式转换为再生模式,直接把再生制动系统产生的电能储存在蓄电池内;按照这种布置方式进行设计有很多的优点;发动机可以保持高效率的运行,使发动机产生的电能在蓄电池和驱动马达之间得到分配;发动机在其最高效率的工况下运行,排放可以大大降低,燃烧每体积的燃料可以产生更多的电能;因为串联动力装置结构简单且成本低,这种动力装置很容在汽车上落实;并联混合动力在并联混合动力汽车中,发动机输出的机械功传到变速箱中;发动机输出的机械功和电动马达输出的功在变速箱内进行机械式的接合,混合的机械功用于驱动汽车行驶;在这种混合动力装置结构中,发动机和驱动轮之间有直接的机械连接;在串联混合动力装置中,电子控制器通过比较驾驶员所需的动力和汽车车速,电动马达输出的转矩,从而决定每个动力源驱动汽车行驶所要输出的能量,以满足汽车行驶性能,获得最佳的效率;正如串联混合装置一样,并联混合动力也以相似的方法控制再生制动;并联混合动力装置通常应用到低功率的电动车中,这两种驱动力可以同时使用,提供更高的行驶性能;与串联混合动力系统相比,并联混合动力系统有很多优势;其中最重要的一项优势是效率高,因为在并联混合动力中,电能和机械能只需转换一次,而在串联混合动力中,电能和机械能需要两次转换;由于并联混合动力可以使发动机和电动马大产生的动力同时结合起来,在不损失汽车行驶性能的前提下,可以使用体积小的电动马达,同时也降低了油耗和排放;最后,并联混合动力汽车在行驶过程中只需使发动机运行,而不需要另一个发电机为蓄电池充电;串、并联混合动力串并联混合动力装置结合了串联和并联动力装置的特点;这种混合方式汽车通过使用动力分配装置来控制双动力源电动马达输出动力,发动机输出动力或者两者同时输出驱动汽车行驶;虽然这种装置形式可以获得串联混合动力装置和并联混合动力装置的优点,因为考虑到汽车实际行驶可能性,这种装置的控制算法会变得非常复杂; 2.1.2混合度现在道路上行驶的混合动力电动汽车大多是串联混合动力,并联混合动力,或者串并联混合动力,因此定义一个‘混合度’变量来评价混合动力电动汽车的电能潜能是非常有意义的;混合度变从传统车辆DOH=0到所有电动车DOH=1之间变化,随着混合度的增加,在汽车上可以使用一个比较小的发动机,同时发动机可以在最接近最佳效率的工况下运行很长的时间,这样就可以减少燃油的消耗和废气的排放;电动马达输出的功用P表emP表示;示,发动机输出的功用ice微混合动力微混合指的是最小混合度,通常是由一个连接到发动机曲轴的综合起动发电机组成;在加速和怠速过程中,综合起动发电机使发动机处于关闭状态,从而节约燃油;加速时,在燃油喷入汽缸之前,综合起动发电机使发动机的曲轴加速旋转;在加速过程中,综合起动发电机对发动机起动协助的作用,在制动过程中,综合起动发电机还可以作为发电机向蓄电池充电;和非混合动力汽车相比,微混合动力汽车的燃油经济性可以提高10%左右;轻混合动力轻混合动力和微混合动力结构相似,有一点不同的是其综合起动发电机是经过改进的,其输出的动力可以超过20KW;但是,轻混合动力的能量储存系统只能储存1KWh左右的能量;轻混合动力汽车只有一个很短的纯电动续航能力,但是可以在加速过程中给发动机提供很大的辅助作用;轻混合动力中的电子元件要比微混合动力中的电子元件复杂的多,且在汽车行驶过程中发挥着更大的作用;和非混合动力的汽车相比,轻混合动力汽车的燃油经济性可以提高20%-25%左右;全混合动力在全混合动力汽车上不再使用综合起动发电机,取代它的是一个独立的电动马达和交流发电机、起动机,这些装置也可以起到综合起动发电机的作用;电动马达可以独立驱动汽车行驶,尤其是在城市道路上走走停停的行驶;能量储存系统也得到了改进,这样就提高了汽车纯电动续航能力,减少了发动机的体积,从而提高燃油经济性和减少排放;与非混合动力汽车相比,全混合动力汽车的燃油消耗量可以减少40%-50%;插电式混合动力插电式混合动力汽车在结构上和全混合动力汽车相似,不同的是插电式混合动力汽车有一个比较大的能量储存系统,可以通过与外部电源连接进行充电;在蓄电池储存能量范围内,可以通过电动马达来驱动汽车行驶,但是当蓄电池的能量降到一定水平后,其运行形势就和全混合动力一样了;2.1.3再生制动原理混合动力电动汽车最重要的特点是可以回收大量的制动能量;在制动过程中,电动马达可以作为发电机来运行操作,将制动过程中的动能转换为电能储存到蓄电池中,这些电能就可以被汽车重复使用;但是,车辆的制动性能就将影响到汽车的安全性;在紧急制动状态下,汽车的制动距离要尽可能的短,还要保证制动时汽车有较好的方向稳定性;汽车具有较好的方向稳定性,就需要控制车轮的制动力分配;一般来说,制动时所需的制动力矩比电动马达产生的制动力矩大得多;因此,机械制动系统需要和电子再生制动系统同时存在,这就需要适当的设计以保证制动时的操作稳定性,不至于影响到汽车的安全性;制动时能量消耗由公式可得,一个质量为1500Kg的汽车以100km/h初速度制动到完全停止,需要消耗的动能;如果这些能量的25%可以通过再生制动系统进行回收,当忽略制动和加速过程中的空气阻力,机械摩擦和滚动阻力,假设电动马达的工作效率100%,利用公式可以估算出,这些能量可以使汽车从0km/h加速到50km/h.这就表明,当汽车行驶在城市道路上,汽车不停加速和制动,混合动力电动车的燃油经济性可以大大增加;需要注意的是,制动能量的回收量受到马达的型号和能量转换率的限制;2.1.4再生制动系统目前,通常使用的有两种再生制动方法;这些方法通常称为串联再生制动和并联制动,每种制动策略都有其优点和缺点,本文对此将进行具体讨论;并联再生制动在并联再生制动系统中,电动马达和机械制动系统同时工作,从而使汽车减速;因为机械制动系统不能独立的控制制动力,使制动时的能量转换为热能而不是电能,因此这不是最有效地再生制动方法;但是并联再生制动结构简单成本低,这就成为其一大优势;并联再生制动的机械制动系统只需要稍加修改,而且电动马达的控制算法也可以很容易在汽车上实现;这种制动方法还有一个额外的优势,当再生制动系统发生故障时,机械制动系统可以起到备用的作用;串联再生制动在串联再生制动中,电动马达只有在制动时才起作用;只有当电动马达和能量储存系统无法接受更多制动时所需的能量时,再生制动系统才起作用;串联再生制动需要独立的控制制动力矩,串联再生制动可以高效率的把动能转换为电能,这是其一项优势;但是它的不足之处在于,制动系统结构复杂,成本高;这种制动方式需要制动踏板模拟器,制动系统也需要重新设计,这都会增加其制造成本;因为制动系统需要装有传感器和信息处理器,这就会增加了结构的复杂度;2.1.5目前的再生制动系统目前大多数混合动力电动汽车的再生制动系统都是比较昂贵的电液制动系统;再生制动系统使用制动踏板模拟器来建立驾驶者的制动需求,这个制动踏板模拟器与液压制动电路独立分开;这样再将制动需求按照一定比例转换为再生制动和机械制动需求,然后将机械制动需求发送到由高压液压泵,蓄能器和比例控制阀的系统;比例控制阀根据制动需求,控制制动液以一定的预定值流到每个车轮的制动轮缸中;。
广东工业大学华立学院本科毕业设计(论文)外文参考文献译文及原文系部城建学部专业土木工程年级 2011级班级名称 11土木工程9班学号 23031109000学生姓名刘林指导教师卢集富2015 年5 月目录一、项目成本管理与控制 0二、Project Budget Monitor and Control (1)三、施工阶段承包商在控制施工成本方面所扮演的作用 (2)四、The Contractor’s Role in Building Cost Reduction After Design (4)一、外文文献译文(1)项目成本管理与控制随着市场竞争的激烈性越来越大,在每一个项目中,进行成本控制越发重要。
本文论述了在施工阶段,项目经理如何成功地控制项目预算成本。
本文讨论了很多方法。
它表明,要取得成功,项目经理必须关注这些成功的方法.1。
简介调查显示,大多数项目会碰到超出预算的问……功控制预算成本.2.项目控制和监测的概念和目的Erel and Raz (2000)指出项目控制周期包括测量成……原因以及决定纠偏措施并采取行动。
监控的目的就是纠偏措施的。
.。
标范围内。
3.建立一个有效的控制体系为了实现预算成本的目标,项目管理者需要建立一……被监测和控制是非常有帮助的。
项目成功与良好的沟通密。
决( Diallo and Thuillier, 2005).4.成本费用的检测和控制4.1对检测的优先顺序进行排序在施工阶段,很多施工活动是基于原来的计……用完了。
第四,项目管理者应该检测高风险活动,高风险活动最有。
..重要(Cotterell and Hughes, 1995)。
4.2成本控制的方法一个项目的主要费用包括员工成本、材料成本以及工期延误的成本。
为了控制这些成本费用,项目管理者首先应该建立一个成本控制系统:a)为财务数据的管理和分析工作落实责任人员b)确保按照项目的结构来合理分配所有的……它的变化-—在成本控制线上准确地记录所有恰..。
外文文献翻译译稿1卡尔曼滤波的一个典型实例是从一组有限的,包含噪声的,通过对物体位置的观察序列(可能有偏差)预测出物体的位置的坐标及速度。
在很多工程应用(如雷达、计算机视觉)中都可以找到它的身影。
同时,卡尔曼滤波也是控制理论以及控制系统工程中的一个重要课题。
例如,对于雷达来说,人们感兴趣的是其能够跟踪目标。
但目标的位置、速度、加速度的测量值往往在任何时候都有噪声。
卡尔曼滤波利用目标的动态信息,设法去掉噪声的影响,得到一个关于目标位置的好的估计。
这个估计可以是对当前目标位置的估计(滤波),也可以是对于将来位置的估计(预测),也可以是对过去位置的估计(插值或平滑)。
命名[编辑]这种滤波方法以它的发明者鲁道夫.E.卡尔曼(Rudolph E. Kalman)命名,但是根据文献可知实际上Peter Swerling在更早之前就提出了一种类似的算法。
斯坦利。
施密特(Stanley Schmidt)首次实现了卡尔曼滤波器。
卡尔曼在NASA埃姆斯研究中心访问时,发现他的方法对于解决阿波罗计划的轨道预测很有用,后来阿波罗飞船的导航电脑便使用了这种滤波器。
关于这种滤波器的论文由Swerling(1958)、Kalman (1960)与Kalman and Bucy(1961)发表。
目前,卡尔曼滤波已经有很多不同的实现。
卡尔曼最初提出的形式现在一般称为简单卡尔曼滤波器。
除此以外,还有施密特扩展滤波器、信息滤波器以及很多Bierman, Thornton开发的平方根滤波器的变种。
也许最常见的卡尔曼滤波器是锁相环,它在收音机、计算机和几乎任何视频或通讯设备中广泛存在。
以下的讨论需要线性代数以及概率论的一般知识。
卡尔曼滤波建立在线性代数和隐马尔可夫模型(hidden Markov model)上。
其基本动态系统可以用一个马尔可夫链表示,该马尔可夫链建立在一个被高斯噪声(即正态分布的噪声)干扰的线性算子上的。
系统的状态可以用一个元素为实数的向量表示。
外文文献原稿和译文原稿The water level control circuit designWater source total ranks sixth in the world, per capita water resources is only a quarter of the world per capita consumption, and geographical distribution is very uneven, the vast region north of the Yangtze River, northin most parts of the medium-sized cities in the dry state, water shortage has become an important factor restricting China's economic development. Reasonable use of water resources has become an important issue for China is now facing. In order to achieve the rational use of water resources, in addition to in beefing water conservancy projects and enhance the people's awareness of water conservation efforts to improve. But more important is the application of new technical information, real-time to accurately understand and master a variety of hydrological information in order to make the right water scheduling and management, so that preventive measures to minimize water wastage . Coupled with long-standing water level measurement of water level has been an important issue in hydrology, water resources department. For the timely detection of the signs of the accident, precautionary measures in the future, economical and practical, reliable water level wireless monitoring systems will play a major role. The water level of dam safety, one of the important parameters for water drainage and irrigation scheduling, water storage, flood discharge.Provides a good foundation for the automation of monitoring, transmission and processing of the water level reservoir modernization. Need to monitor the water level in many areas of industrial and agricultural production. The site may not be able to close without the manpower to monitor, we can RMON, sitting in the control room facing the instrument can be monitored on-site, convenient and save manpower. In order to ensure the safe production of hydroelectric power station to improve power generation efficiency,Hydropower production process need to monitor the water level in the reservoir, trash rack, pressure drop and the tail water level. However, due to the different power plants with a different factual situations, have different technical requirements, and the measurement methods and location of the water level parameters and also the requirements of the monitoring equipment. This often results in the monitoring system equipment of a high degree of variety, interchangeability is not conducive to the maintenance of equipment will increase the equipment design, production, installation complexity. Therefore, on the basis of the actual situation and characteristics of the comprehensive study of hydropower water level monitoring, the use of modern electronic technology, especially single-chip technology and non-volatile memory technology, designed to develop a versatile, high reliability, easy maintenance, the applicable a variety of monitoring the environment, multi-mode automatic water level monitoring system has important practical significance. The subject according to the reservoir water level measurement needs, design a remote microcontroller water level monitoring system, the system automatically detects the water level, time processing, Data GPRS remote upload function. The design of the monitoring system will be significant savings in manpower and resources, low-power 24 hours of continuous monitoring and upload real-time control reservoir water level, to better adapt to the needs of the modern water level measurement, the safety of the dam reservoir, impoundment spillway to provide a basis.Microcontroller embedded microcontrollers are widely used in industrial measurement and control systems, intelligent instruments and household appliances. In real-time detection and automatic control of microcomputer application system, the microcontroller is often as a core component to use. The basic requirements of the water tower water level control system in the case of unattended automatic limit automatically start the motor to reach the water level in the water level in the water tower to the water tower water supply; water tower water level reached the water level upper limit is automatically off the motor to stop water supply. And unusual time to sound the alarm and troubleshooting in the water supply system at any time to ensure that the towers of the external normal water supply role. The water tower is often seen in daily life and industrial applications, water storage devices, external water supply through the control of its water level to meet the needs of its waterlevel control is universal. Regardless of socio-economic rapid water plays an important role in people's normal life and production. Once off the water, ranging from great inconvenience to the people's living standards, weight is likely to cause serious accidents and losses, and thus a higher demand of water supply system to meet the timely, accurate, safe and adequate water supply. If you still use the artificial way, the labor-intensive, low efficiency, safety is hard to guarantee the transformation of the automated control system, which must be carried out. In order to achieve sufficient amount of water, smooth water pressure, water towers, water level automatic control design low-cost, high practical value of the controller. The design uses a separate circuit to achieve high and low warning level processing, and automatic control, save energy, improve the quality of the water supply system.SCM is an integrated circuit chip, VLSI technology with data processing capability of the central processing unit CPU random access memory RAM, read only memory ROM, and a variety of I / O port and interrupt system, timers / timer other functions (which may also include a display driver circuit, pulse width modulation circuit, analog multi-channel converter, A / D converter and other circuit) integrated into a silicon constitute a small computer system. The basic features are as follows: the chip is small, but complete, SCM is one of the main features. Its internal program memory, data memory, a variety of interface circuit. Large processor speed is higher, the median more of the arithmetic unit, processing ability, but need to be configured in the external interface circuit; microcontroller clocked generally 100MHZ less suitable for small products for independent work, cited pin number from a few hundred. The application is simple, flexible, and free assembly language and C language development of SCM products. The working process of the microcontroller: microcontroller automatically complete the tasks entrusted to it, that is, single-chip implementation of the procedure for a section of the instruction execution process, the so-called directive requirements for single-chip implementation of the various operations used in the form of the command is to write down , which corresponds to a basic operation of designers assigned to it by the instruction set, an instruction; Full instructions can be executed by the microcontroller, the microcontroller instruction set, the different types of single-chip, and its instruction set is also different. So that the microcontroller canautomatically complete a specific task, the problem to be solved must be compiled into a series of instructions (these instructions must be selected microcontroller to the identification and implementation of the Directive), a collection of this series of instructions to become the program, the program need to pre- stored in the components - memory storage capabilities. Memory is composed by a number of storage units (the smallest unit of storage), like a large building has many rooms composed of the same, the instructions stored in these units, the instruction fetch unit and perform like the rooms of large buildings, each assigned to only a room number, each memory cell must be assigned to a unique address number, the address is known as the address of the storage unit, so as long as you know the address of the storage unit, you can find the storage unit that stores instructions can be removed, and then be executed. Programs are usually executed in the order, instruction program is a sequential storage, single-chip in the implementation of the program to be able to a section of these instructions out and be implemented, there must be a component to track the address of instruction where this part the program counter PC (included in the CPU), the start of program execution, endowed the address where the first instruction of the program to the PC, and then made for each command to be executed, the PC in the content will automatically increase, increase The amount is determined by the instruction length of this article may be 2 or 3, to point to the starting address of the next instruction to ensure the implementation of the instruction sequence.Microcontroller tower water level control system is the basic design requirements: inside the tower, we have designed a simple water level detection sensor used to detect the three water level, the low water level, the normal water level, water level. Low water to give a high single-chip, driven pumps and water, the red light; water level in the normal range, the pump add water, the green light; high water when the pump without water, the yellow light. The design process using the sensor technology, microcomputer technology, and light alarm technology and weak control the strong power of technology. Technical parameters and design tasks: 1, the use of the MCU to control the water level on the tower;, the water level in the water level detection sensor probe was the tower to give the microcontroller in order to achieve the water pump and water system and display system control; 3, the light alarm display system circuit, pumps and hydropower route relaycontrol;, analysis is drawn on the working principle of the system structure and a system block diagram using the microcontroller as a control chip, the main work process when the water in the tower low water level, water level detection sensor gave a high microcontroller, microcontroller-driven pump to add water and display system so that the red light lit; pump add water when the water level within the normal range, the green light, when the water level in the high-water mark, The microcontroller can not drive the water pump to add water, the yellow light. Light alarm circuit, the relay control circuit it works: When the water level in the low water, low water level detection sensor line is not +5 V power supply guide pass into the regulator circuit is treated in the output of the voltage regulator circuit has a high level, into the P1.0 port of the microcontroller, another high voltage circuit output of the microcontroller P1.1 port SCM After analysis, the P1.2 port outputs a low red light, drive, P1. 5 out a signal so that the optocoupler GDOUHE guide through so that the relay is closed, so that the water pump to add water; when the water level in the normal range, water pump plus P1.3 pin to a low level, so that the green light; when the water level in the high-water zone, the sensor of the two detection lines are conduction, are +5 power conduction into the SCM, SCM After analysis, the P1.4 pin out of a low yellow light, The optocoupler guide a low out of the P1.5-side can not pass, so that the relay can not be closed, the pump can not add water; failure when three flashing light indicates the system.译文水位控制电路设计中国水之源总量居世界第六位,人均占有水资源量仅为世界人均占有量的四分之一,并且在地域上分布很不平衡,长江以北的广大地区,特别是北方大、中城市大部分地区处于缺水状态,水资源短缺已成为制约我国经济发展的一个重要因素。
外文文献翻译译稿1可用性和期望值来自Willliam S.Green, Patrick W.Jordan.产品的愉悦:超越可用性根据人机工程学会(HFES)的观点,人机工程学着眼于“发现和共享可用于各种系统和设备设计的、关于人的特点的知识”。
人们通常只是把它作为生物力学和人体测量所关注的内容,实际上它是从更广泛的意义上的一种对人(产品用户)的全面和综合的理解。
HFES从二战中有军方从事的系统分析中发展而来。
其中的三种主要研究的是人体测量、复杂信息的解释和管理,以及在部队和装备调配中应用的系统分析。
系统分析在尺度和复杂性方面跨度很大,大的系统分析有类似于诺曼底登陆准备的大型系统规划,小到去理解如何从合理性和规模的角度才最佳的布置和装备人员。
诺曼底登陆是20世纪最复杂的事件之一。
他要求建立一个在战斗开始之前还不确定的庞大的人员和物资的合理分配系统。
在更小的规模上,装备和军事人物的布置意味着如何去组织、训练和安排战士,最大限度的发挥他们的长处。
士兵必须迅速地接受训练,并且能够有效地使用和维护在二战中发展起来的一系列技术装备。
其中,对于飞行员、潜艇人员和坦克驾驶员有神采的限制。
复杂的新装备的开发要求找到最好的税收、密码便医院、破译人员、雷达和声纳操作员、轰炸机驾驶员和机组人员。
在战后,随着公司及其产品在尺度、领域和复杂性方面的增长,很多系统分析人员在商用领域找到了发展机会。
尽管是战后的发展才导致了1957年人机工程协会(HFES)的建立,但人机研究的起源可以追溯到大批量生产方式的成型阶段,是当时提高生产效率的要求。
随着工作方式从手工生产和农业生产中的转移,新的工厂工作的概念逐步发展起来。
福特的流水生产线和泰勒的效率理论开始对生产的规划和教育产生影响。
即使在家庭生活中,妇女们也开始接受了现代家庭管理理论,并运用这些理论来组织和规划家庭。
在20世纪末,一种涵盖面更广的人机工程正在发展之中。
新的人机工程学是为了适应已经被广泛意识到的对用户行为模式更深入的需求而诞生的,它开始应用定型研究方法,并探索人的情感和认知因素。
英文文献全文翻译全文共四篇示例,供读者参考第一篇示例:LeGuin, Ursula K. (December 18, 2002). "Dancing at the Edge of the World: Thoughts on Words, Women, Places".《世界边缘的舞蹈:关于语言、女性和地方的思考》Introduction:In "Dancing at the Edge of the World," Ursula K. LeGuin explores the intersection of language, women, and places. She writes about the power of words, the role of women in society, and the importance of our connection to the places we inhabit. Through a series of essays, LeGuin invites readers to think critically about these topics and consider how they shape our understanding of the world.Chapter 1: LanguageConclusion:第二篇示例:IntroductionEnglish literature translation is an important field in the study of language and culture. The translation of English literature involves not only the linguistic translation of words or sentences but also the transfer of cultural meaning and emotional resonance. This article will discuss the challenges and techniques of translating English literature, as well as the importance of preserving the original author's voice and style in the translated text.Challenges in translating English literature第三篇示例:Title: The Importance of Translation of Full English TextsTranslation plays a crucial role in bringing different languages and cultures together. More specifically, translating full English texts into different languages allows for access to valuable information and insights that may otherwise be inaccessible to those who do not speak English. In this article, we will explore the importance of translating full English texts and the benefits it brings.第四篇示例:Abstract: This article discusses the importance of translating English literature and the challenges translators face when putting together a full-text translation. It highlights the skills and knowledge needed to accurately convey the meaning and tone of the original text while preserving its cultural and literary nuances. Through a detailed analysis of the translation process, this article emphasizes the crucial role translators play in bridging the gap between languages and making English literature accessible to a global audience.IntroductionEnglish literature is a rich and diverse field encompassing a wide range of genres, styles, and themes. From classic works by Shakespeare and Dickens to contemporary novels by authors like J.K. Rowling and Philip Pullman, English literature offers something for everyone. However, for non-English speakers, accessing and understanding these works can be a challenge. This is where translation comes in.Translation is the process of rendering a text from one language into another, while striving to preserve the original meaning, tone, and style of the original work. Translating afull-length English text requires a deep understanding of both languages, as well as a keen awareness of the cultural andhistorical context in which the work was written. Additionally, translators must possess strong writing skills in order to convey the beauty and complexity of the original text in a new language.Challenges of Full-text TranslationTranslating a full-length English text poses several challenges for translators. One of the most significant challenges is capturing the nuances and subtleties of the original work. English literature is known for its rich and layered language, with intricate wordplay, metaphors, and symbolism that can be difficult to convey in another language. Translators must carefully consider each word and phrase in order to accurately convey the author's intended meaning.Another challenge of full-text translation is maintaining the author's unique voice and style. Each writer has a distinct way of expressing themselves, and a good translator must be able to replicate this voice in the translated text. This requires a deep understanding of the author's writing style, as well as the ability to adapt it to the conventions of the target language.Additionally, translators must be mindful of the cultural and historical context of the original work. English literature is deeply rooted in the history and traditions of the English-speaking world, and translators must be aware of these influences in orderto accurately convey the author's intended message. This requires thorough research and a nuanced understanding of the social, political, and economic factors that shaped the work.Skills and Knowledge RequiredTo successfully translate a full-length English text, translators must possess a wide range of skills and knowledge. First and foremost, translators must be fluent in both the source language (English) and the target language. This includes a strong grasp of grammar, syntax, and vocabulary in both languages, as well as an understanding of the cultural and historical context of the works being translated.Translators must also have a keen eye for detail and a meticulous approach to their work. Every word, sentence, and paragraph must be carefully considered and translated with precision in order to accurately convey the meaning of the original text. This requires strong analytical skills and a deep understanding of the nuances and complexities of language.Furthermore, translators must possess strong writing skills in order to craft a compelling and engaging translation. Translating a full-length English text is not simply a matter of substituting one word for another; it requires creativity, imagination, and a deep appreciation for the beauty of language. Translators mustbe able to capture the rhythm, cadence, and tone of the original work in their translation, while also adapting it to the conventions of the target language.ConclusionIn conclusion, translating a full-length English text is a complex and challenging task that requires a high level of skill, knowledge, and creativity. Translators must possess a deep understanding of both the source and target languages, as well as the cultural and historical context of the work being translated. Through their careful and meticulous work, translators play a crucial role in making English literature accessible to a global audience, bridging the gap between languages and cultures. By preserving the beauty and complexity of the original text in their translations, translators enrich our understanding of literature and bring the works of English authors to readers around the world.。
(空一行)原□□稿(空一行) IntroductionThe "jumping off" point for this paper is Reengineering the Corporation, by Michael Hammer and James Champy . The paper goes on to review the literature on BPR. It explores the principles and assumptions behind reengineering, looks for commonfactors behind its successes or failures, examines case studies, and presents alternatives to "classical" reengineering theory . The paper pays particular attention to the role of information technology in BPR. In conclusion, the paper offers somespecific recommendations regarding reengineering. Old Wine in New BottlesThe concept of reengineering traces its origins back to management theories developedas early as the nineteenth century . The purpose of reengineering is to "make all your processes the best-in-class." Frederick Taylor suggested in the 1880's that managers use process reengineering methods to discover the best processes for performing work, and that these processes be reengineered to optimize productivity. BPR echoes the classical belief that there is one best way to conduct tasks. In Taylor's time, technology did not allow large companies to design processes in across-functional or cross-departmental manner. Specialization was the state-of-theart method to improve efficiency given the technology of the time.(下略)正文内容:新罗马“TimesNewRoman ”字体,小四号字。
中英文对照外文翻译(文档含英文原文和中文翻译)Bridge research in EuropeA brief outline is given of the development of the European Union, together withthe research platform in Europe. The special case of post-tensioned bridges in the UK is discussed. In order to illustrate the type of European research being undertaken, an example is given from the University of Edinburgh portfolio: relating to the identification of voids in post-tensioned concrete bridges using digital impulse radar.IntroductionThe challenge in any research arena is to harness the findings of different research groups to identify a coherent mass of data, which enables research and practice to be better focused. A particular challenge exists with respect to Europe where language barriers are inevitably very significant. The European Community was formed in the 1960s based upon a political will within continental Europe to avoid the European civil wars, which developed into World War 2 from 1939 to 1945. The strong political motivation formed the original community of which Britain was not a member. Many of the continental countries saw Britain’s interest as being purelyeconomic. The 1970s saw Britain joining what was then the European Economic Community (EEC) and the 1990s has seen the widening of the community to a European Union, EU, with certain political goals together with the objective of a common European currency.Notwithstanding these financial and political developments, civil engineering and bridge engineering in particular have found great difficulty in forming any kind of common thread. Indeed the educational systems for University training are quite different between Britain and the European continental countries. The formation of the EU funding schemes —e.g. Socrates, Brite Euram and other programs have helped significantly. The Socrates scheme is based upon the exchange of students between Universities in different member states. The Brite Euram scheme has involved technical research grants given to consortia of academics and industrial partners within a number of the states—— a Brite Euram bid would normally be led by partners within a number of the statesan industrialist.In terms of dissemination of knowledge, two quite different strands appear to have emerged. The UK and the USA have concentrated primarily upon disseminating basic research in refereed journal publications: ASCE, ICE and other journals. Whereas the continental Europeans have frequently disseminated basic research at conferences where the circulation of the proceedings is restricted.Additionally, language barriers have proved to be very difficult to break down. In countries where English is a strong second language there has been enthusiastic participation in international conferences based within continental Europe —e.g. Germany, Italy, Belgium, The Netherlands and Switzerland. However, countries where English is not a strong second language have been hesitant participants }—e.g. France.European researchExamples of research relating to bridges in Europe can be divided into three types of structure:Masonry arch bridgesBritain has the largest stock of masonry arch bridges. In certain regions of the UK up to 60% of the road bridges are historic stone masonry arch bridges originally constructed for horse drawn traffic. This is less common in other parts of Europe as many of these bridges were destroyed during World War 2.Concrete bridgesA large stock of concrete bridges was constructed during the 1950s, 1960s and 1970s. At the time, these structures were seen as maintenance free. Europe also has a large number of post-tensioned concrete bridges with steel tendon ducts preventing radar inspection. This is a particular problem in France and the UK.Steel bridgesSteel bridges went out of fashion in the UK due to their need for maintenance as perceived in the 1960s and 1970s. However, they have been used for long span and rail bridges, and they are now returning to fashion for motorway widening schemes in the UK.Research activity in EuropeIt gives an indication certain areas of expertise and work being undertaken in Europe, but is by no means exhaustive.In order to illustrate the type of European research being undertaken, an example is given from the University of Edinburgh portfolio. The example relates to the identification of voids in post-tensioned concrete bridges, using digital impulse radar.Post-tensioned concrete rail bridge analysisOve Arup and Partners carried out an inspection and assessment of the superstructure of a 160 m long post-tensioned, segmental railway bridge in Manchester to determine its load-carrying capacity prior to a transfer of ownership, for use in the Metrolink light rail system..Particular attention was paid to the integrity of its post-tensioned steel elements.Physical inspection, non-destructive radar testing and other exploratory methods were used to investigate for possible weaknesses in the bridge.Since the sudden collapse of Ynys-y-Gwas Bridge in Wales, UK in 1985, there has been concern about the long-term integrity of segmental, post-tensioned concrete bridges which may b e prone to ‘brittle’ failure without warning. The corrosion protection of the post-tensioned steel cables, where they pass through joints between the segments, has been identified as a major factor affecting the long-term durability and consequent strength of this type of bridge. The identification of voids in grouted tendon ducts at vulnerable positions is recognized as an important step in the detection of such corrosion.Description of bridgeGeneral arrangementBesses o’ th’ Barn Bridge is a 160 m long, three span, segmental, post-tensionedconcrete railway bridge built in 1969. The main span of 90 m crosses over both the M62 motorway and A665 Bury to Prestwick Road. Minimum headroom is 5.18 m from the A665 and the M62 is cleared by approx 12.5 m.The superstructure consists of a central hollow trapezoidal concrete box section 6.7 m high and 4 m wide. The majority of the south and central spans are constructed using 1.27 m long pre-cast concrete trapezoidal box units, post-tensioned together. This box section supports the in site concrete transverse cantilever slabs at bottom flange level, which carry the rail tracks and ballast.The center and south span sections are of post-tensioned construction. These post-tensioned sections have five types of pre-stressing:1. Longitudinal tendons in grouted ducts within the top and bottom flanges.2. Longitudinal internal draped tendons located alongside the webs. These are deflected at internal diaphragm positions and are encased in in site concrete.3. Longitudinal macalloy bars in the transverse cantilever slabs in the central span .4. Vertical macalloy bars in the 229 mm wide webs to enhance shear capacity.5. Transverse macalloy bars through the bottom flange to support the transverse cantilever slabs.Segmental constructionThe pre-cast segmental system of construction used for the south and center span sections was an alternative method proposed by the contractor. Current thinkingire suggests that such a form of construction can lead to ‘brittle’ failure of the ententire structure without warning due to corrosion of tendons across a construction joint,The original design concept had been for in site concrete construction.Inspection and assessmentInspectionInspection work was undertaken in a number of phases and was linked with the testing required for the structure. The initial inspections recorded a number of visible problems including:Defective waterproofing on the exposed surface of the top flange.Water trapped in the internal space of the hollow box with depths up to 300 mm.Various drainage problems at joints and abutments.Longitudinal cracking of the exposed soffit of the central span.Longitudinal cracking on sides of the top flange of the pre-stressed sections.Widespread sapling on some in site concrete surfaces with exposed rusting reinforcement.AssessmentThe subject of an earlier paper, the objectives of the assessment were:Estimate the present load-carrying capacity.Identify any structural deficiencies in the original design.Determine reasons for existing problems identified by the inspection.Conclusion to the inspection and assessmentFollowing the inspection and the analytical assessment one major element of doubt still existed. This concerned the condition of the embedded pre-stressing wires, strands, cables or bars. For the purpose of structural analysis these elements、had been assumed to be sound. However, due to the very high forces involved,、a risk to the structure, caused by corrosion to these primary elements, was identified.The initial recommendations which completed the first phase of the assessment were:1. Carry out detailed material testing to determine the condition of hidden structural elements, in particularthe grouted post-tensioned steel cables.2. Conduct concrete durability tests.3. Undertake repairs to defective waterproofing and surface defects in concrete.Testing proceduresNon-destructi v e radar testingDuring the first phase investigation at a joint between pre-cast deck segments the observation of a void in a post-tensioned cable duct gave rise to serious concern about corrosion and the integrity of the pre-stress. However, the extent of this problem was extremely difficult to determine. The bridge contains 93 joints with an average of 24 cables passing through each joint, i.e. there were approx. 2200 positions where investigations could be carried out. A typical section through such a joint is that the 24 draped tendons within the spine did not give rise to concern because these were protected by in site concrete poured without joints after the cables had been stressed.As it was clearly impractical to consider physically exposing all tendon/joint intersections, radar was used to investigate a large numbers of tendons and hence locate duct voids within a modest timescale. It was fortunate that the corrugated steel ducts around the tendons were discontinuous through the joints which allowed theradar to detect the tendons and voids. The problem, however, was still highly complex due to the high density of other steel elements which could interfere with the radar signals and the fact that the area of interest was at most 102 mm wide and embedded between 150 mm and 800 mm deep in thick concrete slabs.Trial radar investigations.Three companies were invited to visit the bridge and conduct a trial investigation. One company decided not to proceed. The remaining two were given 2 weeks to mobilize, test and report. Their results were then compared with physical explorations.To make the comparisons, observation holes were drilled vertically downwards into the ducts at a selection of 10 locations which included several where voids were predicted and several where the ducts were predicted to be fully grouted. A 25-mm diameter hole was required in order to facilitate use of the chosen horoscope. The results from the University of Edinburgh yielded an accuracy of around 60%.Main radar sur v ey, horoscope verification of v oids.Having completed a radar survey of the total structure, a baroscopic was then used to investigate all predicted voids and in more than 60% of cases this gave a clear confirmation of the radar findings. In several other cases some evidence of honeycombing in the in site stitch concrete above the duct was found.When viewing voids through the baroscopic, however, it proved impossible to determine their actual size or how far they extended along the tendon ducts although they only appeared to occupy less than the top 25% of the duct diameter. Most of these voids, in fact, were smaller than the diameter of the flexible baroscopic being used (approximately 9 mm) and were seen between the horizontal top surface of the grout and the curved upper limit of the duct. In a very few cases the tops of the pre-stressing strands were visible above the grout but no sign of any trapped water was seen. It was not possible, using the baroscopic, to see whether those cables were corroded.Digital radar testingThe test method involved exciting the joints using radio frequency radar antenna: 1 GHz, 900 MHz and 500 MHz. The highest frequency gives the highest resolution but has shallow depth penetration in the concrete. The lowest frequency gives the greatest depth penetration but yields lower resolution.The data collected on the radar sweeps were recorded on a GSSI SIR System 10.This system involves radar pulsing and recording. The data from the antenna is transformed from an analogue signal to a digital signal using a 16-bit analogue digital converter giving a very high resolution for subsequent data processing. The data is displayed on site on a high-resolution color monitor. Following visual inspection it isthen stored digitally on a 2.3-gigabyte tape for subsequent analysis and signal processing. The tape first of all records a ‘header’ noting the digital radar settings together with the trace number prior to recording the actual data. When the data is played back, one is able to clearly identify all the relevant settings —making for accurate and reliable data reproduction.At particular locations along the traces, the trace was marked using a marker switch on the recording unit or the antenna.All the digital records were subsequently downloaded at the University’s NDT laboratory on to a micro-computer.(The raw data prior to processing consumed 35 megabytes of digital data.) Post-processing was undertaken using sophisticated signal processing software. Techniques available for the analysis include changing the color transform and changing the scales from linear to a skewed distribution in order to highlight、突出certain features. Also, the color transforms could be changed to highlight phase changes. In addition to these color transform facilities, sophisticated horizontal and vertical filtering procedures are available. Using a large screen monitor it is possible to display in split screens the raw data and the transformed processed data. Thus one is able to get an accurate indication of the processing which has taken place. The computer screen displays the time domain calibrations of the reflected signals on the vertical axis.A further facility of the software was the ability to display the individual radar pulses as time domain wiggle plots. This was a particularly valuable feature when looking at individual records in the vicinity of the tendons.Interpretation of findingsA full analysis of findings is given elsewhere, Essentially the digitized radar plots were transformed to color line scans and where double phase shifts were identified in the joints, then voiding was diagnosed.Conclusions1. An outline of the bridge research platform in Europe is given.2. The use of impulse radar has contributed considerably to the level of confidence in the assessment of the Besses o’ th’ Barn Rail Bridge.3. The radar investigations revealed extensive voiding within the post-tensioned cable ducts. However, no sign of corrosion on the stressing wires had been foundexcept for the very first investigation.欧洲桥梁研究欧洲联盟共同的研究平台诞生于欧洲联盟。
2014届毕业设计(论文)外文文献原稿和译文姓名:陈立贺专业:升会计班级: 12.2 学号: 20122804020203 指导教师:刘丽颖外文文献原稿和译文原稿COST MANAGEMENT AND PRODUCTIONCONTROL FOR CONSTRUCTION COMPANIESMarchesan, P.R.C.; Formoso, C. T.ABSTRACTConstruction, like other industries, has been experiencing profound changes involving both the business environment and internal organization. In this context, new cost management information that provides better understanding and helps managing increasingly turbulent and complex production processes is needed. Activity-based costing (ABC) has been suggested as the leading contender method to replace traditional cost accounting systems, due to its capability to make the processes and activities performed in the organization more transparent and observable.This paper reports the main results of a research project which aimed to develop a cost accounting system capable of providing useful information to manage production processes in construction and devise a model that integrates this system to the Production Planning and Control process, based on the new operations management paradigm and on the ABC ideas. The study involved three case studies, a series of interviews with construction managers and the development of prototype software.The main conclusions of this research work are that the cost information provided by the proposed management accounting system (a) makes the production processes more transparent; (b) helps to identify production inefficiencies; (c) encourages managers to introduce corrective actions; and (d) allows the evaluation of corrective actions to be undertaken. The study also indicated that the proposed model has contributed to establishsystematic procedures for production control concerned not only with time management but also with cost management.KEY WORDS:Cost management, activity-based costing, lean construction, production control.1. INTRODUCTIONOver the last decades, many industrial sectors have been experiencing profound changes involving both the business environment and the internal organisation. This process has been so deep and radical as to suggest that a new operations management paradigm has emerged (Bartezzaghi, 1999; Koskela, 2000). In this new competitive and turbulent environment, effective cost management information has become extremely important to drive improvement efforts (Johnson & Kaplan, 1987).However, besides the environmental, managerial and technological changes occurred in the last thirty years, the existing traditional cost management systems are very similar to the ones that have been used since the mid Twenties (Johnson & Kaplan, 1987). In the face of all these changes, traditional cost account information has become mostly irrelevant and even dangerous for managerial purposes (Ploss, 1990). According to Johnson & Kaplan (1987), traditional management accounting information tends to be too late, too aggregated and too distorted to be relevant for production planning and control.The failings of the traditional management accounting systems have three important consequences. Firstly, these systems cannot provide accurate product cost. Costs are distributed to products in a simplistic and arbitrary way that usually does not represent the real demand imposed by each product on the company's resources (Johnson & Kaplan, 1987). Secondly, traditional management accounting systems fail to stimulate decisions that can affect the overall production result. Managers are sometimes encouraged to accomplish short-term goals by reducing expenses with training and investment, or even produce to stock. Although effective in short term, these decisions can seriously affect future results (Goldratt & Cox, 1992). Finally, the cost management information provided by the traditional systems is of little help to managers in their effort to improve production performance. The information provided is past-oriented and too aggregated to be useful inplanning and control decisions, because these systems are developed mostly to satisfy fiscal and financial needs. The lack of transparency allied with the lack of timeliness prevents the traditional cost information to help in the identification and correction of production flow inefficiencies.Particularly in the construction industry, the inadequacy of cost accounting systems has resulted in the dissociation between the cost management and the Production Planning and Control processes. In general, construction cost control consists basically of monitoring actual performance against cost estimates and identifying variances. According to Ballard (2000), the traditional control methods based on the detection of variances appears to assume that the causes of deviation will be apparent and the appropriate corrective action obvious. As a result, the traditional cost control systems has been much more useful to manage contracts than production (Howell & Ballard, 1996; Koskela, 2000).Since the Eighties, some alternatives for traditional cost accounting systems have been developed, aiming to regain the managerial relevance of cost information. One of these alternatives is the Activity-Based Costing (ABC) method, which has been suggested as the leading contender to replace traditional cost accounting methods due to its capability to make the processes and activities performed in an organisation transparent and observable. Over the past decades, ABC has helped many manufacturing and service organisations to improve their competitiveness by enabling them to make better decisions based on cost information.2. BACKGROUND FOR THE RESEARCHConstruction cost control systems has been the subject of a myriad of studies. Despite their relevance in terms of both improving the cost estimate structure and integrating cost and schedule, they have hardly contributed to the integration of cost management and production control systems. Besides the fact that construction cost control systems have not changed much since the Seventies, cost management and productioncontrol are still treated independently, as separated systems.From a managerial point-of-view, the effort to develop, implement and operate a cost system is justifiable only when the cost information provides effective support for decision making (Johnson & Kaplan, 1987; Krieger, 1997).Activity-Based Costing has been increasingly adopted in many industrial and service firms as a method to improve cost management in complex production systems. Despite some similarities to the cost centre method or other traditional costing methods (Horngren & Foster, 1990), the underlying philosophy of ABC is considerably different. ABC is a costing system that is based on the idea that activities consume resources and product /services consume activities.The method is basically a two-stage approach for allocating indirect costs to products based on cost drivers of various levels. In the first stage, resource costs (labour, equipment and power) are assigned to those activities performed in the organisation. During the second stage, activities costs are assigned to the cost objects based on selected cost drivers (e.g., machine set-up, quality inspection and material handling activities), which express a causal relation between the activity demand and the cost object considered. Besides the fact that ABC permits to directly trace manufacturing costs to products, it is possible to determine the costs related to objects different from products, e.g., product family, services and clients.However, the main contribution of ABC to the operations management area is the process view incorporated by the method. The information produced by ABC cost systems can increase process transparency, providing guidance to identify non-valueadding activities and take the necessary corrective actions (Kaplan & Cooper, 1995).Notwithstanding the benefits of its application, ABC presents some drawbacks when compared to traditional cost systems. Perhaps, the most important one is the larger amount of data usually needed in ABC systems. Indeed, according to some authors (Krieger, 1997; Cokins, 1999), the excessive level of detail is a major cause of unsuccessful ABC implementations. This problem can be even worse when one considers unstable and complex production processes, such as those observed in the construction industry.Despite its wide utilisation in manufacturing companies and its capability to increase the transparency of production processes, ABC has been poorly discussed in the construction literature. The few existing studies usually restrict the discussion of ABC concepts to academic applications (Maxwell et al., 1998). Therefore, there is a clear need to discuss the use of ABC concepts and principles in the construction environment. As othermanufacturing concepts and practices transferred to construction, ABC has to be translated and adapted (Lillrank, 1995).3. AN APPLICATION OF ACTIVITY-BASED CONCEPT IN THE CONSTRUCTION INDUSTRYThis paper suggests an integrated model of cost management and production control for construction companies, which is strongly based on concepts related to production planning and control proposed by Laufer & Tucker (1987) and Ballard & Howell (1998). In this model, four subprocesses are identified: the preparation, the production planning and control process itself, the evaluation process and, finally, the cost management system.The information generated by the cost system is introduced into two different control cycles (figure 1) - one is continuous and the other is intermittent. The first is a continuous feedback for long and medium term planning. The aim is to encourage a process of continuous improvement to happen, enabling the managers to identify and correct problems in real time. In the intermittent cycle, cost information is used in the planning evaluation process, aiming to improve the production planning and control process and support decisions to be taken in future projects.Figure 1: An Integrated model for cost management and production control inconstruction companies.ABC was chosen as the conceptual basis of the proposed system for its capability to provide a better understanding about the processes and the activities performed, as well as for its affinity with the new operations management ideas.There are two basic groups of resources that are involved in production: those that perform the work - named production resources -and those that are the object of work typically materials. Production resources are those that can add value to the products. Considering that the cost management of materials is an issue much discussed in literature, this paper focuses on the application of ABC to the management of production resources.Construction production processes tend to be extremely variable. This variability involves not only that caused by manual labour, but also the variability of the process configuration itself. In such a context, the use of a classical ABC structure would demand too many activities, overburdening the cost system.For that reason, a different cost allocation structure was proposed. As shown in figure 2, activities were decomposed into tasks and operations. Resources are consumed by operations and the operations are demanded by different tasks. Differently from typical ABC structures, a restricted set of operations is employed to describe and analyse any construction production process. However, in practice, some resources, such as employees or subcontractors, are typically assigned to tasks. For such resources, cost assignment is made upstream, i.e., from tasks to operations.Figure 2: Cost Allocation Model - adapted from ABCOne of the main challenges in the application of the proposed cost allocation model is the definition of an adequate set of operations. Firstly, the conceptual difference between process and operation must be observed (Shingo, 1988). Briefly, process refers to the material flow and operations refers to the human and the equipment flow. The cost system presented in this work is focused on the operations axis, i.e. the subjects that perform the work.Secondly, the existence of different process types should be recognised (Shingo, 1988). Despite the fact that any process - essential, contributory or preparation, and post-adjustment (figure 3) -consists of the same categories of activities -waiting, inspection, moving and processing -the nature and the purpose of an operation differ considerably in each of those process types.Figure 3: Classification of processes (adapted from Shingo (1988)).In this study, essential processes were defined as those that add a permanent component or prefabricate a permanent component to the owner-specified facility (Pregenzer et al., 1999). Clearly, the definition of what is essential or contributory could be different depending on who classifies the process: owner, contractor, subcontractor or layman. Preparation, post-adjustment and contributory tasks are necessary for the execution of essential activities, but do not directly add a permanent component to the facility, i.e. do not directly add value for the owner. For example, using current methods, formwork must be built for a concrete column to be cast, but the construction of the formwork itself doesnot add a permanent component to facility and, therefore, does not add value for the owner.The following set of operations was defined for the proposed cost system:·Processing: a conversion operation that intersects an essential process. It is a conversion activity that adds a permanent component or prefabricates a permanent component of the owner specified facility, e.g. pouring concrete on slabs, bricklaying, electrical wire installation. Only processing may add value to the final product from the owner point of view;· Pre-processing: a conversion operation that intersects a contributory process, which is a prerequisite of an essential process. Pre-processing operations are considered to be non-value adding, because it does not add a permanent component to the final product, e.g., formwork construction, and scaffolding assembling;·Post-processing: also a conversion process that intersects a contributory process, which, in this case, is subsequent to an essential process, e.g. formwork and scaffolding disassembling;·Moving: an operation in which any material is moved. This operation may belong to either essential or contributory processes;·Cleaning: an operation dedicated to remove rubbish and waste materials from workstations or from construction sites. It belongs to preparation and post-adjustment processes;·Adjustments and corrections: a conversion operation related to both the correction of defects (rework) and the adjustments required by the different production methods;·Travelling: an operation in which a workman travels from one point to another. This operation does not belong to any process, i.e., it does not intersect any material flow;·Waiting: it corresponds to those idle times caused by inclement weather, production disruptions or physiological needs. Like travelling, it does not belong to any process.This set of eight operations has been specifically proposed for this study, considering its objectives and the resources available. In other studies, a different set could be suggested, depending on the cost-benefit analysis of data collection and processing.3.1 DATA GATHERING AND ANALYSISDespite the fact that it is possible to collect data on site, the perception of managersand employees was used in this study to evaluate the intensity which each operation is demanded by each different task. This decision was made due to the difficulty of incorporating objective methods, such as work sampling and work study, in small and medium sized construction companies as systematic procedures for measuring the time spent performing each operation. Moreover, the relative short duration of several construction tasks would imply in the need of making a great number of observations during a very short period of time – sometimes about half a day – in order to get statistical validity.According to Drucker (1995), it is necessary to rely on assessment and judgement rather than on direct measurement when objective measures are nearly impossible to obtain. Furthermore, perception has been successfully used with such purpose in several cases of ABC implementation (Brimson, 1991). Although the use of perception is a limitation of this study, objective measures might be used when the processes are reasonably stable and the proper resources are available. In such situations, objective time measurements can also be used to validate data collected through the perception-based measurement.In general terms, the procedure adopted in this study consisted of measuring the total amount of resources spent in each task (total man-hour) and then determining the intensity that each task demanded from each operation. The time demanded by each operation was evaluated by the managers and the foremen through the assignment of an index (0 to 10) which express the relative intensity of each operation.To illustrate the data gathering and analysis process, consider the example provided in table 1. The table represents a daily register of the tasks performed in a construction site. Each line in the table corresponds to a specific task, defined as a unique combination of an action (e.g. casting, assembling or digging) to be performed on a physical element (e.g. beams, columns, windows or pipes), located in a zone (e.g. a building, a level, an apartment, or a room). For each task, the total cost is given by the multiplication of labour cost per hour and the task duration. For example, assuming that each worker (JJ) cost $40,00 per hour, the task of "pouring concrete on the 14th floor's slab" costed $160,00.Table 1: Cost System Input Data – ExampleThe second step consists of distributing the task cost among the set of operations considered. For every tasks an intensity index is assigned to each operation based on the perception of managers and employees. Accordingly to the assigned index, a share of the task cost is allocated to each operationFinally, the cost assignment to cost objects depends on how these are defined. A cost object might be defined as an element, as a zone, or as a combination of both element and zone. Once defined the cost object, the costs of all those tasks that match the cost object definition are summed up to determine its total cost.CONCLUSIONSThis paper reports the main results of a research project which investigated the application of ABC concepts to improve process transparency in construction cost control.A cost allocation model was suggested aiming to translate and adapt the method for the construction industry context. The need to integrate the cost management information in the production planning and control process was also stressed. Three case studies were carried out, in which the proposed cost system was successfully implemented.The main contributions of this paper are presented below:·The proposed cost system has demonstrated the capability of providing relevant information for production management. Its implementation has increased production processes transparency, making visible important aspects usually neglected by traditional cost systems. The cost information encouraged discussion and reflection about the way production process are managed and performed. In this way, it helped to identify production inefficiencies, as well as to evaluate the corrective actions that were undertaken. The great benefit of the proposed cost system, when compared with operational analysis techniques such as first run studies, work sampling and craftworker questionnaires is that it enables toquantify and evaluate processes costs distribution in a systematic way. The proposed system can be continuously operated by the construction companies, allowing them not only to identify but also to quantify waste, so that corrective actions can be properly evaluated before hand in terms of cost.;· the adoption of ABC concepts in the construction industry demands the translation and adaptation of such concepts in order to create a manageable cost system. The cost allocation model suggested in this paper uses "operations" as the main unit for process analysis. Instead of hundreds of activities, a restricted set of operations, which incorporates the flow and value views by differentiating the value adding from the non-value adding operations, is employed to describe and analyse any production process;· the integration of cost information in the production planning and control process is essential so that systematic procedures for production control concerned not only with time but also with costs can be established.REFERENCESBrimson, J. A. (1991) "Activity Accounting: an activity-based costing approach" John Wiley & Sons, Inc, New York.Ballard, H. G. (2000) "The last planner system of production control". Thesis submitted for the degree of Doctor of Philosophy, Faculty of Engineering, The University of Birmingham. Birmingham, 137pp.Ballard, H. G; Howell, G. (1998). "Shielding production: an essential step in production control". Journal of Construction Engineering and Management, 124(1): pp. 11-17.Bartezzaghi, E. (1999) "The evolution of production models: is a new paradigm emerging?". Intern. J. of Oper. & Prod. Mgmt, 19(2), 229-250.Drucker, P.F. (1995) "The information executives truly need". Harvard Business Review, Boulder, 73(1), 54-62.Goldratt, E.M. and Cox, J.F. (1992). "The Goal: a process of ongoing improvement". 2nd ed., North River, Great Barrington, 337p.Horngren, C.T.; Foster, G. (1990) ―Cost accounting: a managerial emphasis‖. 7th ed. Englewood: Prentice-Hall. 964 p.Howell, G. and Ballard, G. (1996) ―Can project controls do its job?‖. Proceedings of the 4th annual conference of the International Group for Lean Construction, Birminghan, England.Johnson, H.T. and Kaplan, R.S. (1987) "Relevance lost: the rise and fall of management accounting". Harvard Business School, Boston, 269pp.Kaplan, R.S. and Cooper, R. (1995) "Cost and effect" Harvard Business School Press, Boston. 376pp.Koskela, L. (2000) "An exploration towards a production theory and its application to construction". Dr. Thesis, Technical Research Centre of Finland, Helsinki, 296pp.Krieger, J. (1997) "Establishing activity-based costing: Lessons & Pitfalls". Newspaper Financial Executives Quarterly, 3(4), 14-17.Laufer, A. and Tucker, R.L. (1987) "Is construction project planning really doing its job? A critical examination of focus , role and process". Constr. Mgmt, and Econ. London, 5(3), 243-266.Lillrank, P. (1995) "The transfer of management innovations from Japan". Organisation Studies, 16(6), 971-989.Maxwell, D.A.; Back, W.E. and Toon, J. (1998) "Optimazation of crew configurations using activity: based costing". J. Constr. Engrg. and Mgmt., ASCE, New York, NY, 124(2), 162-168.Ploss, G. W. (1990) "Cost Manufacturing: dawn of a new era". Production Planning & Control, London, 1(1), 61-68.Pregenzer, L.J.; Seppanen, P.J.; Kunz, J.C. and Paulson, B.C (1999) "Value-added assessment of construction plans". J. Constr. Engrg. and Mgmt, ASCE, New York, NY, 125(4), 242-248.Shingo, S. (1988). "Non-Stock Production: The Shingo System for Continuous Production" Productivity Press, Cambridge, Massachusetts.译文建筑公司的成本管理和生产控制Marchesan, P.R.C.; Formoso, C. T.摘要建筑和其他的行业一样,在商业环境和内部的组织结构方面已经经历了意义深远的变革。
外文文献原稿和译文原稿INTRODUCTIONNowadays, advanced control systems are playing a major role in plant operations because they allow for effective plant management. The prime advantage of a full plant automation stands in the improvement of plant profitability and productivity that lead to short term payoffs of the investment required for the implementation of advanced automation systems.Typically, advanced control systems rely heavily on real-time process modeling, and this puts strong demands on developing effective process models that, as a prime requirement, have to exhibit real-time responses. Because in many instances detailed process modeling is not viable, efforts have been devoted towards the development of approximate dynamic models. The models suitable for real-time applications result from a trade-off between model complexity and degree of representability of the actual process dynamics.Approximate process models can be classified in structured and unstructured models. While the former are based on first principles, and thus require good understanding of the process physics, the latter are based on some sort of black-box modeling and, in principle, would not require any a priori knowledge of the process. Neural network modeling represents an effective framework to develop unstructured models when relying on an incomplete knowledge of the process under examination. Because of the simplicity of neural models, they exhibit great potentials in all those model-based control applications that require real-time solutions of dynamic process models. The better understanding acquired on neural network modeling has driven its exploitation in many chemical engineering applications: data reconciliation and rectification, process identification and optimization, software sensor development (inferential measurements), state estimation, fault analysis, multivariable control (for a list of references cf. the recent review by Stephanopoulos and Han). The commonbelief that effective and efficient neural can be developed without any a priori understanding of the investigated process is not fully correct. This misconception may lead to neural applications exhibiting poor performance, and this may have been the reason for preventing the full exploitation of neural modeling potentials in real-world applications. Indeed, the selection of the proper inputs and outputs to be fed to the network represents a critical step when developing a neural network model, and an adequate understanding of the process under examination is demanding.Recently Baratti et al. have demonstrated the potentials of neural modeling to develop effective inferential control strategies of industrial multicomponent distillation columns. Figure 1 schematically shows a tray distillation column with the typical controlled (temperatures on specified trays, the so called pilot temperatures) and manipulated (reflux, L0, and boilup, V, flowrates) variables. Figure 1 refers to a case where the pilot temperatures are the ones of the 16th and 3rd trays.The prime aim of this work is to investigate and fully exploit neural modeling in advanced model-based control applications of distillation columns. The basic idea is to construct a neural tool-box that, on the basis of detected disturbances, is capable ofacting properly on the manipulated variable so as to compensate for the upsetting loads as soon as they are detected. This action relies on a process modeling that must provide the description of the effect of disturbances and manipulated variables on the process performance. The direct identification of the action on the manipulated variable so as to meet the specifications requires a process inverse model. In this work, we discuss the use of neural modeling to develop an inverse model of a multicomponent distillation column. The neural model provides the controller action law. The direct action on the manipulated variable allows the control engineer to overcome the problems associated with the optimal tuning of the parameters of conventional feedback controllers. Furthermore, the action of the neural controller resembles that of a feedforward strategy, and thus it may be expected to exhibit a good control performance also in the case of a crude description of the process inverse model.For the case of a single composition control problem of a distillation column, neural network modeling is used to develop a model-based neural control strategy to compensate for upsets in the operating pressure, feed flowrate, and feed composition so as to hold constant the content of the key component in the distillate stream. The neural controller is developed for a five-component distillation column (propane, butane, n-pentane, i-pentane and hexane) that is representative of actual stabilizer units, so commonly encountered in the refinery industries. The performance of the neural controller is compared with that of a conventional temperature control loop and of a neural inferential controller.CASE STUDYThe neural controller is developed for a multicomponent distillation column for which the set of calibration data are produced making use of a dynamic simulator. A 18-trays (ideal stages), five component (propane, hexane, butane, n- and i-pentane) distillation unit is selected as prototype of a gasoline stabilizer tower. The column is also equipped with a reboiler and a total condenser. The feed (saturated liquid) is fed at the 11th stage from the bottom and is constituted of a mixture of propane (l0%), butane (5%), n-pentane (15%), i-pentane (30%) and hexane (40%) (the composition is given on a mole basis). The control objective is to maintain as low as possible thecontent of i-pentane (i-C5)is the distillate and that of propane (C3) is the residue when the main upsets are due to variations or fluctuations in the feed composition, operating pressure and feed flowrate. The variation in the feed composition is the result of mixing the primary stream with a secondary one (slop) whose composition (on mole basis) is: C3, 2%; C4, 2%; C6, 45%; n-C5, 20%; i-C531%.Typically, the secondary stream accounts for about 5-15% of the total feed flowrate entering the distillation unit.The dynamics of distillation columns can be simulated by various model approaches, involving different levels of complexity. Here, we are mainly concerned with addressing the issue of process inversion via neural modeling, thus we have adopted a simple description of the distillation unit that relies on the following assumptions: each tray is an adiabatic ideal equilibrium stage, all the involved species have the same latent heat of vaporization, the vapor and liquid are ideal mixtures, the vapor and liquid holdups are negligible. From these assumptions it follows that the vapor and liquid flowrates are constant along the column but different in the stripping and enriching sections. Moreover, the reboiler and the condenser dynamics have been described as first order processes. Under the prescribed assumptions, the mass and energy balance equations are decoupled, and the process model is simply constituted of the transient liquid mass balances. Within the framework of this description the tray temperatures are evaluated by computing the bubble points corresponding to the composition of the liquids on the trays. The equilibrium vapor pressures are calculated through the Antoine equation. The adopted dynamic model of the distillation column is rather simplified but, nevertheless, is capable of capturing the most significant dynamical features of the unit that are mainly related to the dynamics of the liquid compositions on the tray. Indeed, these are known to exhibit the largest characteristic times.The design specifications are: i-C5content in the distillate less than 0.5% (on mole basis) and C3content in the residue less than 0.1%. The parameters corresponding to the reference case are summarized in Table 1.MODEL-BASED NEURAL CONTROLLERThe relative gain array analysis has been used to select the most proper controlstrategy, and has indicated the E (reflux ratio)/V (vapor boilup) strategy as the most appropriate for the prescribed objectives. In particular, the distillate prescription should be controlled with V while E should be preferred to regulate the residue specification.The control of the residue specification exhibits rather serious problems due to the high volatility of propane that makes the residue C3 content almost insensitive to E. This results in high gains that lead to serious pitfalls when trying to implement the neural controller that is based on process inversion. Overall, the dual composition control problem exhibits rather ill-conditioned characteristics so as to prevent the neural controller from achieving a reasonable performance. Indeed, for the reference case, the condition number of the steadystate gain matrix is >> 1.In this work, we use a multilayer feed-forward neural network trained through the backpropagation algorithm. The adopted network has three layers: the first layer (input) contains n1 nodes corresponding to the actual net inputs, the second layer (hidden) contains n2nodes, and the third layer (output) contains n3nodes that correspond to the number of the monitored state variables. The configuration of such a network is simply indicated as (n1–n2–n3). The feed and hidden layers are also augmented with a bias unit that represents the threshold value. The input value of the bias node is held constant and equal to 1. Two pilot temperatures, located at the 3rd and 16th tray from the bottom, are also used as inputs to the neural network. It is worth pointing out that we have selected as network inputs those that closely resemble the field measurements available for actual distillation units. Further details on the development of the multilayer neural network, on the selection, and on the pre-processing of the inputs are given elsewhere.For the dual composition problem, three control systems based on the E/V strategy have been developed: a conventional temperature control loop, a neural inferential system, and a neural controller based on process inversion. The inferential control strategy relies on a neural observer that provides compositions as set points to a conventional PI feedback controller. The actions of the inferential and neural controllers are schematically shown in Fig. 2.In the neural controller the actuator actions (on V and E)are modeled as first order filters, with a time constant typical of the industrial control valves. The actuator model is given by the following first order filter equation:τdy/dt+y=xwhere τ represents the time constant of the actuator dev ice (control valve) and is given by Δt acq /α;Δt acq represents the acquisition interval; y is the filtered signal (actuation); x is the manipulated signal, that is, the reflux ratio (E) or the boil-up rate (V) as predicted by the process inverse model. The parameter α governs the response speed of the actuator, in other words, it determines the rate at which the control action is able to compensate for the upsetting loads to the unit. The process model inversion is simply carried out by inverting the trained neural network exchanging the manipulated variables (E and V) with the controlled ones (the C3content in the residue and the i-C5content in the distillate). As anticipated earlier, the dual composition control problem exhibited an ill-character that has prevented us from efficiently performing the process inversion, thus, making unsuccessful any attempt to develop a model-based neural controller. The problem has been overcome by reconsidering the control objectives and discovering that the ill-condition character of the distillation unit is mainly related to the weak sensitivity of the residue composition (C, content) on the reflux ratio, E. This suggested releasing the control specification on the residue by only maintaining the one on the distillate, thus reducing the controlproblem from dual composition to single composition. For the single composition control problem, the analysis of the steady-state gain matrix has led to selection of the reflux flowrate as a manipulated variable to control the distillate specification. The neural network used for the single composition problem has been furtherly simplified to a (5-l-l) configuration that has proved to be very efficient for process inversion, even though it was less accurate than the (7-2-2) network in representing the distillation unit.DISCUSSION OF THE RESULTSFor the reference case, the relevant parameters used for simulating the distillation unit are summarized in Table 1. All the results discussed for the neural controller refer to the single composition control problem where the i-C5content in the distillate (controlled variable) is regulated by acting on the reflux flowrate (manipulated variable).The neural networks used to develop the inferential control strategy and the neural controller were trained using sets of calibration data spanning 50 h of operation. The calibration data were generated by making use of the dynamic simulator previously described. The simulations have been designed so as to capture the principal response features of the distillation units to variations in feed flowrate, feed composition, distillate and residue specifications, reflux and boil-up flowrates, and operating pressure.At first, the performance of the neural controller has been evaluated with respect to the value of the parameter α, to test the response sensitivity of the neural controller to changes in the time constant of the actuator device. Increasing α, the response speed of the neural controller increases, at least up to a critical valueα0, above which the controller response is not significantly affected anymore. In reality, for values of αvery close to 1 the neural controller exhibits an unstable behavior. For the case of a change of set point in the i-C5, content in the distillate, the response of the neural controller is shown in Fig. 3 for two values of a, corresponding to valve time constants of 50s (a) and 125s (b). As expected, increasing the value of α, the response rate of the neural controller increases, and thus it takes a shorter time to compensate for the set point change. Above the value of α= 0.1 no detectable changes in thecontroller response occurred. By further increasing α the neural controller exhibited an unstable behavior above the value of 0.9. This is the reason why in all the following simulations the value of α= 0.1 has been adopted. The results of Fig. 3 also show the insensitivity of the residue composition (C3, mole fraction) that does not exhibit any significant change when the operating conditions are changed to accommodate for the new set point in the i-C5 content in the distillate stream. This explains why the neural controller is capable of meeting both the specifications on distillate and residue, even though it directly acts only on the manipulated variable that regulates the distillate composition.A second series of tests have been carried out to compare the performance of the neural controller with that of the inferential strategy. The results, shown in Fig. 4, refer to a set point change in the distillate composition. The neural observer is based on a (7-2-2) neural network configuration, while the neural controller relies on a (5- 1-1) network topology. The responses of the two control systems (inferential and neural controller) are compared in terms of the capability in set point tracking. The neural controller exhibits a much faster response than the inferential strategy. The first is able to accommodate to the new operating conditions within about 8min, while thesecond requires about 20min. To verify that the control action of the neural controller is indeed achievable, the required load on the manipulated variable is shown in Fig. 5 in terms of the percentage load variation with respect to the reference case. The results show that the neural controller is able to achieve a good setpoint tracking performance by requiring at most a 3% change in the manipulated variable.Finally, the performance of the three control structures (temperature loop, inferential and neural controllers) are compared in Fig. 6 in terms of the disturbance rejection characteristics. The data of Fig. 6, spanning over an operating window of about 1000 h, refer to unit upsets due to fluctuations in feed flowrate (±1-2%), feed composition (±5-15%) and operating pressure (±3-4%). The results show that the control system based on the temperature loop (Fig. 6(a)) fails to meet the desired distillate specification (i-C5 composition), even though the temperature set point, not shown here, is met. The nonunique relationship between temperature and composition, for multicomponent systems, is the prime reason for the observed faulty performance of the temperature control strategy. On the other hand, the C3specification on the residue stream is well satisfied, and this is due to its insensitivity to the operating conditions rather than to an actual efficacy of the temperature loop policy. The tall peaks exhibited by the i-C5response correspond to perturbations in the feed composition that represent the most critical type of upsets to be faced by the temperature loop control.The response of the inferential system is illustrated in Fig. 6(b), which clearly shows the improvement achieved over the temperature loop performance. Theinferential controller also performs well in rejecting the critical disturbances in the feed composition, the response peaks are attenuated by a factor of about 4, with respect to those detected with the temperature loop strategy. The neural controller response is illustrated in Fig. 6(c). The results show that this control strategy by far outperforms the previous two policies since it is capable of completely compensating for the upsetting loadings. In practice, with the neural controller a complete disturbance rejection is achieved. This is mainly due to the intrinsic characteristics of the neural controller that acts on the manipulated variable as soon as the upsetting conditions are detected, and this makes the correction action much faster. CONCLUSIONSFor a multicomponent distillation column representative of an actual stabilizer unit, three control policies are discussed: a conventional temperature loop, an inferential strategy and a neural controller. The results show that control strategies based on process modeling (the inferential and neural controllers) perform better than the conventional temperature control loop.The neural controller is based on process inversion that performed by simply inverting the neural network model of the distillation unit. The results indicate that for ill-conditioned processes, such as the one investigate here, the implementation of nonlinear control strategies based on process inversion requires a careful revision of the control objectives as well as of the input set to be fed to the neural network.The neural controller by far outperforms the temperature loop and the inferential control strategies. In reality, the neural controller is the only one capable of achieving an almost complete disturbance rejection, even for the most critical upsetting loadings represented by fluctuations in the feed composition. . All this suggests that for multicomponent distillation units the conventional temperature loop strategies should be abandoned in favor of more advanced control policies.The results clearly demonstrate the potentials of neural modeling for developing advanced nonlinear control strategies based on process inversion. It is remarkable that even for ill-conditioned processes, such as the one investigated here, neural modeling provides an efficient framework to perform process inversion.译文概论如今,先进的控制系统在工厂运营中发挥着重要的作用,因为它们可以使工厂进行有效的管理。
外文文献原稿和译文原稿The introduction of the electronic commerceWith the rapid development of Internet, websites surging forward are after frying the foot " concept ", it is " practical " to all change direction one after another, and " practical one of the more distinct characteristics ": It is most websites that are all attempting to do the true " e-commerce ". What is " e-commerce " so? The so-called e-commerce (Electronic Commerce ) utilizes technology of the computer, network technology and long-range communication technology, realize the electrification , digitization and networked in the course of the whole commercial affair (buys and sells ). People are no longer face-to-face , look at the true goods , depend on the paper medium document (including the cash ) to carry on the business transaction. But provide and deliver system and fund convenient to be safe and settle account and do a deal systematically through a feast for the eyes goods information on net , perfect logistics through the network (buy and sell).In fact, the course of the whole trade can be divided into three stages:The first stage is information interchange stage: To trade company, in order to release information stage this stage. Mainly choose one's own outstanding goods , organize one's own goods information meticulously, set up one's own webpage, then join in famous websites with greater reputation , with stronger influence power , of higher clicking rate, let as many as possible people understand you know you. To buyer, this stage is the stage going to look for the goods and goods information on the net. It is mainly according to the needs of one's own one, surf the Net and look for the information and goods that oneself needs, and the trade company choosing the service with a high reputation to be well cheap.The second stage signs the contract stage of the goods: As B2B (trade company vs. trade company ), this stage is exchange course of signing the contract , finishing the essential trade and business note. The ones that should pay attention to are: Complicatedproblems , such as accuracy , dependability , can't be alteredding of the data ,etc.. As B2C (trade company vs. personal customer ), this stage finishes doing shopping the order form of the course signs course, the customer should refer goods , one's own contacting information , way , payment method to deliver goods which you choose to the trade company after in the network good sign well, the trade company should send mails or the telephone and verify above-mentioned contents after receiving the order form .The third stage carries on goods handing over , fund settlement stage according to the contract: This stage is stage when trade of the whole goods is very key, not merely involve the putting in place correctly and safely on the net of fund, involve the accuracy that the goods provide and deliver , put in place on time too at the same time . There is banking , provide and deliver systematic involvement at this stage, technically, in law, such respects as the standard is first-class have higher requests. The succeeding or not for a stage just here of online trade.The characteristics of developmentWider environment: People are not restricted by time , is not restricted by space , a great deal of restrictions not doing shopping by the tradition , can conclude the business on the net whenever and wherever possible .Wider market: This world will become very small on the net, a trade company can face global consumers , and a consumer can do shopping in any global trade company .Faster circulation and cheap price: E-commerce has reduce the intermediate link of circulation of commodities, has saved a large amount of expenses, thus reduced the cost of the circulation of commodities and trade greatly .Accord with the demands for times even more: People pursue the fashion , stress the individual character more and more nowadays, pay attention to the environment that does shopping, the online shopping, the shopping course that best reflects individualization. Influence1.E-commerce will change the way of the commercial activityThe traditional most typical scene of commercial activity is that " the salesman exists everywhere ", " the purchasing agent runs everywhere ", " reveal the mouth , run and break the leg ", consumers look for the goods that oneself need all pooped outlay in the market .Now, it is if touch it all right to pass Internet, people can enter the online market to browse through , purchase all kinds of products, and can also get online service , trade company can get in touch with the customer on the net , utilize the network to carry on the payment for goods and close serving, the government can also carry on electron calling for bid , government procurement ,etc. conveniently .2.E-commerce will change people's consumption patternThe biggest characteristic of online shopping is consumers' leading factor, do shopping the will is grasped in consumer's hands, at the same time consumers can also finish the trade by way of a kind of relaxed and free self-service , consumer's sovereignty can be fully embodied in the shopping at network.3.E-commerce will change the modes of production of enterprisesBecause e-commerce is a kind of swift , convenient shopping means, consumers' individualization , priviledges need and be able to be totally shown in front of the manufacturer through the network, in order to please customers, stress the design style of the products, a lot of enterprises in the manufacturing industry develop and popularize e-commerce one after another , if American Ford Motor Company connects with inside network of the company in 120,000 computer work stations in the whole world of March of 1998, and include 15,000 distributors of the whole world in the inside network , the final purpose of Ford Company is to realize being according to users' different requests, supply the automobile as required.4.E-commerce will bring a revolution to the traditional tradeE-commerce is in the course of the whole commercial activity, through the combination of people and telecommunications way , improve the efficiency of the commercial activity greatly, reduce the unnecessary intermediate link, the traditional manufacturing industry enters the eras of short run , much varieties by this, " zero stock " become possible; Traditional retail business and wholesale business open the new mode of " online marketing " of " having no shop "; Various kinds of online services have offered the brand-new method of service for service trade of the tradition.5.E-commerce will bring a brand-new financial circlesBecause pay the key link that is e-commerce by mails online, it is basic condition that e-commerce can be developed smoothly too, with the break-through of e-commerce on the electronic trade link, such services as online bank , bank card pay the network , electronic payment system of bank and electronic cheque , electronic cash, bring traditional financial circles into a brand-new field.6.E-commerce will change the government's behaviorThe government is undertaking the function of the management and service of a large number of societies , economy , culture, in e-commerce era, as enterprises use e-commerce to carry on the production and management, the bank computerizes financial services, and while consumers realize cosumption online, will put forward the new request , an electronic government or calling the online government for government's administration behaviour too, will become an important social role as e-commerce is developed.In a word, as a kind of commercial activity course, e-commerce will bring an unprecedented revolution, its impact on social economy far exceeds the commercial affair itself. Besides these above-mentioned influence, it will also bring enormous influence on employment , legal system and culture and education ,etc., e-commerce will bring the mankind into the information-intensive society .In the modern information-intensive society, e-commerce can enable mastering the enterprises and individuals of the information technology and business rule , utilizing various kinds of electronic tools and networks systematically, the low cost with high efficiency is engaged in various kinds of commercial trade activities realized in electronic way. According to using with the function , can divide e-commerce into three levels or 3S , namely SHOW , SALE , SERVE:1.SHOW (show)Offer the electronic market conditions, enterprises release goods and other information on the net by way of webpage, with advertise etc. on the net, through SHOW, enterprises can establish one's own corporate image , expand the popularity of enterprises, propagate the service of one's own products , look for the new trade cooperative partner.2.SALE (trade)Realize the whole course of the transaction activity of the traditional form in electronic way in the network, for instance online shopping ,etc.. Enterprises can finish the whole course of the trade through SALE, expand the range traded, improve the efficiency of the work , reduce the cost of the trade , thus obtain the economic and social benefit.3.SERVE (serve)And the service after sale before meaning various kinds of related to commercial activity that enterprises launch through the network and selling, through such SERVE on net, enterprises can perfect one's own e-commerce system , consolidate the already existing customer , attract the new customer , thus expand the management business of enterprises, obtain greater economic benefits and social benefit. The enterprise is a protagonist who launches e-commerce.KindsE-commerce inside enterprise is mainly to deal with and exchange trade and business information through the way of the network (Intranet ) inside enterprises. The network (Intranet ) is a kind of effective commercial tool inside enterprises, pass the fire wall , enterprises isolate one's own inside network from Internet, it can be used for automaticallying process the commercial affair to operate and work to flow, strengthen it to the access of important system and key data, share experience , solve customer's problem together, and keep the connection while organizing. E-commerce inside enterprises, can bring the following advantage to enterprise : Increase agility that commercial activity punish, to market situation can reacting quickly, can provide service for the customer better . E-commerce (B-B mode ) among enterprise and enterprise mainly carries on the electronic commercial activity through INTERNET or the special-purpose network way. E-commerce among enterprises is merits attention and probes into most three kinds of modes of e-commerce, because it has potentiality of development most.E-commerce (B-C mode ) between enterprise and consumer is mainly that enterprises offer a new-type shopping environment to consumer through INTERNET --The online shop, consumers do shopping on the net through the network , pay on the net. Because this kind of mode has saved customer and time and space of both sides of enterprise, have improve trade efficiency greatly, has saved the unnecessary expenses.Enterprise and inter-governmental e-commerce (B-G mode) the commercial activity cover enterprise and government organize every thing among, mainly including government procurement , declare at the Customs , declare dutiable goods etc.on the net.译文电子商务介绍随着因特网的迅速发展,风起云涌的网站在炒足了“概念”之后,都纷纷转向了“务实”,而“务实”比较鲜明的特点之一:是绝大多数的网站都在试图做实实在在的“电子商务”。
外文文献翻译原文Analysis of Con tin uous Prestressed Concrete BeamsChris BurgoyneMarch 26, 20051、IntroductionThis conference is devoted to the development of structural analysis rather than the strength of materials, but the effective use of prestressed concrete relies on an appropriate combination of structural analysis techniques with knowledge of the material behaviour. Design of prestressed concrete structures is usually left to specialists; the unwary will either make mistakes or spend inordinate time trying to extract a solution from the various equations.There are a number of fundamental differences between the behaviour of prestressed concrete and that of other materials. Structures are not unstressed when unloaded; the design space of feasible solutions is totally bounded;in hyperstatic structures, various states of self-stress can be induced by altering the cable profile, and all of these factors get influenced by creep and thermal effects. How were these problems recognised and how have they been tackled?Ever since the development of reinforced concrete by Hennebique at the end of the 19th century (Cusack 1984), it was recognised that steel and concrete could be more effectively combined if the steel was pretensioned, putting the concrete into compression. Cracking could be reduced, if not prevented altogether, which would increase stiffness and improve durability. Early attempts all failed because the initial prestress soon vanished, leaving the structure to be- have as though it was reinforced; good descriptions of these attempts are given by Leonhardt (1964) and Abeles (1964).It was Freyssineti’s observations of the sagging of the shallow arches on three bridges that he had just completed in 1927 over the River Allier near Vichy which led directly to prestressed concrete (Freyssinet 1956). Only the bridge at Boutiron survived WWII (Fig 1). Hitherto, it had been assumed that concrete had a Young’s modulus which remained fixed, but he recognised that the de- ferred strains due to creep explained why the prestress had been lost in the early trials. Freyssinet (Fig. 2) also correctly reasoned that high tensile steel had to be used, so that some prestress would remain after the creep had occurred, and alsothat high quality concrete should be used, since this minimised the total amount of creep. The history of Freyssineti’s early prestressed concrete work is written elsewhereFigure1:Boutiron Bridge,Vic h yFigure 2: Eugen FreyssinetAt about the same time work was underway on creep at the BRE laboratory in England ((Glanville 1930) and (1933)). It is debatable which man should be given credit for the discovery of creep but Freyssinet clearly gets the credit for successfully using the knowledge to prestress concrete.There are still problems associated with understanding how prestressed concrete works, partly because there is more than one way of thinking about it. These different philosophies are to some extent contradictory, and certainly confusing to the young engineer. It is also reflected, to a certain extent, in the various codes of practice.Permissible stress design philosophy sees prestressed concrete as a way of avoiding cracking by eliminating tensile stresses; the objective is for sufficient compression to remain after creep losses. Untensionedreinforcement, which attracts prestress due to creep, is anathema. This philosophy derives directly from Freyssinet’s logic and is primarily a working stress concept.Ultimate strength philosophy sees prestressing as a way of utilising high tensile steel as reinforcement. High strength steels have high elastic strain capacity, which could not be utilised when used as reinforcement; if the steel is pretensioned, much of that strain capacity is taken out before bonding the steel to the concrete. Structures designed this way are normally designed to be in compression everywhere under permanent loads, but allowed to crack under high live load. The idea derives directly from the work of Dischinger (1936) and his work on the bridge at Aue in 1939 (Schonberg and Fichter 1939), as well as that of Finsterwalder (1939). It is primarily an ultimate load concept. The idea of partial prestressing derives from these ideas.The Load-Balancing philosophy, introduced by T.Y. Lin, uses prestressing to counter the effect of the permanent loads (Lin 1963). The sag of the cables causes an upward force on the beam, which counteracts the load on the beam. Clearly, only one load can be balanced, but if this is taken as the total dead weight, then under that load the beam will perceive only the net axial prestress and will have no tendency to creep up or down.These three philosophies all have their champions, and heated debates take place between them as to which is the most fundamental.2、Section designFrom the outset it was recognised that prestressed concrete has to be checked at both the working load and the ultimate load. For steel structures, and those made from reinforced concrete, there is a fairly direct relationship between the load capacity under an allowable stress design, and that at the ultimate load under an ultimate strength design. Older codes were based on permissible stresses at the working load; new codes use moment capacities at the ultimate load. Different load factors are used in the two codes, but a structure which passes one code is likely to be acceptable under the other.For prestressed concrete, those ideas do not hold, since the structure is highly stressed, even when unloaded. A small increase of load can cause some stress limits to be breached, while a large increase in load might be needed to cross other limits. The designer has considerable freedom to vary both the working load and ultimate load capacities independently; both need to be checked.A designer normally has to check the tensile and compressive stresses, in both the top and bottom fibre of the section, for every load case. The critical sections are normally, but not always, the mid-span and the sections over piers but other sections may become critical ,when the cable profile has to be determined.The stresses at any position are made up of three components, one of which normally has a different sign from the other two; consistency of sign convention is essential.If P is the prestressing force and e its eccentricity, A and Z are the area of the cross-section and its elastic section modulus, while M is the applied moment, then where ft and fc are the permissible stresses in tension and compression.c e t f ZM Z P A P f ≤-+≤Thus, for any combination of P and M , the designer already has four in- equalities to deal with.The prestressing force differs over time, due to creep losses, and a designer isusually faced with at least three combinations of prestressing force and moment;• the applied moment at the time the prestress is first applied, before creep losses occur,• the maximum applied moment after creep losses, and• the minimum applied moment after creep losses.Figure 4: Gustave MagnelOther combinations may be needed in more complex cases. There are at least twelve inequalities that have to be satisfied at any cross-section, but since an I-section can be defined by six variables, and two are needed to define the prestress, the problem is over-specified and it is not immediately obvious which conditions are superfluous. In the hands of inexperienced engineers, the design process can be very long-winded. However, it is possible to separate out the design of the cross-section from the design of the prestress. By considering pairs of stress limits on the same fibre, but for different load cases, the effects of the prestress can be eliminated, leaving expressions of the form:rangestress e Perm issibl Range Mom entZ These inequalities, which can be evaluated exhaustively with little difficulty, allow the minimum size of the cross-section to be determined.Once a suitable cross-section has been found, the prestress can be designed using a construction due to Magnel (Fig.4). The stress limits can all be rearranged into the form:()M fZ PA Z e ++-≤1 By plotting these on a diagram of eccentricity versus the reciprocal of the prestressing force, a series of bound lines will be formed. Provided the inequalities (2) are satisfied, these bound lines will always leave a zone showing all feasible combinations of P and e. The most economical design, using the minimum prestress, usually lies on the right hand side of the diagram, where the design is limited by the permissible tensile stresses.Plotting the eccentricity on the vertical axis allows direct comparison with the crosssection, as shown in Fig. 5. Inequalities (3) make no reference to the physical dimensions of the structure, but these practical cover limits can be shown as wellA good designer knows how changes to the design and the loadings alter the Magnel diagram. Changing both the maximum andminimum bending moments, but keeping the range the same, raises and lowers the feasible region. If the moments become more sagging the feasible region gets lower in the beam.In general, as spans increase, the dead load moments increase in proportion to the live load. A stage will be reached where the economic point (A on Fig.5) moves outside the physical limits of the beam; Guyon (1951a) denoted the limiting condition as the critical span. Shorter spans will be governed by tensile stresses in the two extreme fibres, while longer spans will be governed by the limiting eccentricity and tensile stresses in the bottom fibre. However, it does not take a large increase in moment ,at which point compressive stresses will govern in the bottom fibre under maximum moment.Only when much longer spans are required, and the feasible region moves as far down as possible, does the structure become governed by compressive stresses in both fibres.3、Continuous beamsThe design of statically determinate beams is relatively straightforward; the engineer can work on the basis of the design of individual cross-sections, as outlined above. A number of complications arise when the structure is indeterminate which means that the designer has to consider, not only a critical section,but also the behaviour of the beam as a whole. These are due to the interaction of a number of factors, such as Creep, Temperature effects and Construction Sequence effects. It is the development of these ideas whichforms the core of this paper. The problems of continuity were addressed at a conference in London (Andrew and Witt 1951). The basic principles, and nomenclature, were already in use, but to modern eyes concentration on hand analysis techniques was unusual, and one of the principle concerns seems to have been the difficulty of estimating losses of prestressing force.3.1 Secondary MomentsA prestressing cable in a beam causes the structure to deflect. Unlike the statically determinate beam, where this motion is unrestrained, the movement causes a redistribution of the support reactions which in turn induces additional moments. These are often termed Secondary Moments, but they are not always small, or Parasitic Moments, but they are not always bad.Freyssinet’s bridge across the Marne at Luzancy, started in 1941 but not completed until 1946, is often thought of as a simply supported beam, but it was actually built as a two-hinged arch (Harris 1986), with support reactions adjusted by means of flat jacks and wedges which were later grouted-in (Fig.6). The same principles were applied in the later and larger beams built over the same river.Magnel built the first indeterminate beam bridge at Sclayn, in Belgium (Fig.7) in 1946. The cables are virtually straight, but he adjusted the deck profile so that the cables were close to the soffit near mid-span. Even with straight cables the sagging secondary momentsare large; about 50% of the hogging moment at the central support caused by dead and live load.The secondary moments cannot be found until the profile is known but the cablecannot be designed until the secondary moments are known. Guyon (1951b) introduced the concept of the concordant profile, which is a profile that causes no secondary moments; es and ep thus coincide. Any line of thrust is itself a concordant profile.The designer is then faced with a slightly simpler problem; a cable profile has to be chosen which not only satisfies the eccentricity limits (3) but is also concordant. That in itself is not a trivial operation, but is helped by the fact that the bending moment diagram that results from any load applied to a beam will itself be a concordant profile for a cable of constant force. Such loads are termed notional loads to distinguish them from the real loads on the structure. Superposition can be used to progressively build up a set of notional loads whose bending moment diagram gives the desired concordant profile.3.2 Temperature effectsTemperature variations apply to all structures but the effect on prestressed concrete beams can be more pronounced than in other structures. The temperature profile through the depth of a beam (Emerson 1973) can be split into three components for the purposes of calculation (Hambly 1991). The first causes a longitudinal expansion, which is normally released by the articulation of the structure; the second causes curvature which leads to deflection in all beams and reactant moments in continuous beams, while the third causes a set of self-equilibrating set of stresses across the cross-section.The reactant moments can be calculated and allowed-for, but it is the self- equilibrating stresses that cause the main problems for prestressed concrete beams. These beams normally have high thermal mass which means that daily temperature variations do not penetrate to the core of the structure. The result is a very non-uniform temperature distribution across the depth which in turn leads to significant self-equilibrating stresses. If the core of the structure is warm, while the surface is cool, such as at night, then quite large tensile stresses can be developed on the top and bottom surfaces. However, they only penetrate a very short distance into the concrete and the potential crack width is very small. It can be very expensive to overcome the tensile stress by changing the section or the prestress。
外文文献翻译译稿数字频率计使用DMA终端计数停止方法Digital frequency meter using DMATerminal Count stop method1.N.Suresh Kumar,Assistant Professor,Dept of IT,GITAM University.2.Dr.D.V.RamaKotiReddy,Associate professor,College of Engineering,AndhraUniversity.3.Mr.B.pramod Kumar,M.Tech.Student,Gitam University,Visakhapatnam.4.Mr.A.Harish,M.Tech,Lecturer,Dept.of IT.摘要本文提出了一种新型宽量程转速测量方法,使用直接内存访问(DMA)终端计数寄存器(TCR)。
DMA的方法是基于在不断采样的过程中在终端计数DMA控制器的停止销的脉冲计数。
提出了单片机实现所需的硬件配置和演算方法。
通过DMA 控制器和定时器来使用单片机的这个方法是适合系统需求的。
同时对该方法的局限性和来源的误差进行了详细讨论。
DMA终端计数寄存器的方法是适用于实时速度控制系统的。
一.导言:可以通过下列方法对速度进行测量1.测量时间,确定脉冲之间的间隔时间[1]2.脉冲计数,计数采样时间内输入的脉冲[2]3.相结合的方法[3]4.持续访问时长法(CET)[4]5.DMA传输方法。
[4][5]当DMA传输方法的硬件配置执行了一个长的处理器指令,在接收到下一个输入脉冲的上升沿之前,DMA应答信号(DACK)可能无法被接收到。
因此,下一个脉冲将不会被检测出来。
为了不丢失任何输入脉冲,就需要对未履行的DMA请求计数器。
这也可以通过一个简单的H/W配置和更低的功耗方法来解决,这个方法可以通过终端计数寄存器和DMA控制器的TC停止销来实现。
DMA控制器的停止销在固定数量的DMA周期后会改变自己的状态。
儿童教育外文翻译文献(文档含中英文对照即英文原文和中文翻译)原文:The Role of Parents and Community in the Educationof the Japanese ChildHeidi KnipprathAbstractIn Japan, there has been an increased concern about family and community participation in the child’s educat ion. Traditionally, the role of parents and community in Japan has been one of support and less one of active involvement in school learning. Since the government commenced education reforms in the last quarter of the 20th century, a more active role for parents and the community in education has been encouraged. These reforms have been inspired by the need to tackle various problems that had arisen, such as the perceived harmful elements of society’spreoccupation with academic achievement and the problematic behavior of young people. In this paper, the following issues are examined: (1) education policy and reform measures with regard to parent and community involvement in the child’s education; (2) the state of parent and community involvement at the eve of the 20th century.Key Words: active involvement, community, education reform, Japan, parents, partnership, schooling, supportIntroduction: The Discourse on the Achievement GapWhen western observers are tempted to explain why Japanese students attain high achievement scores in international comparative assessment studies, they are likely to address the role of parents and in particular of the mother in the education of the child. Education mom is a phrase often brought forth in the discourse on Japanese education to depict the Japanese mother as being a pushy, and demanding home-bound tutor, intensely involved in the child’s education due to severe academic competition. Although this image of the Japanese mother is a stereotype spread by the popular mass media in Japan and abroad, and the extent by which Japanese mothers are absorbed in their children is exaggerated (Benjamin, 1997, p. 16; Cummings, 1989, p. 297; Stevenson & Stigler, 1992, p. 82), Stevenson and Stigler (1992) argue that Japanese parents do play an indispensable role in the academic performance of their children. During their longitudinal and cross-national research project, they and their collaborators observed that Japanese first and fifth graders persistently achieved higher on math tests than American children. Besides reciting teacher’s teaching style, cultural beliefs, and organization of schooling, Stevenson and Stigler (1992) mention parent’s role in supporting the learning conditions of the child to explain differences in achievement between elementary school students of the United States and students of Japan. In Japan, children receive more help at home with schoolwork (Chen & Stevenson, 1989; Stevenson & Stigler, 1992), and tend to perform less household chores than children in the USA (Stevenson et al., 1990; Stevenson & Stigler, 1992). More Japanese parents than American parents provide space and a personal desk and purchase workbooks for their children to supplement their regular text-books at school (Stevenson et al., 1990; Stevenson & Stigler, 1992). Additionally, Stevenson and Stigler (1992) observed that American mothers are much more readily satisfied with their child’s performance than Asian parents are, have less realistic assessments of their child’s academic perform ance, intelligence, and other personality characteristics, and subsequently have lower standards. Based on their observation of Japanese, Chinese and American parents, children and teachers, Stevenson and Stigler (1992) conclude that American families can increase the academic achievement of their children by strengthening the link between school and home, creating a physical and psychological environment that is conducive to study, and by making realistic assessments and raising standards. Also Benjamin (1997), who performed ‘day-to-day ethnography’ to find out how differences in practice between American and Japanese schools affect differences in outcomes, discusses the relationship between home and school and how the Japanese mother is involved in the academic performance standards reached by Japanese children. She argues that Japanese parents are willing to pay noticeable amounts of money for tutoring in commercial establishments to improve the child’s performance on entrance examinations, to assist in ho mework assignments, to facilitate and support their children’s participation in school requirements and activities, and to check notebooks of teachers on the child’s progress and other school-related messages from the teacher. These booklets are read and written daily by teachers and parents. Teachers regularly provide advice and reminders to parents, and write about homework assignments of the child, special activities and the child’s behavior (Benjamin, 1997, p. 119, p. 1993–1995). Newsletters, parents’ v isits to school, school reports, home visits by the teacher and observation days sustain communication in later years at school. According toBenjamin (1997), schools also inform parents about how to coach their children on proper behavior at home. Shimahara (1986), Hess and Azuma (1991), Lynn (1988) and White (1987) also try to explain national differences in educational achievement. They argue that Japanese mothers succeed in internalizing into their children academic expectations and adaptive dispositions that facilitate an effective teaching strategy, and in socializing the child into a successful person devoted to hard work.Support, Support and SupportEpstein (1995) constructed a framework of six types of involvement of parents and the community in the school: (1) parenting: schools help all families establish home environments to support children as students; (2) communicating: effective forms of school-to-home and home-to-school communications about school programs and children’s progress; (3) volu nteering: schools recruit and organize parents help and support; (4) learning at home: schools provide information and ideas to families about how to help students at home with homework and other curriculum-related activities, decisions and planning; (5) decision making: schools include parents in school decisions, develop parent leaders and representatives; and (6) collaborating with the community: schools integrate resources and services from the community to strengthen school programs, family practices, and student learning and development. All types of involvement mentioned in studies of Japanese education and in the discourse on the roots of the achievement gap belong to one of Epstein’s first four types of involvement: the creation of a conducive learn ing environment (type 4), the expression of high expectations (type 4), assistance in homework (type 4), teachers’ notebooks (type 2), mother’s willingness to facilitate school activities (type3) teachers’ advice about the child’s behavior (type 1), observ ation days by which parents observe their child in the classroom (type 2), and home visits by the teachers (type 1). Thus, when one carefully reads Stevenson and Stigler’s, Benjamin’s and other’s writings about Japanese education and Japanese students’ high achievement level, one notices that parents’ role in the child’s school learning is in particular one of support, expected and solicited by the school. The fifth type (decision making) as well as the sixth type (community involvement) is hardly ever mentioned in the discourse on the achievement gap.In 1997, the OECD’s Center for Educational Research and Innovation conducted a cross-national study to report the actual state of parents as partners in schooling in nine countries, including Japan. In its report, OECD concludes that the involvement of Japanese parents in their schools is strictly limited, and that the basis on which it takes place tends to be controlled by the teacher (OECD, 1997, p. 167). According to OECD (1997), many countries are currently adopting policies to involve families closely in the education of their children because (1) governments are decentralizing their administrations; (2) parents want to be increasingly involved; and (3) because parental involvement is said to be associated with higher achievement in school (p. 9). However, parents in Japan, where students already score highly on international achievement tests, are hardly involved in governance at the national and local level, and communication between school and family tends to be one-way (Benjamin, 1997; Fujita, 1989; OECD, 1997). Also parent–teacher associations (PTA, fubo to kyoshi no kai ) are primarily presumed to be supportive of school learning and not to participate in school governance (cf. OECD, 2001, p. 121). On the directionsof the occupying forces after the second world war, PTA were established in Japanese schools and were considered with the elective education boards to provide parents and the community an opportunity to participate actively in school learning (Hiroki, 1996, p. 88; Nakata, 1996, p. 139). The establishment of PTA and elective education boards are only two examples of numerous reform measures the occupying forces took to decentralize the formal education system and to expand educational opportunities. But after they left the country, the Japanese government was quick to undo liberal education reform measures and reduced the community and parental role in education. The stipulation that PTA should not interfere with personnel and other administrative tasks of schools, and the replacement of elective education boards by appointed ones, let local education boards believe that parents should not get involved with school education at all (Hiroki, 1996, p. 88). Teachers were regarded to be the experts and the parents to be the laymen in education (Hiroki, 1996, p. 89).In sum, studies of Japanese education point into one direction: parental involvement means being supportive, and community involvement is hardly an issue at all. But what is the actual state of parent and community involvement in Japanese schools? Are these descriptions supported by quantitative data?Statistics on Parental and Community InvolvementTo date, statistics of parental and community involvement are rare. How-ever, the school questionnaire of the TIMSS-R study did include some interesting questions that give us a clue about the degree of involvement relatively compared to the degree of involvement in other industrialized countries. The TIMSS-R study measured science and math achievement of eighth graders in 38 countries. Additionally, a survey was held among principals, teachers and students. Principals answered questions relating to school management, school characteristics, and involvement. For convenience, the results of Japan are only compared with the results of those countries with a GNP of 20650 US dollars or higher according to World Bank’s indicators in 1999.Unfortunately, only a very few items on community involvement were measured. According to the data, Japanese principals spend on average almost eight hours per month on representing the school in the community (Table I). Australian and Belgian principals spend slightly more hours and Dutch and Singaporean principals spend slightly less on representing the school and sustaining communication with the community. But when it comes to participation from the community, Japanese schools report a nearly absence of involvement (Table II). Religious groups and the business community have hardly any influence on the curriculum of the school. In contrast, half of the principals report that parents do have an impact in Japan. On one hand, this seems a surprising result when one is reminded of the centralized control of the Ministry of Education. Moreover, this control and the resulting uniform curriculum are often cited as a potential explanation of the high achievement levels in Japan. On the other hand, this extent of parental impact on the curriculum might be an indicator of the pressure parents put on schools to prepare their children appropriately for the entrance exams of senior high schools.In Table III, data on the extent of other types of parental involvement in Japan and other countries are given. In Japan, parental involvement is most common in case of schools volunteering for school projects and programs, and schools expecting parents to make sure that thechild completes his or her homework. The former is together with patrolling the grounds of the school to monitor student behavior most likely materialized through the PTA. The kinds and degree of activities of PTA vary according to the school, but the activities of the most active and well-organized PTA’s of 395 elementary schools investigated by Sumida (2001)range from facilitating sport and recreation for children, teaching greetings, encouraging safe traffic, patrolling the neighborhood, publishing the PTA newspaper to cleaning the school grounds (pp. 289–350). Surprisingly, less Japanese principals expect from the parents to check one’s child’s completion of homework than principals of other countries. In the discourse on the achievement gap, western observers report that parents and families in Japan provide more assistance with their children’s homework than parents and families outside Japan. This apparent contradiction might be the result of the fact that these data are measured at the lower secondary level while investigations of the roots of Japanese students’ high achievement levels focus on childhood education and learning at primary schools. In fact, junior high school students are given less homework in Japan than their peers in other countries and less homework than elementary school students in Japan. Instead, Japanese junior high school students spend more time at cram schools. Finally, Japanese principals also report very low degrees of expectations toward parents with regard to serving as a teacher aid in the classroom, raising funds for the school, assisting teachers on trips, and serving on committees which select school personnel and review school finances. The latter two items measure participation in school governance.In other words, the data support by and large the descriptions of parental of community involvement in Japanese schooling. Parents are requested to be supportive, but not to mount the territory of the teacher nor to be actively involved in governance. Moreover, whilst Japanese principals spend a few hours per month on communication toward the community, involvement from the community with regard to the curriculum is nearly absent, reflecting the nearly absence of accounts of community involvement in studies on Japanese education. However, the reader needs to be reminded that these data are measured at the lower secondary educational level when participation by parents in schooling decreases (Epstein, 1995; OECD, 1997; Osakafu Kyoiku Iinkai, unpublished report). Additionally, the question remains what stakeholders think of the current state of involvement in schooling. Some interesting local data provided by the Osaka Prefecture Education Board shed a light on their opinion.ReferencesBenjamin, G. R. (1997). Japanese lessons. New York: New York University Press.Cave, P. (2003). Educational reform in Japan in the 1990s: ‘Individuality’ and other uncertainties. Comparative Education Review, 37(2), 173–191.Chen, C., & Stevenson, H. W. (1989). Homework: A cross-cultural examination. Child Development, 60(3), 551–561.Chuo Kyoiku Shingikai (1996). 21 seiki o tenbo shita wagakuni no kyoiku no arikata ni tsu-ite [First Report on the Model for Japanese Education in the Perspective of theCummings, W. K. (1989). The American perception of Japanese parative Education, 25(3), 293–302.Epstein, J. L. (1995). School/family/community partnerships. Phi Delta Kappan , 701–712.Fujita, M. (1989). It’s all mother’s fault: childcare and the socialization of working mothers in Japan. The Journal of Japanese Studies , 15(1), 67–91.Harnish, D. L. (1994). Supplemental education in Japan: juku schooling and its implication. Journal of Curriculum Studies , 26(3), 323–334.Hess, R. D., & Azuma, H. (1991). Cultural support for schooling, contrasts between Japanand the United States. Educational Researcher , 20(9), 2–8, 12.Hiroki, K. (1996). Kyoiku ni okeru kodomo, oya, kyoshi, kocho no kenri, gimukankei[Rights and duties of principals, teachers, parents and children in education. InT. Horio & T. Urano (Eds.), Soshiki toshite no gakko [School as an organization](pp. 79–100). Tokyo: Kashiwa Shobo. Ikeda, H. (2000). Chiiki no kyoiku kaikaku [Local education reform]. Osaka: Kaiho Shup-pansha.Kudomi, Y., Hosogane, T., & Inui, A. (1999). The participation of students, parents and the community in promoting school autonomy: case studies in Japan. International Studies in Sociology of Education, 9(3), 275–291.Lynn, R. (1988).Educational achievement in Japan. London: MacMillan Press.Martin, M. O., Mullis, I. V. S., Gonzalez, E. J., Gregory, K. D., Smith, T. A., Chrostowski,S. J., Garden, R. A., & O’Connor, K. M. (2000). TIMSS 1999 Intern ational science report, findings from IEA’s Repeat of the Third International Mathematics and ScienceStudy at the Eight Grade.Chestnut Hill: The International Study Center.Mullis, I. V. S., Martin, M. O., Gonzalez, E. J., Gregory, K. D., Garden, R. A., O’Connor, K. M.,Chrostowski, S. J., & Smith, T. A.. (2000). TIMSS 1999 International mathemat-ics report, findings from IEA’s Repeat of the Third International Mathematics and Science Study at the Eight Grade.Chestnut Hill: The International Study Center. Ministry of Education, Science, Sports and Culture (2000).Japanese government policies in education, science, sports and culture. 1999, educational reform in progress. Tokyo: PrintingBureau, Ministry of Finance.Monbusho Ed. (1999).Heisei 11 nendo, wagakuni no bunkyoshisaku : Susumu kaikaku [Japanese government policies in education, science, sports and culture 1999: Educational reform in progress]. Tokyo: Monbusho.Educational Research for Policy and Practice (2004) 3: 95–107 © Springer 2005DOI 10.1007/s10671-004-5557-6Heidi KnipprathDepartment of MethodologySchool of Business, Public Administration and TechnologyUniversity of Twente P.O. Box 2177500 AE Enschede, The Netherlands译文:家长和社区在日本儿童教育中的作用摘要在日本,人们越来越关心家庭和社区参与到儿童教育中。
外文文献翻译译稿1卡尔曼滤波的一个典型实例是从一组有限的,包含噪声的,通过对物体位置的观察序列(可能有偏差)预测出物体的位置的坐标及速度。
在很多工程应用(如雷达、计算机视觉)中都可以找到它的身影。
同时,卡尔曼滤波也是控制理论以及控制系统工程中的一个重要课题。
例如,对于雷达来说,人们感兴趣的是其能够跟踪目标。
但目标的位置、速度、加速度的测量值往往在任何时候都有噪声。
卡尔曼滤波利用目标的动态信息,设法去掉噪声的影响,得到一个关于目标位置的好的估计。
这个估计可以是对当前目标位置的估计(滤波),也可以是对于将来位置的估计(预测),也可以是对过去位置的估计(插值或平滑)。
命名[编辑]这种滤波方法以它的发明者鲁道夫.E.卡尔曼(Rudolph E. Kalman)命名,但是根据文献可知实际上Peter Swerling在更早之前就提出了一种类似的算法。
斯坦利。
施密特(Stanley Schmidt)首次实现了卡尔曼滤波器。
卡尔曼在NASA埃姆斯研究中心访问时,发现他的方法对于解决阿波罗计划的轨道预测很有用,后来阿波罗飞船的导航电脑便使用了这种滤波器。
关于这种滤波器的论文由Swerling(1958)、Kalman (1960)与Kalman and Bucy(1961)发表。
目前,卡尔曼滤波已经有很多不同的实现。
卡尔曼最初提出的形式现在一般称为简单卡尔曼滤波器。
除此以外,还有施密特扩展滤波器、信息滤波器以及很多Bierman, Thornton开发的平方根滤波器的变种。
也许最常见的卡尔曼滤波器是锁相环,它在收音机、计算机和几乎任何视频或通讯设备中广泛存在。
以下的讨论需要线性代数以及概率论的一般知识。
卡尔曼滤波建立在线性代数和隐马尔可夫模型(hidden Markov model)上。
其基本动态系统可以用一个马尔可夫链表示,该马尔可夫链建立在一个被高斯噪声(即正态分布的噪声)干扰的线性算子上的。
系统的状态可以用一个元素为实数的向量表示。
随着离散时间的每一个增加,这个线性算子就会作用在当前状态上,产生一个新的状态,并也会带入一些噪声,同时系统的一些已知的控制器的控制信息也会被加入。
同时,另一个受噪声干扰的线性算子产生出这些隐含状态的可见输出。
卡尔曼滤波是一种递归的估计,即只要获知上一时刻状态的估计值以及当前状态的观测值就可以计算出当前状态的估计值,因此不需要记录观测或者估计的历史信息。
卡尔曼滤波器与大多数滤波器不同之处,在于它是一种纯粹的时域滤波器,它不需要像低通滤波器等频域滤波器那样,需要在频域设计再转换到时域实现。
卡尔曼滤波器的状态由以下两个变量表示:,在时刻k的状态的估计;,误差相关矩阵,度量估计值的精确程度。
卡尔曼滤波器的操作包括两个阶段:预测与更新。
在预测阶段,滤波器使用上一状态的估计,做出对当前状态的估计。
在更新阶段,滤波器利用对当前状态的观测值优化在预测阶段获得的预测值,以获得一个更精确的新估计值。
预测(预测状态)(预测估计协方差矩阵)更新首先要算出以下三个量:(测量余量,measurement residual)(测量余量协方差)(最优卡尔曼增益)然后用它们来更新滤波器变量x与P:(更新的状态估计)(更新的协方差估计)使用上述公式计算仅在最优卡尔曼增益的时候有效。
使用其他增益的话,公式要复杂一些不变量(Invariant)如果模型准确,而且与的值准确的反映了最初状态的分布,那么以下不变量就保持不变:所有估计的误差均值为零且协方差矩阵准确的反映了估计的协方差:请注意,其中表示的期望值, 。
实例考虑在无摩擦的、无限长的直轨道上的一辆车。
该车最初停在位置0处,但时不时受到随机的冲击。
我们每隔Δt秒即测量车的位置,但是这个测量是非精确的;我们想建立一个关于其位置以及速度的模型。
我们来看如何推导出这个模型以及如何从这个模型得到卡尔曼滤波器。
因为车上无动力,所以我们可以忽略掉B k和u k。
由于F、H、R和Q是常数,所以时间下标可以去掉。
车的位置以及速度(或者更加一般的,一个粒子的运动状态)可以被线性状态空间描述如下:其中是速度,也就是位置对于时间的导数。
我们假设在(k− 1)时刻与k时刻之间,车受到a k的加速度,其符合均值为0,标准差为σa的正态分布。
根据牛顿运动定律,我们可以推出其中且我们可以发现(因为σa是一个标量)。
在每一时刻,我们对其位置进行测量,测量受到噪声干扰。
我们假设噪声服从正态分布,均值为0,标准差为σz。
其中且如果我们知道足够精确的车最初的位置,那么我们可以初始化并且,我们告诉滤波器我们知道确切的初始位置,我们给出一个协方差矩阵:如果我们不确切的知道最初的位置与速度,那么协方差矩阵可以初始化为一个对角线元素是B的矩阵,B取一个合适的比较大的数。
此时,与使用模型中已有信息相比,滤波器更倾向于使用初次测量值的信息。
§推§推导后验协方差矩阵按照上边的定义,我们从误差协方差开始推导如下:代入再代入与整理误差向量,得因为测量误差v k与其他项是非相关的,因此有利用协方差矩阵的性质,此式可以写作使用不变量P k|k-1以及R k的定义这一项可以写作:这一公式对于任何卡尔曼增益K k都成立。
如果K k是最优卡尔曼增益,则可以进一步简化,请见下文。
§最优卡尔曼增益的推导卡尔曼滤波器是一个最小均方误差估计器,后验状态误差估计(英文:a posteriori state estimate)是我们最小化这个矢量幅度平方的期望值,,这等同于最小化后验估计协方差矩阵P k|k的迹(trace)。
将上面方程中的项展开、抵消,得到:当矩阵导数是0的时候得到P k|k的迹(trace)的最小值:此处须用到一个常用的式子,如下:从这个方程解出卡尔曼增益K k:这个增益称为最优卡尔曼增益,在使用时得到最小均方误差。
§后验误差协方差公式的化简在卡尔曼增益等于上面导出的最优值时,计算后验协方差的公式可以进行简化。
在卡尔曼增益公式两侧的右边都乘以S k K k T得到根据上面后验误差协方差展开公式,最后两项可以抵消,得到.这个公式的计算比较简单,所以实际中总是使用这个公式,但是需注意这公式仅在使用最优卡尔曼增益的时候它才成立。
如果算术精度总是很低而导致数值稳定性出现问题,或者特意使用非最优卡尔曼增益,那么就不能使用这个简化;必须使用上面导出的后验误差协方差公式。
自适应滤波器是能够根据输入信号自动调整性能进行数字信号处理的数字滤波器。
作为对比,非自适应滤波器有静态的滤波器系数,这些静态系数一起组成传递函数。
对于一些应用来说,由于事先并不知道所需要进行操作的参数,例如一些噪声信号的特性,所以要求使用自适应的系数进行处理。
在这种情况下,通常使用自适应滤波器,自适应滤波器使用反馈来调整滤波器系数以及频率响应。
总的来说,自适应的过程涉及到将代价函数用于确定如何更改滤波器系数从而减小下一次迭代过程成本的算法。
价值函数是滤波器最佳性能的判断准则,比如减小输入信号中的噪声成分的能力。
随着数字信号处理器性能的增强,自适应滤波器的应用越来越常见,时至今日它们已经广泛地用于手机以及其它通信设备、数码录像机和数码照相机以及医疗监测设备中假设医院正在监测一个患者的心脏跳动,即心电图,这个信号受到50 Hz(许多国家供电所用频率)噪声的干扰剔除这个噪声的方法之一就是使用50Hz 的陷波滤波器(en:notch filter)对信号进行滤波。
但是,由于医院的电力供应会有少许波动,所以我们假设真正的电力供应可能会在47Hz 到53Hz 之间波动。
为了剔除47 到53Hz 之间的频率的静态滤波器将会大幅度地降低心电图的质量,这是因为在这个阻带之内很有可能就有心脏跳动的频率分量。
为了避免这种可能的信息丢失,可以使用自适应滤波器。
自适应滤波器将患者的信号与电力供应信号直接作为输入信号,动态地跟踪噪声波动的频率。
这样的自适应滤波器通常阻带宽度更小,这就意味着这种情况下用于医疗诊断的输出信号就更加准确。
扩展卡尔曼滤波器在扩展卡尔曼滤波器(Extended Kalman Filter,简称EKF)中状态转换和观测模型不需要是状态的线性函数,可替换为(可微的)函数。
函数f可以用来从过去的估计值中计算预测的状态,相似的,函数h可以用来以预测的状态计算预测的测量值。
然而f和h不能直接的应用在协方差中,取而代之的是计算偏导矩阵(Jacobian)。
在每一步中使用当前的估计状态计算Jacobian矩阵,这几个矩阵可以用在卡尔曼滤波器的方程中。
这个过程,实质上将非线性的函数在当前估计值处线性化了。
这样一来,卡尔曼滤波器的等式为:预测使用Jacobians矩阵更新模型更新预测如同扩展卡尔曼滤波器(EKF)一样, UKF的预测过程可以独立于UKF的更新过程之外,与一个线性的(或者确实是扩展卡尔曼滤波器的)更新过程合并来使用;或者,UKF的预测过程与更新过程在上述中地位互换亦可。
外文文献翻译原文1Kalman filtering, also known as linear quadratic estimation(LQE), isan algorithm that uses a series of measurements observed over time,containing noise(random variations) and other inaccuracies, and producesestimates of unknown variables that tend to be more precise than those based on a single measurement alone. More formally, the Kalman filter operates recursively on streams of noisy input data to produce a statistically optimal estimate of the underlying system state. The filter is named after Rudolf (Rudy) E. Kálmán, one of the primary developers of its theory.The Kalman filter has numerous applications in technology. A common application is for guidance, navigation and control of vehicles, particularly aircraft and spacecraft. Furthermore, the Kalman filter is a widely applied concept in timeseries analysis used in fields such as signal processing and econometrics. Kalman filters also are one of the main topics in the field of Robotic motion planning and control, and sometimes included in Trajectory optimization.The algorithm works in a two-step process. In the prediction step, the Kalman filter produces estimates of the current state variables, along with their uncertainties. Once the outcome of the next measurement (necessarily corrupted with some amount of error, including random noise) is observed, these estimates are updated using a weighted average, with more weight being given to estimates with higher certainty. Because of the algorithm's recursive nature, it can run in real time using only the present input measurements and the previously calculated state and its uncertainty matrix; no additional past information is required.It is a common misconception that the Kalman filter assumes that all error terms and measurements are Gaussian distributed. Kalman's original paper derived the filter using orthogonal projection theory to show that the covariance is minimized, and this result does not require any assumption, e.g., that the errors are Gaussian.[1]He then showed that the filter yields the exact conditional probability estimate in the special case that all errors are Gaussian-distributed.Extensions and generalizations to the method have also been developed, such asthe extended Kalman filter and the unscented Kalman filter which work on nonlinear systems. The underlying model is a Bayesian model similar toa hidden Markov model but where the state space of the latent variables is continuous and where all latent and observed variables have Gaussian distributions.The Kalman filters are based on linear dynamic systems discretized in the time domain.They are modelled on a Markov chain built on linear operators perturbed by errors thatmay include Gaussian noise. The state of the system is represented as a vector of realnumbers. At each discrete time increment, a linear operator is applied to the state togenerate the new state, with some noise mixed in, and optionally some information fromthe controls on the system if they are known. Then, another linear operator mixed withmore noise generates the observed outputs from the true ("hidden") state. The Kalmanfilter may be regarded as analogous to the hidden Markov model, with the key difference that the hidden state variables take values in a continuous space (as opposed to a discrete state space as in the hidden Markov model).The Kalman filter is a recursive estimator. This means that only the estimated state from the previous time step and the current measurement are needed to compute the estimate for the current state. In contrast to batch estimation techniques, no history ofobservations and/or estimates is required. In what follows, the notation represents the estimate of at time n given observations up to, and including at time m ≤ n.The state of the filter is represented by two variables:•, the a posteriori state estimate at time k given observations up to and including at time k;•, the a posteriori error covariance matrix (a measure of the estimated accuracy of the state estimate).The Kalman filter can be written as a single equation, however it is most often conceptualized as two distinct phases: "Predict" and "Update". The predict phase uses the state estimate from the previous timestep to produce an estimate of the state at the current timestep. This predicted state estimate is also known as the a priori state estimate because, although it is an estimate of the state at the current timestep, it does not include observation information from the current timestep. In the update phase, the current a priori prediction is combined with current observation information to refine the state estimate. This improved estimate is termed the a posteriori state estimate.Typically, the two phases alternate, with the prediction advancing the state until the next scheduled observation, and the update incorporating the observation. However, this is not necessary; if an observation is unavailable for some reason, the update may be skipped and multiple prediction steps performed. Likewise, if multiple independent observations are available at the same time, multiple update steps may be performed (typically with different observation matrices H k).[14][15]§PredictPredicted (a priori) state estimatePredicted (a priori) estimate covariance§UpdateInnovation or measurement residualInnovation (or residual) covarianceOptimal Kalman gainUpdated (a posteriori) state estimateUpdated (a posteriori) estimate covarianceThe formula for the updated estimate and covariance above is only valid for the optimal Kalman gain. Usage of other gain values require a more complex formula found inthe derivations section.InvariantsIf the model is accurate, and the values for and accurately reflect thedistribution of the initial state values, then the following invariants are preserved: (all estimates have a mean error of zero)••where is the expected value of , and covariance matrices accurately reflect the covariance of estimates•••Example application, technical[edit]Consider a truck on frictionless, straight rails. Initially the truck is stationary at position 0, but it is buffeted this way and that by random uncontrolled forces. We measure the position of the truck every Δt seconds, but these measurements are imprecise; we want to maintain a model of where the truck is and what its velocity is. We show here how wederive the model from which we create our Kalman filter.Since are constant, their time indices are dropped.The position and velocity of the truck are described by the linear state spacewhere is the velocity, that is, the derivative of position with respect to time.We assume that between the (k − 1) and k timestep uncontrolled forces cause a constant acceleration of ak that is normally distributed, with mean 0 and standard deviationσa. From Newton's laws of motion we conclude that(note that there is no term since we have no known control inputs. Instead, we assume that ak is the effect of an unknown input and applies that effect to the state vector) whereandso thatwhere andAt each time step, a noisy measurement of the true position of the truck is made. Let us suppose the measurement noise vk is also normally distributed, with mean 0 and standard deviation σz.whereandWe know the initial starting state of the truck with perfect precision, so we initializeand to tell the filter that we know the exact position and velocity, we give it a zero covariance matrix:If the initial position and velocity are not known perfectly the covariance matrix should be initialized with a suitably large number, say L, on its diagonal.The filter will then prefer the information from the first measurements over the information already in the model.Deriving the a posteriori estimate covariance matrixStarting with our invariant on the error covariance Pk | k as abovesubstitute in the definition ofand substituteandand by collecting the error vectors we getSince the measurement error vk is uncorrelated with the other terms, this becomesby the properties of vector covariance this becomeswhich, using our invariant on Pk | k−1 and the definition of Rk becomesThis formula (sometimes known as the "Joseph form" of the covariance update equation) is valid for any value of Kk. It turns out that if Kk is the optimal Kalman gain, this can be simplified further as shown below.Kalman gain derivationThe Kalman filter is a minimum mean-square error estimator. The error in the a posteriori state estimation isWe seek to minimize the expected value of the square of the magnitude of this vector, . This is equivalent to minimizing the trace of the a posterioriestimate covariance matrix . By expanding out the terms in the equation above and collecting, we get:The trace is minimized when its matrix derivative with respect to the gain matrix is zero. Using the gradient matrix rules and the symmetry of the matrices involved we find thatSolving this for Kk yields the Kalman gain:This gain, which is known as the optimal Kalman gain, is the one that yields MMSE estimates when used.Simplification of the a posteriori error covariance formulaThe formula used to calculate the a posteriori error covariance can be simplified when the Kalman gain equals the optimal value derived above. Multiplying both sides of our Kalman gain formula on the right by SkKkT, it follows thatReferring back to our expanded formula for the a posteriori error covariance,we find the last two terms cancel out, givingThis formula is computationally cheaper and thus nearly always used in practice, but is only correct for the optimal gain. If arithmetic precision is unusually low causing problems with numerical stability, or if a non-optimal Kalman gain is deliberately used, this simplification cannot be applied; the a posteriori error covariance formula as derived above must be used.An adaptive filter is a system with a linear filter that has a transfer function controlled by variable parameters and a means to adjust those parameters according to an optimization algorithm. Because of the complexity of the optimization algorithms, most adaptive filters are digital filters. Adaptive filters are required for some applications because some parameters of the desired processing operation (for instance, the locations of reflective surfaces in a reverberant space) are not known in advance or are changing. The closed loop adaptive filter uses feedback in the form of an error signal to refine its transfer function.Generally speaking, the closed loop adaptive process involves the use of a cost function, which is a criterion for optimum performance of the filter, to feed an algorithm, whichdetermines how to modify filter transfer function to minimize the cost on the next iteration. The most common cost function is the mean square of the error signal.As the power of digital signal processors has increased, adaptive filters have become much more common and are now routinely used in devices such as mobile phones and other communication devices, camcorders and digital cameras, and medical monitoring equipment.Assuming the hospital is monitoring a patient's heart beating, namely, ECG, the signal is 50 Hz (frequency is used by many countries supply) noiseNotch filter method to eliminate noise of this is the use of 50Hz (en:notch filter) of the signal filtering. However, because of the power supply in hospital. There will be a little fluctuation, sowe assume that the power supply real may fluctuate in the 47Hz to 53Hz. In order to eliminate47 to static filters will greatly reduce the frequency of 53Hz between the ECG quality, this isbecause in the stopband within might well have a frequency component of beating heart.In order to avoid the possible loss of information, you can use the adaptive filter. The adaptive filter will supply signal and power of patients directly as the input signal, dynamicallytracking noise fluctuation frequency. Adaptive filter this usually stopband width is smaller,which means in this case an output signal for medical diagnosis is more accurate.Hybrid Kalman filter[edit]Most physical systems are represented as continuous-time models while discrete-time measurements are frequently taken for state estimation via a digital processor. Therefore, the system model and measurement model are given bywhere.InitializePredictThe prediction equations are derived from those of continuous-time Kalman filter without update from measurements, i.e., . The predicted state and covariance are calculated respectively by solving a set of differential equations with the initial value equal to the estimate at the previous step.UpdateThe update equations are identical to those of the discrete-time Kalman filter.外文文献翻译译稿2外文文献翻译原文2。