Hadoop云计算外文翻译文献

Hadoop云计算外文翻译文献

(文档含中英文对照即英文原文和中文翻译)

原文:

Meet Hadoop

In pioneer days they used oxen for heavy pulling, and when one ox couldn’t budge a log, they didn’t try to grow a larger ox. We shouldn’t be trying for bigger computers, but for more systems of computers.

—Grace Hopper Data!

We live in the data age. It’s not easy to measure the total volume of data stored electronically, but an IDC estimate put the size of the “digital universe” at 0.18 zettabytes in

2006, and is forecasting a tenfold growth by 2011 to 1.8 zettabytes. A zettabyte is 1021 bytes, or equivalently one thousand exabytes, one million petabytes, or one billion terabytes. That’s roughly the same order of magnitude as one disk drive for every person in the world.

This flood of data is coming from many sources. Consider the following:

? The New York Stock Exchange generates about one terabyte of new trade data per

day.

? Facebook hosts approximately 10 billion photos, taking up one petabyte of storage.

? https://www.360docs.net/doc/1d229508.html,, the genealogy site, stores around 2.5 petabytes of data.

? The Internet Archive stores around 2 petabytes of data, and is growing at a rate of

20 terabytes per month.

? The Large Hadron Collider near Geneva, Switzerland, will produce about 15 petabytes of data per year.

So there’s a lot of data out there. But you are probably wondering how it affects you.

Most of the data is locked up in the largest web properties (like search engines), or

scientific or financial institutions, isn’t it? Does the advent of “Big Data,” as it is being called, affect smaller organizations or individuals?

I argue that it does. Take photos, for example. My wife’s grandfather was an avid photographer, and took photographs throughout his adult life. His entire corpus of medium format, slide, and 35mm film, when scanned in at high-resolution, occupies around 10 gigabytes. Compare this to the digital photos that my family took last year,which take up about 5 gigabytes of space. My family is producing photographic data at 35 times the rate my wife’s grandfather’s did, and the rate is increasing every year as it becomes easier to take more and more photos.

More generally, the digital streams that individuals are producing are growing apace. Microsoft Research’s MyLifeBits project gives a glimpse of archiving of pe rsonal information that may become commonplace in the near future. MyLifeBits was an experiment where an individual’s interactions—phone calls, emails, documents were captured electronically and stored for later access. The data gathered included a photo taken every minute, which resulted in an overall data volume of one gigabyte a month. When storage costs come down enough to make it feasible to store continuous audio and video, the data volume for a future MyLifeBits service will be many times that.

The t rend is for every individual’s data footprint to grow, but perhaps more importantly the amount of data generated by machines will be even greater than that generated by people. Machine logs, RFID readers, sensor networks, vehicle GPS traces, retail transactions—all of these contribute to the growing mountain of data.

The volume of data being made publicly available increases every year too. Organizations no longer have to merely manage their own data: success in the future will be dictated to a large extent by their ability to extract value from other organizations’ data.Initiatives such as Public Data Sets on Amazon Web Services, https://www.360docs.net/doc/1d229508.html,, and https://www.360docs.net/doc/1d229508.html, exist to foster the “information commons,” where data can be freely (or in the case of AWS, for a modest price) shared for anyone to download and analyze. Mashups between different information sources make for unexpected and hitherto unimaginable applications.

Take, for example, the https://www.360docs.net/doc/1d229508.html, project, which watches the Astrometry group

on Flickr for new photos of the night sky. It analyzes each image, and identifies which part of the sky it is from, and any interesting celestial bodies, such as stars or galaxies. Although it’s still a new and experimental service, it shows the kind of things that are possible when data (in this case, tagged photographic images) is made available andused for something (image analysis) that was not anticipated by the creator.

It has been said that “More data usually beats better algorithms,” which is to say that for some problems (such as recommending movies or music based on past preferences),however fiendish your algorithms are, they can often be beaten simply by having more data (and a less sophisticated algorithm).

The good news is that Big Data is here. The bad news is that we are struggling to store and analyze it.

Data Storage and Analysis

The problem is simple: while the storage capacities of hard drives have increased massively over the years, access speeds--the rate at which data can be read from drives--have not kept up. One typical drive from 1990 could store 1370 MB of data and had a transfer speed of 4.4 MB/s, so you could read all the data from a full drive in around five minutes. Almost 20

years later one terabyte drives are the norm, but the transfer speed is around 100 MB/s, so it takes more than two and a half hours to read all the data off the disk.

This is a long time to read all data on a single drive and writing is even slower. The obvious way to reduce the time is to read from multiple disks at once. Imagine if we had 100 drives, each holding one hundredth of the data. Working in parallel, we could read the data in under two minutes.

Only using one hundredth of a disk may seem wasteful. But we can store one hundred datasets, each of which is one terabyte, and provide shared access to them. We can imagine that the users of such a system would be happy to share access in return for shorter analysis times, and, statistically, that their analysis jobs would be likely to be spread over time, so they wouldn`t interfere with each other too much.

There`s more to being able to read and write data in parallel to or from multiple disks, though. The first problem to solve is hardware failure: as soon as you start using many pieces of hardware, the chance that one will fail is fairly high. A common way of avoiding data loss is through replication: redundant copies of the data are kept by the system so that in the event of failure, there is another copy available. This is how RAID works, for instance, although Hadoop`s filesystem, the Hadoop Distributed Filesystem (HDFS),takes a slightly different approach, as you shall see later. The second problem is that most analysis tasks need to be able to combine the data in some way; data read from one disk may need to be combined with the data from any of the other 99 disks. Various distributed systems allow data to be combined from multiple sources, but doing this correctly is notoriously challenging. MapReduce provides a programming model that abstracts the problem from disk reads and writes, transforming it into a computation over sets of keys and values. We will look at the details of this model in later chapters, but the important point for the present discussion is that there are two parts to the computation, the map a nd the reduce, and it’s the interface between the two where the “mixing” occurs. Like HDFS, MapReduce has reliability built-in.

This, in a nutshell, is what Hadoop provides: a reliable shared storage and analysis system. The storage is provided by HDFS, and analysis by MapReduce. There are other parts to Hadoop, but these capabilities are its kernel.

Comparison with Other Systems

The approach taken by MapReduce may seem like a brute-force approach. The premise is that the entire dataset—or at least a good portion of it—is processed for each query. But this is its power. MapReduce is a batch query processor, and the ability to run an ad hoc query against your whole dataset and get the results in a reasonable time is transformative. It changes the way you think about data, and unlocks data that was previously archived on tape or disk. It gives people the opportunity to innovate with data. Questions that took too long to get answered before can now be answered, which in turn leads to new questions and new insights.

For example, Mailtrust, Rackspace’s mail division, used Hadoop for processing email logs. One ad hoc query they wrote was to find the geographic distribution of their users.

In their words: This data was so useful that we’ve scheduled the MapReduce job to run monthly and we will be using this data to help us decide which Rackspace data centers to place new mail servers in as we grow. By bringing several hundred gigabytes of data together and having the tools to analyze it, the Rackspace engineers were able to gain an understanding of the data that they otherwise would never have had, and, furthermore, they were able to use what they had learned to improve the service for their customers. You can read more about how Rackspace uses Hadoop in Chapter 14.

RDBMS

Why can’t we use databases with lots of disks to do large-scale batch analysis? Why is MapReduce needed? The answer to these questions comes from another trend in disk drives: seek time is improving more slowly than transfer rate. Seeking is the process of moving the disk’s head to a particular place on the disk to read or write data. It characterizes the latency of a disk operation, whereas the transfer rate corresponds to a disk’s bandwidth.

If the data access pattern is dominated by seeks, it will take longer to read or write large portions of the dataset than streaming through it, which operates at the transfer rate. On the other hand, for updating a small proportion of records in a database, a traditional B-Tree (the data structure used in relational databases, which is limited by the rate it can perform seeks) works well. For updating the majority of a database, a B-Tree is less efficient than MapReduce, which uses Sort/Merge to rebuild the database.

In many ways, MapReduce can be seen as a complement to an RDBMS. (The differences between the two systems are shown in Table 1-1.) MapReduce is a good fit for problems that

need to analyze the whole dataset, in a batch fashion, particularly for ad hoc analysis. An RDBMS is good for point queries or updates, where the dataset has been indexed to deliver low-latency retrieval and update times of a relatively small amount of data. MapReduce suits applications where the data is written once, and read many times, whereas a relational database is good for datasets that are continually updated.

Table 1-1. RDBMS compared to MapReduce

Traditional RDBMS MapReduce

Data size Gigabytes Petabytes

Access Interactive and batch Batch

Write once, read many times Updates Read and write many

times

Structure Static schema Dynamic schema

Integrity High Low

Scaling Nonlinear Linear

Another difference between MapReduce and an RDBMS is the amount of structure in the datasets that they operate on. Structured data is data that is organized into entities that have a defined format, such as XML documents or database tables that conform to a particular predefined schema. This is the realm of the RDBMS. Semi-structured data, on the other hand, is looser, and though there may be a schema, it is often ignored, so it may be used only as a guide to the structure of the data: for example, a spreadsheet, in which the structure is the grid of cells, although the cells themselves may hold anyform of data. Unstructured data does not have any particular internal structure: for example, plain text or image data. MapReduce works well on unstructured or semistructured data, since it is designed to interpret the data at processing time. In other words, the input keys and values for MapReduce are not an intrinsic property of the data, but they are chosen by the person analyzing the data.

Relational data is often normalized to retain its integrity, and remove redundancy. Normalization poses problems for MapReduce, since it makes reading a record a nonlocal

operation, and one of the central assumptions that MapReduce makes is that it is possible to perform (high-speed) streaming reads and writes.

A web server log is a good example of a set of records that is not normalized (for example, the client hostnames are specified in full each time, even though the same client may appear many times), and this is one reason that logfiles of all kinds are particularly well-suited to analysis with MapReduce.

MapReduce is a linearly scalable programming model. The programmer writes two functions—a map function and a reduce function—each of which defines a mapping from one set of key-value pairs to another. These functions are oblivious to the size of the data or the cluster that they are operating on, so they can be used unchanged for a small dataset and for a massive one. More importantly, if you double the size of the input data, a job will run twice as slow. But if you also double the size of the cluster, a job will run as fast as the original one. This is not generally true of SQL queries.

Over time, however, the differences between relational databases and MapReduce systems are likely to blur. Both as relational databases start incorporating some of the ideas from MapReduce (such as Aster Data’s and Greenplum’s databases), and, from the other direction, as higher-level query languages built on MapReduce (such as Pig and Hive) make MapReduce systems more approachable to traditional database programmers.

Grid Computing

The High Performance Computing (HPC) and Grid Computing communities have been doing large-scale data processing for years, using such APIs as Message Passing Interface (MPI). Broadly, the approach in HPC is to distribute the work across a cluster of machines, which access a shared filesystem, hosted by a SAN. This works well for predominantly compute-intensive jobs, but becomes a problem when nodes need to access larger data volumes (hundreds of gigabytes, the point at which MapReduce really starts to shine), since the network bandwidth is the bottleneck, and compute nodes become idle.

MapReduce tries to colocate the data with the compute node, so data access is fast since it is local. This feature, known as data locality, is at the heart of MapReduce and is the reason for its good performance. Recognizing that network bandwidth is the most precious resource in a data center environment (it is easy to saturate network links by copying data around),

MapReduce implementations go to great lengths to preserve it by explicitly modelling network topology. Notice that this arrangement does not preclude high-CPU analyses in MapReduce.

MPI gives great control to the programmer, but requires that he or she explicitly handle the mechanics of the data flow, exposed via low-level C routines and constructs, such as sockets, as well as the higher-level algorithm for the analysis. MapReduce operates only at the higher level: the programmer thinks in terms of functions of key and value pairs, and the data flow is implicit.

Coordinating the processes in a large-scale distributed computation is a challenge. The hardest aspect is gracefully handling partial failure—when you don’t know if a remote process has failed or not—and still making progress with the overall computation. MapReduce spares the programmer from having to think about failure, since the implementation detects failed map or reduce tasks and reschedules replacements on machines that are healthy. MapReduce is able to do this since it is a shared-nothing architecture, meaning that tasks have no dependence on one other. (This is a slight oversimplification, since the output from mappers is fed to the reducers, but this is under the control of the MapReduce system; in this case, it needs to take more care rerunning a failed reducer than rerunning a failed map, since it has to make sure it can retrieve the necessary map outputs, and if not, regenerate them by running the relevant maps again.) So from the programmer’s point of view, the order in which the tasks run doesn’t matter. By contrast, MPI programs have to explicitly manage their own checkpointing and recovery, which gives more control to the programmer, but makes them more difficult to write.

MapReduce might sound like quite a restrictive programming model, and in a sense it

is: you are limited to key and value types that are related in specified ways, and mappers and reducers run with very limited coordination between one another (the mappers pass keys and values to reducers). A natural question to ask is: can you do anything useful or nontrivial with it?

The answer is yes. MapReduce was invented by engineers at Google as a system for building production search indexes because they found themselves solving the same problem over and over again (and MapReduce was inspired by older ideas from the functional programming, distributed computing, and database communities), but it has since been used for many other applications in many other industries. It is pleasantly surprising to see the range of algorithms that can be expressed in MapReduce, from image analysis, to graph-based problems,

to machine learning algorithms. It can’t solve every problem, of course, but it is a general data-processing tool.

You can see a sample of some of the applications that Hadoop has been used for in Chapter 14.

Volunteer Computing

When people first hear about Hadoop and MapReduce, they oft en ask, “How is it different from SETI@home?” SETI, the Search for Extra-Terrestrial Intelligence, runs a project called SETI@home in which volunteers donate CPU time from their otherwise idle computers to analyze radio telescope data for signs of intelligent life outside earth. SETI@home is the most well-known of many volunteer computing projects; others include the Great Internet Mersenne Prime Search (to search for large prime numbers) and Folding@home (to understand protein folding, and how it relates to disease).

Volunteer computing projects work by breaking the problem they are trying to solve into chunks called work units, which are sent to computers around the world to be analyzed. For example, a SETI@home work unit is about 0.35 MB of radio telescope data, and takes hours or days to analyze on a typical home computer. When the analysis is completed, the results are sent back to the server, and the client gets another work unit. As a precaution to combat cheating, each work unit is sent to three different machines, and needs at least two results to agree to be accepted.

Although SETI@home may be superficially similar to MapReduce (breaking a problem into independent pieces to be worked on in parallel), there are some significant differences. The SETI@home problem is very CPU-intensive, which makes it suitable for running on hundreds of thousands of computers across the world, since the time to transfer the work unit is dwarfed by the time to run the computation on it. Volunteers are donating CPU cycles, not bandwidth.

MapReduce is designed to run jobs that last minutes or hours on trusted, dedicated hardware running in a single data center with very high aggregate bandwidth interconnects. By contrast, SETI@home runs a perpetual computation on untrusted machines on the Internet with highly variable connection speeds and no data locality.

译文:

初识Hadoop

古时候,人们用牛来拉重物,当一头牛拉不动一根圆木的时候,他们不曾想过培育个头更大的牛。同样,我们也不需要尝试更大的计算机,而是应该开发更多的计算系统。

--格蕾斯·霍珀数据

我们生活在数据时代!很难估计全球存储的电子数据总量是多少,但是据IDC估计2006年"数字全球"项目(digital universe)的数据总量为0.18 ZB,并且预测到2011年这个数字将达到1.8 ZB,为2006年的10倍。1 ZB相当于10的21次方字节的数据,或者相当于1000 EB,1000000 PB,或者大家更熟悉的10亿TB的数据!这相当于世界上每个人一个磁盘驱动器的数量级。

这一数据洪流有许多来源。考虑下文:

纽约证券交易所每天产生1 TB的交易数据。

著名社交网站Facebook的主机存储着约100亿张照片,占据PB级存储空间。

https://www.360docs.net/doc/1d229508.html,,一个家谱网站,存储着2.5 PB数据。

互联网档案馆(The Internet Archive)存储着约2 PB数据,并以每月至少20 TB的速度增长。

瑞士日内瓦附近的大型强子对撞机每年产生约15 PB的数据。

此外还有大量数据。但是你可能会想它对自己有何影响。大部分数据被锁定在最大的

网页内容里面(如搜索引擎)或者是金融和科学机构,对不对?是不是所谓的"大数据"的出现会影响到较小的组织或个人?

新技术云计算外文文献

云计算——新兴的计算技术 摘要:云计算是涉及通过互联网提供托管服务的总称。这些服务大致分为三类:基础设施即服务(IaaS)、平台即服务(PaaS)和软件即服务(SaaS)。云计算这个名字的灵感来自于云符号经常用来代表在互联网上流程图和图表。这是在继主机计算、个人电脑计算、客户端服务器计算和Web计算之后的第五代计算技术。本文将围绕云计算进行讨论。 关键词:云计算,IaaS(基础设施即服务),PaaS的(平台即服务),SaaS(软件即服务) 1引言 云服务有三个鲜明的特点区别于传统的主机服务模式,它们分别是:云服务的出售通常按分钟或小时收取费用;云服务是有弹性的,一个用户可以在不同的时间拥有可多可少的服务;云服务完全由供应商托管(消费者只需要通过个人电脑和互联网就可以使用)。虚拟化的重大创新、分布式计算的发展,以及高速互联网的建设和经济的衰落,都加速了对云计算的兴趣。 云可以是私有的或公有的。公有云向互联网上的任何人销售(目前,亚马逊的网络服务是最大的公有云服务提供商)。私有云是一个专有网络或数据中心,向一部分人提供托管服务。当服务提供商使用公有云资源来创建自己的私有云,这样的结果被称为虚拟化的私有云。私有云或公共云的云计算目标是提供方便的、可扩展的计算资源和IT服务[1]。 2云计算的优势 云计算具有的优势是什么? (a)最小化的资本开支 (b)位置和设备独立性 答案:供应商的视角:申请厂商更容易吸引新客户。 (a)提供最低成本的方法和配套应用; (b)能够使用商品服务器和存储硬件; 3云计算的障碍 从客户的视角来看,云计算的障碍有: (a)数据安全; (b)很多客户不希望他们的数据迁移到可以信任的“云”上; (c)数据必须进行本地保留; (d)延迟; (e)云可以走多少毫秒; (f)不是实时应用的理想选择; (g)应用程序可用性; (h)无法通过现有的传统应用进行切换; (i)等效的云应用不存在; 总结,并非所有的应用程序都要工作在公共云之上。

ASP外文翻译原文

https://www.360docs.net/doc/1d229508.html, https://www.360docs.net/doc/1d229508.html, 是一个统一的 Web 开发模型,它包括您使用尽可能少的代码生成企业级 Web 应用程序所必需的各种服务。https://www.360docs.net/doc/1d229508.html, 作为 .NET Framework 的一部分提供。当您编写 https://www.360docs.net/doc/1d229508.html, 应用程序的代码时,可以访问 .NET Framework 中的类。您可以使用与公共语言运行库 (CLR) 兼容的任何语言来编写应用程序的代码,这些语言包括 Microsoft Visual Basic、C#、JScript .NET 和 J#。使用这些语言,可以开发利用公共语言运行库、类型安全、继承等方面的优点的https://www.360docs.net/doc/1d229508.html, 应用程序。 https://www.360docs.net/doc/1d229508.html, 包括: ?页和控件框架 ?https://www.360docs.net/doc/1d229508.html, 编译器 ?安全基础结构 ?状态管理功能 ?应用程序配置 ?运行状况监视和性能功能 ?调试支持 ?XML Web services 框架 ?可扩展的宿主环境和应用程序生命周期管理 ?可扩展的设计器环境 https://www.360docs.net/doc/1d229508.html, 页和控件框架是一种编程框架,它在 Web 服务器上运行,可以动态地生成和呈现 https://www.360docs.net/doc/1d229508.html, 网页。可以从任何浏览器或客户端设备请求 https://www.360docs.net/doc/1d229508.html, 网页,https://www.360docs.net/doc/1d229508.html, 会向请求浏览器呈现标记(例如 HTML)。通常,您可以对多个浏览器使用相同的页,因为 https://www.360docs.net/doc/1d229508.html, 会为发出请求的浏览器呈现适当的标记。但是,您可以针对诸如 Microsoft Internet Explorer 6 的特定浏览器设计https://www.360docs.net/doc/1d229508.html, 网页,并利用该浏览器的功能。https://www.360docs.net/doc/1d229508.html, 支持基于 Web 的设备(如移动电话、手持型计算机和个人数字助理 (PDA))的移动控件。

人工智能专业外文翻译-机器人

译文资料: 机器人 首先我介绍一下机器人产生的背景,机器人技术的发展,它应该说是一个科学技术发展共同的一个综合性的结果,同时,为社会经济发展产生了一个重大影响的一门科学技术,它的发展归功于在第二次世界大战中各国加强了经济的投入,就加强了本国的经济的发展。另一方面它也是生产力发展的需求的必然结果,也是人类自身发展的必然结果,那么随着人类的发展,人们在不断探讨自然过程中,在认识和改造自然过程中,需要能够解放人的一种奴隶。那么这种奴隶就是代替人们去能够从事复杂和繁重的体力劳动,实现人们对不可达世界的认识和改造,这也是人们在科技发展过程中的一个客观需要。 机器人有三个发展阶段,那么也就是说,我们习惯于把机器人分成三类,一种是第一代机器人,那么也叫示教再现型机器人,它是通过一个计算机,来控制一个多自由度的一个机械,通过示教存储程序和信息,工作时把信息读取出来,然后发出指令,这样的话机器人可以重复的根据人当时示教的结果,再现出这种动作,比方说汽车的点焊机器人,它只要把这个点焊的过程示教完以后,它总是重复这样一种工作,它对于外界的环境没有感知,这个力操作力的大小,这个工件存在不存在,焊的好与坏,它并不知道,那么实际上这种从第一代机器人,也就存在它这种缺陷,因此,在20世纪70年代后期,人们开始研究第二代机器人,叫带感觉的机器人,这种带感觉的机器人是类似人在某种功能的感觉,比如说力觉、触觉、滑觉、视觉、听觉和人进行相类比,有了各种各样的感觉,比方说在机器人抓一个物体的时候,它实际上力的大小能感觉出来,它能够通过视觉,能够去感受和识别它的形状、大小、颜色。抓一个鸡蛋,它能通过一个触觉,知道它的力的大小和滑动的情况。第三代机器人,也是我们机器人学中一个理想的所追求的最高级的阶段,叫智能机器人,那么只要告诉它做什么,不用告诉它怎么去做,它就能完成运动,感知思维和人机通讯的这种功能和机能,那么这个目前的发展还是相对的只是在局部有这种智能的概念和含义,但真正完整意义的这种智能机器人实际上并没有存在,而只是随着我们不断的科学技术的发展,智能的概念越来越丰富,它内涵越来越宽。 下面我简单介绍一下我国机器人发展的基本概况。由于我们国家存在很多其

毕业设计外文翻译原文.

Optimum blank design of an automobile sub-frame Jong-Yop Kim a ,Naksoo Kim a,*,Man-Sung Huh b a Department of Mechanical Engineering,Sogang University,Shinsu-dong 1,Mapo-ku,Seoul 121-742,South Korea b Hwa-shin Corporation,Young-chun,Kyung-buk,770-140,South Korea Received 17July 1998 Abstract A roll-back method is proposed to predict the optimum initial blank shape in the sheet metal forming process.The method takes the difference between the ?nal deformed shape and the target contour shape into account.Based on the method,a computer program composed of a blank design module,an FE-analysis program and a mesh generation module is developed.The roll-back method is applied to the drawing of a square cup with the ˉange of uniform size around its periphery,to con?rm its validity.Good agreement is recognized between the numerical results and the published results for initial blank shape and thickness strain distribution.The optimum blank shapes for two parts of an automobile sub-frame are designed.Both the thickness distribution and the level of punch load are improved with the designed blank.Also,the method is applied to design the weld line in a tailor-welded blank.It is concluded that the roll-back method is an effective and convenient method for an optimum blank shape design.#2000Elsevier Science S.A.All rights reserved. Keywords:Blank design;Sheet metal forming;Finite element method;Roll-back method

指纹识别系统(文献综述)

指纹识别方法的综述 摘 要: 对在指纹的预处理和特征提取、指纹分类、指纹的匹配过程中的方向图、滤波器、神经网络等关 键性原理和技术做了详细的说明,并对在各个过程中用到的方法做了进一步的比较,讨论了各种方法的优越性。 0 引 言 自动指纹识别是上世纪六十年代兴起的,利用计算机取代人工来进行指纹识别的一种方法。近年 来,随着计算机技术的飞速发展,低价位指纹采集仪的出现以及高可靠算法的实现,更使得自动指纹识 别技术越来越多地进入到人们的生活和工作中,自动指纹识别系统的研究和开发正在成为国内外学术 界和商业界的热点。相对于其他生物特征鉴别技术例如语音识别及虹膜识别,指纹识别具有许多独到 的优点,更重要的是它具有很高的实用性和可行性,已经被认为是一种理想的身份认证技术,有着十分 广泛的应用前景,是将来生物特征识别技术的主流。 1 指纹取像 图 1 是一个自动指纹识别系统AFIS(Automated Fingerprint Identification System) 的简单流程。 → → → ↓ ↑ ———— 将一个人的指纹采集下来输入计算机进行处理是指纹自动识别的首要步骤。指纹图像的获取主要利用设备取像,方便实用,比较适合AFIS 。利用设备取像的主要方法又利用光学设备、晶体传感器和超声波来进行。光学取像设备是根据光的全反射原理来设计的。晶体传感器取像是根据谷线和脊线皮肤与传感器之间距离不同而产生的电容不同来设计的。超声波设备取像也是采用光波来取像,但由于超声波波长较短,抗干扰能力较强,所以成像的质量非常好。 2 图像的预处理与特征提取 无论采取哪种方法提取指纹,总会给指纹图像带来各种噪声。预处理的目的就是去除图像中的噪 音,把它变成一幅清晰的点线图,以便于提取正确的指纹特征。预处理是指纹自动识别过程的第一步, 它的好坏直接影响着指纹识别的效果。常用的预处理与特征提取( Image Preprocessing and Feature Ex 2 t raction) 方法的主要步骤包括方向图计算、图像滤波、二值化、细化、提取特征和后处理。当然这些步骤 可以根据系统和应用的具体情况再进行适当变化。文献[ 1 ]提出了基于脊线跟踪的方法能够指纹取像 图像预处理 特征提取 指纹识别 数据库管理

最新云计算中侧信道攻击的防御-毕业论文外文翻译整理

附录一英文文献 Security against Side Channel Attack in Cloud Computing Bhrugu Sevak Abstract--Cloud computing is a word that delivering hosted service over the internet. Cloud computing has been ideate as the next generation architecture of IT enterprise ecause of it’s provides ubiquitous network, cost reducing, flexibility and scalability to users. Now days with the fast growing of cloud computing technology introduces new more vulnerabilities so security is considered to be one of the most critical aspect in clod computing environment due to the confidential and important information stored in the cloud. As per AMAZONE EC2 service case study it is possible to identify the particular target VM(virtual machine) in internal cloud infrastructure and then placed new VM with targeted VM and extract confidential information from targeted VM on same physical machine called as simple side channel attack. This paper introduces how to avert the side channel attack in cloud computing. This is accomplished by using combination of Virtual firewall appliance and randomly encryption decryption (using concept of confusion diffusion) and provide RAS (Reliability, Availability, and Security) of client’s data or information. Keywords--Cloud computing, side channel attack, Amazon EC2 service case study, virtual firewall appliance, randomly encryption decryption. I. INTRODUCTION Cloud computing is a word that delivering hosted service over the internet.

文献综述_人工智能

人工智能的形成及其发展现状分析 冯海东 (长江大学管理学院荆州434023) 摘要:人工智能的历史并不久远,故将从人工智能的出现、形成、发展现 状及前景几个方面对其进行分析,总结其发展过程中所出现的问题,以及发展现状中的不足之处,分析其今后的发展方向。 关键词:人工智能,发展过程,现状分析,前景。 一.引言 人工智能最早是在1936年被英国的科学家图灵提出,并不为多数人所认知。 当时,他编写了一个下象棋的程序,这就是最早期的人工智能的应用。也有著名的“图灵测试”,这也是最初判断是否是人工智能的方案,因此,图灵被尊称为“人工智能之父”。人工智能从产生到发展经历了一个起伏跌宕的过程,直到目前为止,人工智能的应用技术也不是很成熟,而且存在相当的缺陷。 通过搜集的资料,将详细的介绍人工智能这个领域的具体情况,剖析其面临的挑战和未来的前景。 二.人工智能的发展历程 1. 1956年前的孕育期 (1) 从公元前伟大的哲学家亚里斯多德(Aristotle)到16世纪英国哲学家培根(F. Bacon),他们提出的形式逻辑的三段论、归纳法以及“知识就是力量”的警句,都对人类思维过程的研究产生了重要影响。 (2)17世纪德国数学家莱布尼兹(G..Leibniz)提出了万能符号和推理计算思想,为数理逻辑的产生和发展奠定了基础,播下了现代机器思维设计思想的种子。而19世纪的英国逻辑学家布尔(G. Boole)创立的布尔代数,实现了用符号语言描述人类思维活动的基本推理法则。 (3) 20世纪30年代迅速发展的数学逻辑和关于计算的新思想,使人们在计算机出现之前,就建立了计算与智能关系的概念。被誉为人工智能之父的英国天才的数学家图灵(A. Tur-ing)在1936年提出了一种理想计算机的数学模型,即图灵机之后,1946年就由美国数学家莫克利(J. Mauchly)和埃柯特(J. Echert)研制出了世界上第一台数字计算机,它为人工智能的研究奠定了不可缺少的物质基础。1950年图灵又发表了“计算机与智能”的论文,提出了著名的“图灵测试”,形象地指出什么是人工智能以及机器具有智能的标准,对人工智能的发展产生了极其深远的影响。 (4) 1934年美国神经生理学家麦克洛奇(W. McCulloch) 和匹兹(W. Pitts )建立了第一个神经网络模型,为以后的人工神经网络研究奠定了基础。 2. 1956年至1969年的诞生发育期 (1)1956年夏季,麻省理工学院(MIT)的麦卡锡(J.McCarthy)、明斯基(M. Minshy)、塞尔夫里奇(O. Selfridge)与索罗门夫(R. Solomonff)、 IBM的洛

中国的对外贸易外文翻译及原文

外文翻译 原文 Foreign T rade o f China Material Source:W anfang Database Author:Hitomi Iizaka 1.Introduction On December11,2001,China officially joined the World T rade Organization(WTO)and be c a me its143rd member.China’s presence in the worl d economy will continue to grow and deepen.The foreign trade sector plays an important andmultifaceted role in China’s economic development.At the same time, China’s expanded role in the world economy is beneficial t o all its trading partners. Regions that trade with China benefit from cheaper and mor e varieties of imported consumer goods,raw materials and intermediate products.China is also a large and growing export market.While the entry of any major trading nation in the global trading system can create a process of adjustment,the o u t c o me is fundamentally a win-win situation.In this p aper we would like t o provide a survey of the various institutions,laws and characteristics of China’s trade.Among some of the findings, we can highlight thefollowing: ?In2001,total trade to gross domestic pr oduct(GDP)ratio in China is44% ?In2001,47%of Chinese trade is processed trade1 ?In2001,51%of Chinese trade is conduct ed by foreign firms in China2 ?In2001,36%of Chinese exports originate from Gu an gdon g province ?In2001,39%of China’s exports go through Hong Kong to be re-exported elsewhere 2.Evolution of China’s Trade Regime Equally remarkable are the changes in the commodity composition of China’s exports and imports.Table2a shows China’s annu al export volumes of primary goods and manufactured goods over time.In1980,primary goods accounted for 50.3%of China’s exports and manufactured goods accounted for49.7%.Although the share of primary good declines slightly during the first half of1980’s,it remains at50.6%in1985.Since then,exports of manufactured goods have grown at a much

虹膜识别外文翻译文献

虹膜识别外文翻译文献 虹膜识别外文翻译文献 (文档含中英文对照即英文原文和中文翻译) 外文: The first chapter 1.1 The research background of iris recognition Biometrics is a technology for personal identification using physiological characteristics and behavior characteristics inherent in the human body. Can be used for the biological characteristics of biological recognition, fingerprint, hand type face, iris, retina, pulse, ear etc.. Behavior has the following characteristics: signature, voice, gait, etc.. Based on these characteristics, it has been the development of hand shape recognition, fingerprint recognition, facial recognition, iris recognition, signature recognition and other biometric technology, many techniques have been formed and mature to application of. Biological recognition technology in a , has a long history, the ancient Egyptians through identification of each part of the body size measure to carry out identity may be the earliest human based on the earliest history of biometrics. But the modern biological recognition technology began in twentieth Century 70 time metaphase, as biometric devices early is relatively expensive, so only a higher security level atomic test, production base.due to declining cost of microprocessor and various electronic components, precision gradually improve, control device of a biological recognition technology has been gradually applied to commerce authorized, such as access control, attendance management, management system, safety certification field etc.. All biometric technology, iris recognition is currently used as a convenient and accurate.

Hadoop云计算外文翻译文献

Hadoop云计算外文翻译文献 (文档含中英文对照即英文原文和中文翻译) 原文: Meet Hadoop In pioneer days they used oxen for heavy pulling, and when one ox couldn’t budge a log, they didn’t try to grow a larger ox. We shouldn’t be trying for bigger computers, but for more systems of computers. —Grace Hopper Data! We live in the data age. It’s not easy to measure the total volume of data stored electronically, but an IDC estimate put the size of the “digital universe” at 0.18 zettabytes in

2006, and is forecasting a tenfold growth by 2011 to 1.8 zettabytes. A zettabyte is 1021 bytes, or equivalently one thousand exabytes, one million petabytes, or one billion terabytes. That’s roughly the same order of magnitude as one disk drive for every person in the world. This flood of data is coming from many sources. Consider the following: ? The New York Stock Exchange generates about one terabyte of new trade data per day. ? Facebook hosts approximately 10 billion photos, taking up one petabyte of storage. ? https://www.360docs.net/doc/1d229508.html,, the genealogy site, stores around 2.5 petabytes of data. ? The Internet Archive stores around 2 petabytes of data, and is growing at a rate of 20 terabytes per month. ? The Large Hadron Collider near Geneva, Switzerland, will produce about 15 petabytes of data per year. So there’s a lot of data out there. But you are probably wondering how it affects you. Most of the data is locked up in the largest web properties (like search engines), or scientific or financial institutions, isn’t it? Does the advent of “Big Data,” as it is being called, affect smaller organizations or individuals? I argue that it does. Take photos, for example. My wife’s grandfather was an avid photographer, and took photographs throughout his adult life. His entire corpus of medium format, slide, and 35mm film, when scanned in at high-resolution, occupies around 10 gigabytes. Compare this to the digital photos that my family took last year,which take up about 5 gigabytes of space. My family is producing photographic data at 35 times the rate my wife’s grandfather’s did, and the rate is increasing every year as it becomes easier to take more and more photos. More generally, the digital streams that individuals are producing are growing apace. Microsoft Research’s MyLifeBits project gives a glimpse of archiving of pe rsonal information that may become commonplace in the near future. MyLifeBits was an experiment where an individual’s interactions—phone calls, emails, documents were captured electronically and stored for later access. The data gathered included a photo taken every minute, which resulted in an overall data volume of one gigabyte a month. When storage costs come down enough to make it feasible to store continuous audio and video, the data volume for a future MyLifeBits service will be many times that.

英文翻译与英文原文.陈--

翻译文献:INVESTIGATION ON DYNAMIC PERFORMANCE OF SLIDE UNIT IN MODULAR MACHINE TOOL (对组合机床滑台动态性能的调查报告) 文献作者:Peter Dransfield, 出处:Peter Dransfield, Hydraulic Control System-Design and Analysis of TheirDynamics, Springer-Verlag, 1981 翻译页数:p139—144 英文译文: 对组合机床滑台动态性能的调查报告 【摘要】这一张纸处理调查利用有束缚力的曲线图和状态空间分析法对组合机床滑台的滑动影响和运动平稳性问题进行分析与研究,从而建立了滑台的液压驱动系统一自调背压调速系统的动态数学模型。通过计算机数字仿真系统,分析了滑台产生滑动影响和运动不平稳的原因及主要影响因素。从那些中可以得出那样的结论,如果能合理地设计液压缸和自调背压调压阀的结构尺寸. 本文中所使用的符号如下: s1-流源,即调速阀出口流量; S el—滑台滑动摩擦力 R一滑台等效粘性摩擦系数: I1—滑台与油缸的质量 12—自调背压阀阀心质量 C1、c2—油缸无杆腔及有杆腔的液容; C2—自调背压阀弹簧柔度; R1, R2自调背压阀阻尼孔液阻, R9—自调背压阀阀口液阻 S e2—自调背压阀弹簧的初始预紧力; I4, I5—管路的等效液感 C5、C6—管路的等效液容: R5, R7-管路的等效液阻; V3, V4—油缸无杆腔及有杆腔内容积; P3, P4—油缸无杆腔及有杆腔的压力 F—滑台承受负载, V—滑台运动速度。本文采用功率键合图和状态空间分折法建立系统的运动数学模型,滑台的动态特性可以能得到显著改善。

论文《人工智能》---文献检索结课作业

人工智能 【摘要】:人工智能是一门极富挑战性的科学,但也是一门边沿学科。它属于自然科学和社会科学的交叉。涉及的学科主要有哲学、认知科学、数学、神经生理学、心理学、计算机科学、信息论、控制论、不定性论、仿生学等。人工智能(Artificial Intelligence),英文缩写为AI。它是研究、开发用于模拟、延伸和扩展人的智能的理论、方法、技术及应用系统的一门新的技术科学。人工智能是计算机科学的一个分支,它企图了解智能的实质,并生产出一种新的能以人类智能相似的方式做出反应的智能机器,该领域的研究包括机器人、语言识别、图像识别、自然语言处理和专家系统等1。 【关键词】:人工智能;应用领域;发展方向;人工检索。 1.人工智能描述 人工智能(Artificial Intelligence) ,英文缩写为AI。它是研究、开发用于模拟、延伸和扩展人的智能的理论、方法、技术及应用系统的一门新的技术科学2。人工智能是计 算机科学的一个分支,它企图了解智 能的实质,并生产出一种新的能以人 类智能相似的方式作出反应的智能 机器,该领域的研究包括机器人、语 言识别、图像识别、自然语言处理和 专家系统等。“人工智能”一词最初 是在1956 年Dartmouth学会上提出 的。从那以后,研究者们发展了众多 理论和原理,人工智能的概念也随之扩展。人工智能是一门极富挑战性的科学,从事这项工作的人必须懂得计算机知识,心理学和哲学。人工智能是包括十分广泛的科学,它由不同的领域组成,如机器学习,计算机视觉等等,总的说来,人工智能研究的一个主要目标是使机器能够胜任一些通常需要人类智能才能完成的复杂工作。但不同的时代、不同的人对这种“复杂工作”的理解是不同的。例如繁重的科学和工程计算本来是要人脑来承担的,现在计算机不但能完成这种计算, 而且能够比人脑做得更快、更准确,因之当代人已不再把这种计算看作是“需要人类智能才能完成的复 1.蔡自兴,徐光祐.人工智能及其应用.北京:清华大学出版社,2010 2元慧·议当人工智能的应用领域与发展状态〖J〗.2008

污水处理外文翻译(带原文)

提高塔式复合人工湿地处理农村生活污水的 脱氮效率1 摘要: 努力保护水源,尤其是在乡镇地区的饮用水源,是中国污水处理当前面临的主要问题。氮元素在水体富营养化和对水生物的潜在毒害方面的重要作用,目前废水脱氮已成为首要关注的焦点。人工湿地作为一种小型的,处理费用较低的方法被用于处理乡镇生活污水。比起活性炭在脱氮方面显示出的广阔前景,人工湿地系统由于溶解氧的缺乏而在脱氮方面存在一定的制约。为了提高脱氮效率,一种新型三阶段塔式混合湿地结构----人工湿地(thcw)应运而生。它的第一部分和第三部分是水平流矩形湿地结构,第二部分分三层,呈圆形,呈紊流状态。塔式结构中水流由顶层进入第二层及底层,形成瀑布溢流,因此水中溶解氧浓度增加,从而提高了硝化反应效率,反硝化效率也由于有另外的有机物的加入而得到了改善,增加反硝化速率的另一个原因是直接通过旁路进入第二部分的废水中带入的足量有机物。常绿植物池柏(Taxodium ascendens),经济作物蔺草(Schoenoplectus trigueter),野茭白(Zizania aquatica),有装饰性的多花植物睡莲(Nymphaea tetragona),香蒲(Typha angustifolia)被种植在湿地中。该系统对总悬浮物、化学需氧量、氨氮、总氮和总磷的去除率分别为89%、85%、83%、83% 和64%。高水力负荷和低水力负荷(16 cm/d 和32 cm/d)对于塔式复合人工湿地结构的性能没有显著的影响。通过硝化活性和硝化速率的测定,发现硝化和反硝化是湿地脱氮的主要机理。塔式复合人工湿地结构同样具有观赏的价值。 关键词: 人工湿地;硝化作用;反硝化作用;生活污水;脱氮;硝化细菌;反硝化细菌 1. 前言 对于提高水源水质的广泛需求,尤其是提高饮用水水源水质的需求是目前废水深度处理的技术发展指向。在中国的乡镇地区,生活污水是直接排入湖泊、河流、土壤、海洋等水源中。这些缺乏处理的污水排放对于很多水库、湖泊不能达到水质标准是有责任的。许多位于中国的乡镇地区的社区缺乏足够的生活污水处理设备。由于山区地形、人口分散、经济基础差等原因,废水的收集和处理是很成问题的。由于资源短缺,经济欠发达地区所采取的废水处理技术必须低价高效,并且要便于施用,能量输入及维护费用较低,而且要保证出水能达标。建造在城市中基于活性污泥床的废水集中处理厂,对于小乡镇缺乏经济适用性,主要是由于污水收集结构的建造费用高。 1Ecological Engineering,Fen xia ,Ying Li。

外文翻译---超越台式机:一个关于云计算的介绍

附录A 译文 超越台式机:一个关于云计算的介绍 在这个世界上几乎每一天都能看到新技术趋势的崛起和衰退,一种新的趋势带来了更长久的生命力。这种趋势被称为云计算,它将改变你使用电脑和互联网的方式。 云计算预示着我们在如何存储信息和运行应用程序问题上的重大变革。而不是在单独的计算机上运行程序的散的数据,一切都托管在“云”中,通过英特网去访问“云”中的其他计算机和服务器。云计算让你能够在世界上的任何地方访问你所有的应用程序和文档,摆脱了台式机的束缚,让你更容易在不同的地方和同事进行协作。 第一部分认识云计算 云计算的出现就相当于一个世纪前电力革命的发生。在电力公司出现之前,每一个农场和企业都通过自己独立的发电机发电。在电网产生以后,农场和企业关闭他们自己的的发电机,并且从电力公司以一个比他们用自己发电机发电低得多的价格(并且更具有可靠性)买电。 为了寻找相同类型革命的发生,使得云计算成为定局。由于我们所期望的教育普及,每时每刻的可靠性和无处不在的云计算协作承诺使得以台式机为计算中心的观念半途而废。云计算将是未来发展的方向。 云计算:它是什么和它的不同 通过传统的桌面计算,你可以在你的每一台电脑上运行程序的副本,你所创建的文件都保存在你创建文件的计算机上。虽然它们能够被这个网络上的其他计算机访问,但是它们不能被外网的计算机访问。 整个的场景都是以个人计算机为中心。 通过云计算,你使用的软件程序不需要运行在你的个人电脑上,但是却可以通过访问英特网上的服务器使用这个软件程序。假如你的电脑死机了,但是这个

软件程序仍然可以供其他人使用。这种模式仍然适用于你创建的文件,它们被储存在一个可以通过互联网访问的集合中。经过许可的任何人不仅可以访问文件,还可以实时地编辑和处理这些文件。不同于传统的计算,云计算模式不以个人计算机为中心,而是以文档为中心,个人计算机使用简单的文件访问并不重要。 但是这只是一个简化。让我们来看看更详细的云计算是什么样子的?它和我们了解云计算的不同同样重要。 云计算的不同 首先,云计算不是网络计算。通过网络计算,应用程序和文档被托管在一个公司的服务器中并且通过公司的网络可以访问那些应用程序和文档。云计算要比网络计算大很多,它包含了许多个公司,许多个服务器和许多个网络。此外,与网络计算不同,云服务和储存通过网络在世界的任何地方都可以访问,但是通过网络计算,只能在自己公司的网络访问到。 云计算也不是传统的像一个公司将计算服务外包(分包)给一个外部公司。虽然外包公司可以托管一个公司的数据或者应用程序,但是这些程序和计划只有这个公司的员工通过公司的网络才能访问到,而不是通过整个英特网。 因此,尽管表面上很相似,但是网络计算和外包不是云计算。 什么是云计算 云计算定义的关键是“云”本身。对于我们而言,云是一大群互联的计算机。这些计算机可以是个人电脑或者网络服务器,它们可以是公共或者私人的。 例如,谷歌的云是由小的个人电脑和更大的服务器组成的。谷歌云是私人的(即,谷歌拥有它),是公开访问的(通过谷歌用户)。 这种计算机云继承并且超越了单一的公司或者企业。跨企业的跨平台的广大用户群可以访问云服务所提供的应用程序和数据。通过互联网访问,任何被授权的用户可以访问连接在英特网上的任何计算机上的这些文档和应用程序。并且,对于用户而言,云技术和云背后的基础设施是不可见的。 云技术是否基于HTTP,HTML协议,XML,JavaScript或者其他特定的技术是不明显的。(在大多数情况下是无所谓的)。 云计算是用户为中心的。一旦你作为一个用户连接到云,无论储存在哪里的文件,信息,图像,应用程序,无论什么都将变成你的。此外,你的数据也可以

外文翻译原文

204/JOURNAL OF BRIDGE ENGINEERING/AUGUST1999

JOURNAL OF BRIDGE ENGINEERING /AUGUST 1999/205 ends.The stress state in each cylindrical strip was determined from the total potential energy of a nonlinear arch model using the Rayleigh-Ritz method. It was emphasized that the membrane stresses in the com-pression region of the curved models were less than those predicted by linear theory and that there was an accompanying increase in ?ange resultant force.The maximum web bending stress was shown to occur at 0.20h from the compression ?ange for the simple support stiffness condition and 0.24h for the ?xed condition,where h is the height of the analytical panel.It was noted that 0.20h would be the optimum position for longitudinal stiffeners in curved girders,which is the same as for straight girders based on stability requirements.From the ?xed condition cases it was determined that there was no signi?cant change in the membrane stresses (from free to ?xed)but that there was a signi?cant effect on the web bend-ing stresses.Numerical results were generated for the reduc-tion in effective moment required to produce initial yield in the ?anges based on curvature and web slenderness for a panel aspect ratio of 1.0and a web-to-?ange area ratio of 2.0.From the results,a maximum reduction of about 13%was noted for a /R =0.167and about 8%for a /R =0.10(h /t w =150),both of which would correspond to extreme curvature,where a is the length of the analytical panel (modeling the distance be-tween transverse stiffeners)and R is the radius of curvature.To apply the parametric results to developing design criteria for practical curved girders,the de?ections and web bending stresses that would occur for girders with a curvature corre-sponding to the initial imperfection out-of-?atness limit of D /120was used.It was noted that,for a panel with an aspect ratio of 1.0,this would correspond to a curvature of a /R =0.067.The values of moment reduction using this approach were compared with those presented by Basler (Basler and Thurlimann 1961;Vincent 1969).Numerical results based on this limit were generated,and the following web-slenderness requirement was derived: 2 D 36,500a a =1?8.6?34 (1) ? ??? t R R F w ?y where D =unsupported distance between ?anges;and F y =yield stress in psi. An extension of this work was published a year later,when Culver et al.(1973)checked the accuracy of the isolated elas-tically supported cylindrical strips by treating the panel as a unit two-way shell rather than as individual strips.The ?ange/web boundaries were modeled as ?xed,and the boundaries at the transverse stiffeners were modeled as ?xed and simple.Longitudinal stiffeners were modeled with moments of inertias as multiples of the AASHO (Standard 1969)values for straight https://www.360docs.net/doc/1d229508.html,ing analytical results obtained for the slenderness required to limit the plate bending stresses in the curved panel to those of a ?at panel with the maximum allowed out-of-?atness (a /R =0.067)and with D /t w =330,the following equa-tion was developed for curved plate girder web slenderness with one longitudinal stiffener: D 46,000a a =1?2.9 ?2.2 (2) ? ? ? t R f R w ?b where the calculated bending stress,f b ,is in psi.It was further concluded that if longitudinal stiffeners are located in both the tension and compression regions,the reduction in D /t w will not be required.For the case of two stiffeners,web bending in both regions is reduced and the web slenderness could be de-signed as a straight girder panel.Eq.(1)is currently used in the ‘‘Load Factor Design’’portion of the Guide Speci?cations ,and (2)is used in the ‘‘Allowable Stress Design’’portion for girders stiffened with one longitudinal stiffener.This work was continued by Mariani et al.(1973),where the optimum trans-verse stiffener rigidity was determined analytically. During almost the same time,Abdel-Sayed (1973)studied the prebuckling and elastic buckling behavior of curved web panels and proposed approximate conservative equations for estimating the critical load under pure normal loading (stress),pure shear,and combined normal and shear loading.The linear theory of shells was used.The panel was simply supported along all four edges with no torsional rigidity of the ?anges provided.The transverse stiffeners were therefore assumed to be rigid in their directions (no strains could be developed along the edges of the panels).The Galerkin method was used to solve the governing differential equations,and minimum eigenvalues of the critical load were calculated and presented for a wide range of loading conditions (bedding,shear,and combined),aspect ratios,and curvatures.For all cases,it was demonstrated that the critical load is higher for curved panels over the comparable ?at panel and increases with an increase in curvature. In 1980,Daniels et al.summarized the Lehigh University ?ve-year experimental research program on the fatigue behav-ior of horizontally curved bridges and concluded that the slen-derness limits suggested by Culver were too severe.Equations for ‘‘Load Factor Design’’and for ‘‘Allowable Stress Design’’were developed (respectively)as D 36,500a =1?4?192(3)? ?t R F w ?y D 23,000a =1?4 ?170 (4) ? ? t R f w ?b The latter equation is currently used in the ‘‘Allowable Stress Design’’portion of the Guide Speci?cations for girders not stiffened longitudinally. Numerous analytical and experimental works on the subject have also been published by Japanese researchers since the end of the CURT project.Mikami and colleagues presented work in Japanese journals (Mikami et al.1980;Mikami and Furunishi 1981)and later in the ASCE Journal of Engineering Mechanics (Mikami and Furunishi 1984)on the nonlinear be-havior of cylindrical web panels under bending and combined bending and shear.They analyzed the cylindrical panels based on Washizu’s (1975)nonlinear theory of shells.The governing nonlinear differential equations were solved numerically by the ?nite-difference method.Simple support boundary condi-tions were assumed along the curved boundaries (top and bot-tom at the ?ange locations)and both simple and ?xed support conditions were used at the straight (vertical)boundaries.The large displacement behavior was demonstrated by Mi-kami and Furunishi for a range of geometric properties.Nu-merical values of the load,de?ection,membrane stress,bend-ing stress,and torsional stress were obtained,but no equations for design use were presented.Signi?cant conclusions include that:(1)the compressive membrane stress in the circumfer-ential direction decreases with an increase in curvature;(2)the panel under combined bending and shear exhibits a lower level of the circumferential membrane stress as compared with the panel under pure bending,and as a result,the bending moment carried by the web panel is reduced;and (3)the plate bending stress under combined bending and shear is larger than that under pure bending.No formulations or recommendations for direct design use were made. Kuranishi and Hiwatashi (1981,1983)used the ?nite-ele-ment method to demonstrate the elastic ?nite displacement be-havior of curved I-girder webs under bending using models with and without ?ange rigidities.Rotation was not allowed (?xed condition)about the vertical axis at the ends of the panel (transverse stiffener locations).Again,the nonlinear distribu-

相关文档
最新文档