数字图像处理与边缘检测中英文对照外文翻译文献

数字图像处理与边缘检测中英文对照外文翻译文献
数字图像处理与边缘检测中英文对照外文翻译文献

中英文资料对照外文翻译

Digital Image Processing and Edge Detection

Digital Image Processing

Interest in digital image processing methods stems from two principal applica- tion areas: improvement of pictorial information for human interpretation; and processing of image data for storage, transmission, and representation for au- tonomous machine perception.

An image may be defined as a two-dimensional function, f(x, y), where x and y are spatial (plane) coordinates, and the amplitude of f at any pair of coordinates (x, y) is called the intensity or gray level of the image at that point. When x, y, and the amplitude values of f are all finite, discrete quantities, we call the image a digital image. The field of digital image processing refers to processing digital images by means of a digital computer. Note that a digital image is composed of a finite number of elements, each of which has a particular location and value. These elements are referred to as picture elements, image elements, pels, and pixels. Pixel is the term most widely used to denote the elements of a digital image.

Vision is the most advanced of our senses, so it is not surprising that images play the single most important role in human perception. However, unlike humans, who are limited to the visual band of the electromagnetic (EM) spec- trum, imaging machines cover almost the entire EM spectrum, ranging from gamma to radio waves. They can operate on images generated by sources that humans are not accustomed to associating with images. These include ultra- sound, electron microscopy, and computer-generated images. Thus, digital image processing encompasses a wide and varied field of applications.

There is no general agreement among authors regarding where image processing stops and other related areas, such as image analysis and computer vi- sion, start. Sometimes a distinction is made by defining image processing as a discipline in which both the input and output of a process are images. We believe this to be a limiting and somewhat artificial boundary. For example, under this definition, even the trivial task of computing the average intensity of an image (which yields a single number) would not be considered an image processing operation. On the other hand, there are fields such as computer vision whose ultimate goal is to use computers to emulate human vision, including learning and being able to make inferences and take actions based on visual inputs. This area itself is a branch of artificial intelligence (AI) whose objective is to emulate human intelligence. The field of AI is in its earliest stages of infancy in terms of development, with progress having been much slower than

originally anticipated. The area of image analysis (also called image understanding) is in be- tween image processing and computer vision.

There are no clearcut boundaries in the continuum from image processing at one end to computer vision at the other. However, one useful paradigm is to consider three types of computerized processes in this continuum: low-, mid-, and highlevel processes. Low-level processes involve primitive opera- tions such as image preprocessing to reduce noise, contrast enhancement, and image sharpening. A low-level process is characterized by the fact that both its inputs and outputs are images. Mid-level processing on images involves tasks such as segmentation (partitioning an image into regions or objects), description of those objects to reduce them to a form suitable for computer processing, and classification (recognition) of individual objects. A midlevel process is characterized by the fact that its inputs generally are images, but its outputs are attributes extracted from those images (e.g., edges, contours, and the identity of individual objects). Finally, higherlevel processing involves “makin g sense”of an ensemble of recognized objects, as in image analysis, and, at the far end of the continuum, performing the cognitive functions normally associated with vision.

Based on the preceding comments, we see that a logical place of overlap between image processing and image analysis is the area of recognition of individual regions or objects in an image. Thus, what we call in this book digital image processing encompasses processes whose inputs and outputs are images and, in addition, encompasses processes that extract attributes from images, up to and including the recognition of individual objects. As a simple illustration to clarify these concepts, consider the area of automated analysis of text. The processes of acquiring an image of the area containing the text, preprocessing that image, extracting (segmenting) the individual characters, describing the characters in a form suitable for computer processing, and recognizing those individual characters are in the scope of what we call digital image processing in this book. Making sense of the content of the page may be viewed as being in the domain of image analysis and even computer vision, depending on the level of complexity implied by the statement “making sense.”As will become evident shortly, digital image processing, as we have defined it, is used successfully in a broad range of areas of exceptional social and economic value.

The areas of application of digital image processing are so varied that some form of organization is desirable in attempting to capture the breadth of this field. One of the simplest ways to develop a basic understanding of the extent of image processing applications is to categorize images according to their source (e.g., visual, X-ray, and so on). The principal energy source for images in use today is the electromagnetic energy spectrum. Other important sources of energy include acoustic, ultrasonic, and electronic (in the form of electron beams used in electron microscopy). Synthetic images, used for modeling and visualization, are generated by computer. In this section we discuss briefly how images are generated in these various categories and the areas in which they are applied.

Images based on radiation from the EM spectrum are the most familiar, es- pecially images in the X-ray and visual bands of the spectrum. Electromagnet- ic waves can be conceptualized as propagating sinusoidal waves of varying

wavelengths, or they can be thought of as a stream of massless particles, each traveling in a wavelike pattern and moving at the speed of light. Each massless particle contains a certain amount (or bundle) of energy. Each bundle of energy is called a photon. If spectral bands are grouped according to energy per photon, we obtain the spectrum shown in fig. below, ranging from gamma rays (highest energy) at one end to radio waves (lowest energy) at the other. The bands are shown shaded to convey the fact that bands of the EM spectrum are not distinct but rather transition smoothly from one to the other.

Image acquisition is the first process. Note that acquisition could be as simple as being given an image that is already in digital form. Generally, the image acquisition stage involves preprocessing, such as scaling.

Image enhancement is among the simplest and most appealing areas of digital image processing. Basically, the idea behind enhancement techniques is to bring out detail that is obscured, or simply to highlight certain features of interest in an image. A familiar example of enhancement is when we increase the contrast of an image because “it looks better.” It is important to keep in mind that enhancement is a very subjective area of image processing. Image restoration is an area that also deals with improving the appearance of an image. However, unlike enhancement, which is subjective, image restoration is objective, in the sense that restoration techniques tend to be based on mathematical or probabilistic models of image degradation. Enhancement, on the other hand, is based on human subjective preferences regarding what constitutes a “good”enhancement result.

Color image processing is an area that has been gaining in importance because of the significant increase in the use of digital images over the Internet. It covers a number of fundamental concepts in color models and basic color processing in a digital domain. Color is used also in later chapters as the basis for extracting features of interest in an image.

Wavelets are the foundation for representing images in various degrees of resolution. In particular, this material is used in this book for image data compression and for pyramidal representation, in which images are subdivided successively into smaller regions.

Compression, as the name implies, deals with techniques for reducing the storage required to save an image, or the bandwidth required to transmi it.Although storage technology has improved significantly over the past decade, the same cannot be said for transmission capacity. This is true particularly in uses of the Internet, which are characterized by significant pictorial content. Image compression is familiar (perhaps inadvertently) to most users of computers in the form of image file extensions, such as the jpg file extension used in the JPEG (Joint Photographic Experts Group) image compression standard.

Morphological processing deals with tools for extracting image components that are useful in the representation and description of shape. The material in this chapter begins a transition from processes that output images to processes that output image attributes.

Segmentation procedures partition an image into its constituent parts or objects. In general, autonomous segmentation is one of the most difficult tasks in digital image processing. A rugged segmentation procedure brings the process a long way toward successful solution of imaging problems that require objects to be identified individually. On the other hand, weak or erratic segmentation algorithms almost always guarantee eventual failure. In general, the more accurate the segmentation, the more likely recognition is to succeed.

Representation and description almost always follow the output of a segmentation stage, which usually is raw pixel data, constituting either the bound- ary of a region (i.e., the set of pixels separating one image region from another) or all the points in the region itself. In either case, converting the data to a form suitable for computer processing is necessary. The first decision that must be made is whether the data should be represented as a boundary or as a complete region. Boundary representation is appropriate when the focus is on external shape characteristics, such as corners and inflections. Regional representation is appropriate when the focus is on internal properties, such as texture or skeletal shape. In some applications, these representations complement each other. Choosing a representation is

only part of the solution for trans- forming raw data into a form suitable for subsequent computer processing. A method must also be specified for describing the data so that features of interest are highlighted. Description, also called feature selection, deals with extracting attributes that result in some quantitative information of interest or are basic for differentiating one class of objects from another.

Recognition is the process that assigns a label (e.g., “vehicle”) to an object based on its descriptors. As detailed before, we conclude our coverage of digital image processing with the development of methods for recognition of individual objects.

So far we have said nothing about the need for prior knowledge or about the interaction between the knowledge base and the processing modules in Fig2 above. Knowledge about a problem domain is coded into an image processing system in the form of a knowledge database. This knowledge may be as sim- ple as detailing regions of an image where the information of interest is known to be located, thus limiting the search that has to be conducted in seeking that information. The knowledge base also can be quite complex, such as an interrelated list of all major possible defects in a materials inspection problem or an image database containing high-resolution satellite images of a region in con- nection with change-detection applications. In addition to guiding the operation of each processing module, the knowledge base also controls the interaction between modules. This distinction is made in Fig2 above by the use of double-headed arrows between the processing modules and the knowledge base, as op- posed to single-headed arrows linking the processing modules.

Edge detection

Edge detection is a terminology in image processing and computer vision, particularly in the areas of feature detection and feature extraction, to refer to algorithms which aim at identifying points in a digital image at which the image brightness changes sharply or more formally has discontinuities.Although point and line detection certainly are important in any discussion on segmentation,edge dectection is by far the most common approach for detecting meaningful discounties in gray level.

Although certain literature has considered the detection of ideal step edges, the edges obtained from natural images are usually not at all ideal step edges. Instead they are normally affected by one or several of the following effects:1.focal blur caused by a finite depth-of-field and finite point spread function; 2.penumbral blur caused by shadows created by light sources of non-zero radius; 3.shading at a smooth object edge; 4.local specularities or interreflections in the vicinity of object edges.

A typical edge might for instance be the border between a block of red color and a block of yellow. In contrast a line (as can be extracted by a ridge detector) can be a small number of pixels of a different color on an otherwise unchanging background. For a line, there may therefore usually be one edge on each side of the line.

To illustrate why edge detection is not a trivial task, let us consider the problem of detecting edges in the following one-dimensional signal. Here, we may intuitively say that there should be an edge between the 4th and 5th pixels.

5 7

6 4 152 148 149

If the intensity difference were smaller between the 4th and the 5th pixels and if the intensity differences between the adjacent neighbouring pixels were higher, it would not be as easy to say that there should be an edge in the corresponding region. Moreover, one could argue that this case is one in which there are several edges.Hence, to firmly state a specific threshold on how large the intensity change between two neighbouring pixels must be for us to say that there should be an edge between these pixels is not always a simple problem. Indeed, this is one of the reasons why edge detection may be a non-trivial problem unless the objects in the scene are particularly simple and the illumination conditions can be well controlled.

There are many methods for edge detection, but most of them can be grouped into two categories,search-based and zero-crossing based. The search-based methods detect edges by first computing a measure of edge strength, usually a first-order derivative expression such as the gradient magnitude, and then searching for local directional maxima of the gradient magnitude using a computed estimate of the local orientation of the edge, usually the gradient direction. The zero-crossing based methods search for zero crossings in a second-order derivative expression computed from the image in order to find edges, usually the zero-crossings of the Laplacian or the zero-crossings of a non-linear differential expression, as will be described in the section on differential edge detection following below. As a pre-processing step to edge detection, a smoothing stage, typically Gaussian smoothing, is almost always applied (see also noise reduction).

The edge detection methods that have been published mainly differ in the types of smoothing filters that are applied and the way the measures of edge strength are computed. As many edge detection methods rely on the computation of image gradients, they also differ in the types of filters used for computing gradient estimates in the x- and y-directions.

Once we have computed a measure of edge strength (typically the gradient magnitude), the next stage is to apply a threshold, to decide whether edges are present or not at an image point. The lower the threshold, the more edges will be detected, and the result will be increasingly susceptible to noise, and also to picking out irrelevant features from the image. Conversely a high threshold may miss subtle edges, or result in fragmented edges.

If the edge thresholding is applied to just the gradient magnitude image, the resulting edges will in general be thick and some type of edge thinning post-processing is necessary. For edges detected with non-maximum suppression however, the edge curves are thin by definition and the edge pixels can be linked into edge polygon by an edge linking (edge tracking) procedure. On a discrete grid, the non-maximum suppression stage can be implemented by estimating the gradient direction using first-order derivatives, then rounding off the gradient direction to multiples of 45 degrees, and finally comparing the values of the gradient magnitude in the estimated gradient direction.

A commonly used approach to handle the problem of appropriate thresholds for

thresholding is by using thresholding with hysteresis. This method uses multiple thresholds to find edges. We begin by using the upper threshold to find the start of an edge. Once we have a start point, we then trace the path of the edge through the image pixel by pixel, marking an edge whenever we are above the lower threshold. We stop marking our edge only when the value falls below our lower threshold. This approach makes the assumption that edges are likely to be in continuous curves, and allows us to follow a faint section of an edge we have previously seen, without meaning that every noisy pixel in the image is marked down as an edge. Still, however, we have the problem of choosing appropriate thresholding parameters, and suitable thresholding values may vary over the image.

Some edge-detection operators are instead based upon second-order derivatives of the intensity. This essentially captures the rate of change in the intensity gradient. Thus, in the ideal continuous case, detection of zero-crossings in the second derivative captures local maxima in the gradient.

We can come to a conclusion that,to be classified as a meaningful edge point,the transition in gray level associated with that point has to be significantly stronger than the background at that point.Since we are dealing with local computations,the method of choice to determine whether a value is “significant” or not id to use a threshold.Thus we define a point in an image as being as being an edge point if its two-dimensional first-order derivative is greater than a specified criterion of connectedness is by definition an edge.The term edge segment generally is used if the edge is short in relation to the dimensions of the image.A key problem in segmentation is to assemble edge segments into longer edges.An alternate definition if we elect to use the second-derivative is simply to define the edge ponits in an image as the zero crossings of its second derivative.The definition of an edge in this case is the same as above.It is important to note that these definitions do not guarantee success in finding edge in an image.They simply give us a formalism to look for them.First-order derivatives in an image are computed using the gradient.Second-order derivatives are obtained using the Laplacian.

中文对照

数字图像处理与边缘检测

数字图像处理

数字图像处理方法的研究源于两个主要应用领域:其一是为了便于人们分析而对图像信息进行改进:其二是为使机器自动理解而对图像数据进行存储、传输及显示。

一幅图像可定义为一个二维函数f(x,y),这里x和y是空间坐标,而在任何一对空间坐标(x,y)上的幅值f 称为该点图像的强度或灰度。当x,y和幅值f 为有限的、离散的数值时,称该图像为数字图像。数字图像处理是指借用数字计算机处理数字图像,值得提及的是数字图像是由有限的元素组成的,每一个元素都有一个特定的位置和幅值,这些元素称为图像元素、画面元素或像素。像素是广泛用于表示数字图像元素的词汇。

视觉是人类最高级的感知器官,所以,毫无疑问图像在人类感知中扮演着最重要的角色。然而,人类感知只限于电磁波谱的视觉波段,成像机器则可覆盖几乎全部电磁波谱,从伽马射线到无线电波。它们可以对非人类习惯的那些图像源进行加工,这些图像源包括超声波、电子显微镜及计算机产生的图像。因此,数字图像处理涉及各种各样的应用领域。

图像处理涉及的范畴或其他相关领域(例如,图像分析和计算机视觉)的界定在初创人之间并没有一致的看法。有时用处理的输入和输出内容都是图像这一特点来界定图像处理的范围。我们认为这一定义仅是人为界定和限制。例如,在这个定义下,甚至最普通的计算一幅图像灰度平均值的工作都不能算做是图像处理。另一方面,有些领域(如计算机视觉)研究的最高目标是用计算机去模拟人类视觉,包括理解和推理并根据视觉输入采取行动等。这一领域本身是人工智能

的分支,其目的是模仿人类智能。人工智能领域处在其发展过程中的初期阶段,它的发展比预期的要慢的多,图像分析(也称为图像理解)领域则处在图像处理和计算机视觉两个学科之间。

从图像处理到计算机视觉这个连续的统一体内并没有明确的界线。然而,在这个连续的统一体中可以考虑三种典型的计算处理(即低级、中级和高级处理)来区分其中的各个学科。

低级处理涉及初级操作,如降低噪声的图像预处理,对比度增强和图像尖锐化。低级处理是以输入、输出都是图像为特点的处理。中级处理涉及分割(把图像分为不同区域或目标物)以及缩减对目标物的描述,以使其更适合计算机处理及对不同目标的分类(识别)。中级图像处理是以输入为图像,但输出是从这些图像中提取的特征(如边缘、轮廓及不同物体的标识等)为特点的。最后,高级处理涉及在图像分析中被识别物体的总体理解,以及执行与视觉相关的识别函数(处在连续统一体边缘)等。

根据上述讨论,我们看到,图像处理和图像分析两个领域合乎逻辑的重叠区域是图像中特定区域或物体的识别这一领域。这样,在研究中,我们界定数字图像处理包括输入和输出均是图像的处理,同时也包括从图像中提取特征及识别特定物体的处理。举一个简单的文本自动分析方面的例子来具体说明这一概念。在自动分析文本时首先获取一幅包含文本的图像,对该图像进行预处理,提取(分割)字符,然后以适合计算机处理的形式描述这些字符,最后识别这些字符,而所有这些操作都在本文界定的数字图像处理的范围内。理解一页的内容可能要根据理解的复杂度从图像分析或计算机视觉领域考虑问题。这样,我们定义的数字图像处理的概念将在有特殊社会和经济价值的领域内通用。

数字图像处理的应用领域多种多样,所以文本在内容组织上尽量达到该技术应用领域的广度。阐述数字图像处理应用范围最简单的一种方法是根据信息源来分类(如可见光、X射线,等等)。在今天的应用中,最主要的图像源是电磁能谱,其他主要的能源包括声波、超声波和电子(以用于电子显微镜方法的电子束形式)。建模和可视化应用中的合成图像由计算机产生。

建立在电磁波谱辐射基础上的图像是最熟悉的,特别是X射线和可见光谱图像。电磁波可定义为以各种波长传播的正弦波,或者认为是一种粒子流,每个粒子包含一定(一束)能量,每束能量成为一个光子。如果光谱波段根据光谱能量进行分组,我们会得到下图1所示的伽马射线(最高能量)到无线电波(最低能量)的光谱。如图所示的加底纹的条带表达了这样一个事实,即电磁波谱的各波段间并没有明确的界线,而是由一个波段平滑地过渡到另一个波段。

图像获取是第一步处理。注意到获取与给出一幅数字形式的图像一样简单。通常,图像获取包括如设置比例尺等预处理。

图像增强是数字图像处理最简单和最有吸引力的领域。基本上,增强技术后面的思路是显现那些被模糊了的细节,或简单地突出一幅图像中感兴趣的特征。一个图像增强的例子是增强图像的对比度,使其看起来好一些。应记住,增强是图像处理中非常主观的领域,这一点很重要。

图像复原也是改进图像外貌的一个处理领域。然而,不像增强,图像增强是主观的,而图像复原是客观的。在某种意义上说,复原技术倾向于以图像退化的数学或概率模型为基础。另一方面,增强以怎样构成好的增强效果这种人的主观偏爱为基础。

彩色图像处理已经成为一个重要领域,因为基于互联网的图像处理应用在不断增长。就使得在彩色模型、数字域的彩色处理方面涵盖了大量基本概念。在后续发展,彩色还是图像中感兴趣特征被提取的基础。

小波是在各种分辨率下描述图像的基础。特别是在应用中,这些理论被用于图像数据压缩及金字塔描述方法。在这里,图像被成功地细分为较小的区域。

压缩,正如其名称所指的意思,所涉及的技术是减少图像的存储量,或者在传输图像时降低频带。虽然存储技术在过去的十年内有了很大改进,但对传输能力我们还不能这样说,尤其在互联网上更是如此,互联网是以大量的图片内容为特征的。图像压缩技术对应的图像文件扩展名对大多数计算机用户是很熟悉的(也许没注意),如JPG文件扩展名用于JPEG(联合图片专家组)图像压缩标准。

形态学处理设计提取图像元素的工具,它在表现和描述形状方面非常有用。

红外数据通信技术外文翻译文献

红外数据通信技术外文翻译文献(文档含中英文对照即英文原文和中文翻译) Infrared Remote Control System Abstract Red outside data correspondence the technique be currently within the scope of world drive extensive usage of a kind of wireless conjunction technique, drive numerous hardware and software platform support. Red outside the transceiver product have cost low, small scaled turn, the baud rate be quick, point to point SSL, be free from electromagnetism thousand Raos

etc. characteristics, can realization information at dissimilarity of the product fast, convenience, safely exchange and transmission, at short distance wireless deliver aspect to own very obvious of advantage. Along with red outside the data deliver a technique more and more mature, the cost descend, red outside the transceiver necessarily will get at the short distance communication realm more extensive of application. The purpose that design this system is transmit customer’s operation information with infrared rays for transmit media, then demodulate original signal with receive circuit. It use coding chip to modulate signal and use decoding chip to demodulate signal. The coding chip is PT2262 and decoding chip is PT2272. Both chips are made in Taiwan. Main work principle is that we provide to input the information for the PT2262 with coding keyboard. The input information was coded by PT2262 and loading to high frequent load wave whose frequent is 38 kHz, then modulate infrared transmit dioxide and radiate space outside when it attian enough power. The receive circuit receive the signal and demodulate original information. The original signal was decoded by PT2272, so as to drive some circuit to accomplish customer’s operation demand. Keywords: Infrared dray;Code;Decoding;LM386;Red outside transceiver 1 Introduction 1.1 research the background and significance Infrared Data Communication Technology is the world wide use of a wireless connection technology, by the many hardware and software platforms supported. Is a data through electrical pulses and infrared optical pulse switch between the wireless data transceiver technology.

外文翻译---基于模糊逻辑技术图像上边缘检测

译文二: 1基于模糊逻辑技术图像上边缘检测[2] 摘要:模糊技术是经营者为了模拟在数学水平的代偿行为过程的决策或主观评价而引入的。下面介绍经营商已经完成了的计算机视觉应用。本文提出了一种基于模糊逻辑推理战略为基础的新方法,它被建议使用在没有确定阈值的数字图像边缘检测上。这种方法首先将用3?3的浮点二进制矩阵将图像分割成几个区域。边缘像素被映射到一个属性值与彼此不同的范围。该方法的鲁棒性所得到的不同拍摄图像将与线性Sobel运算所得到的图像相比较。并且该方法给出了直线的线条平滑度、平直度和弧形线条的良好弧度这些永久的效果。同时角位可以更清晰并且可以更容易的定义。关键词:模糊逻辑,边缘检测,图像处理,电脑视觉,机械的部位,测量 1.引言 在过去的几十年里,对计算机视觉系统的兴趣,研究和发展已经增长了不少。如今,它们出现在各个生活领域,从停车场,街道和商场各角落的监测系统到主要食品生产的分类和质量控制系统。因此,引进自动化的视觉检测和测量系统是有必要的,特别是二维机械对象[1,8]。部分原因是由于那些每天产生的数字图像大幅度的增加(比如,从X光片到卫星影像),并且对于这样图片的自动处理有增加的需求[9,10,11]。因此,现在的许多应用例如对医学图像进行计算机辅助诊断,将遥感图像分割和分类成土地类别(比如,对麦田,非法大麻种植园的鉴定,以及对作物生长的估计判断),光学字符识别,闭环控制,基于目录检索的多媒体应用,电影产业上的图像处理,汽车车牌的详细记录的鉴定,以及许多工业检测任务(比如,纺织品,钢材,平板玻璃等的缺陷检测)。历史上的许多数据已经被生成图像,以帮助人们分析(相比较于数字表之类的,图像显然容易理解多了)[12]。所以这鼓励了数字分析技术在数据处理方面的使用。此外,由于人类善于理解图像,基于图像的分析法在算法发展上提供了一些帮助(比如,它鼓励几何分析),并且也有助于非正式确认的结果。虽然计算机视觉可以被总结为一个自动(或半自动)分析图像的系统,一些变化也是可能的[9,13]。这些图像可以来自超出正常灰度和色彩的照片,例如红外光,X射线,以及新一代的高光谱 [1]Abdallah A. Alshennawy, A yman A. Aly. Edge Detection in Digital Images Using Fuzzy Logic Technique[]J. World Academy of Science, Engineering and Technology, 2009, 51:178-186.

人工智能专业外文翻译-机器人

译文资料: 机器人 首先我介绍一下机器人产生的背景,机器人技术的发展,它应该说是一个科学技术发展共同的一个综合性的结果,同时,为社会经济发展产生了一个重大影响的一门科学技术,它的发展归功于在第二次世界大战中各国加强了经济的投入,就加强了本国的经济的发展。另一方面它也是生产力发展的需求的必然结果,也是人类自身发展的必然结果,那么随着人类的发展,人们在不断探讨自然过程中,在认识和改造自然过程中,需要能够解放人的一种奴隶。那么这种奴隶就是代替人们去能够从事复杂和繁重的体力劳动,实现人们对不可达世界的认识和改造,这也是人们在科技发展过程中的一个客观需要。 机器人有三个发展阶段,那么也就是说,我们习惯于把机器人分成三类,一种是第一代机器人,那么也叫示教再现型机器人,它是通过一个计算机,来控制一个多自由度的一个机械,通过示教存储程序和信息,工作时把信息读取出来,然后发出指令,这样的话机器人可以重复的根据人当时示教的结果,再现出这种动作,比方说汽车的点焊机器人,它只要把这个点焊的过程示教完以后,它总是重复这样一种工作,它对于外界的环境没有感知,这个力操作力的大小,这个工件存在不存在,焊的好与坏,它并不知道,那么实际上这种从第一代机器人,也就存在它这种缺陷,因此,在20世纪70年代后期,人们开始研究第二代机器人,叫带感觉的机器人,这种带感觉的机器人是类似人在某种功能的感觉,比如说力觉、触觉、滑觉、视觉、听觉和人进行相类比,有了各种各样的感觉,比方说在机器人抓一个物体的时候,它实际上力的大小能感觉出来,它能够通过视觉,能够去感受和识别它的形状、大小、颜色。抓一个鸡蛋,它能通过一个触觉,知道它的力的大小和滑动的情况。第三代机器人,也是我们机器人学中一个理想的所追求的最高级的阶段,叫智能机器人,那么只要告诉它做什么,不用告诉它怎么去做,它就能完成运动,感知思维和人机通讯的这种功能和机能,那么这个目前的发展还是相对的只是在局部有这种智能的概念和含义,但真正完整意义的这种智能机器人实际上并没有存在,而只是随着我们不断的科学技术的发展,智能的概念越来越丰富,它内涵越来越宽。 下面我简单介绍一下我国机器人发展的基本概况。由于我们国家存在很多其

指纹识别系统(文献综述)

指纹识别方法的综述 摘 要: 对在指纹的预处理和特征提取、指纹分类、指纹的匹配过程中的方向图、滤波器、神经网络等关 键性原理和技术做了详细的说明,并对在各个过程中用到的方法做了进一步的比较,讨论了各种方法的优越性。 0 引 言 自动指纹识别是上世纪六十年代兴起的,利用计算机取代人工来进行指纹识别的一种方法。近年 来,随着计算机技术的飞速发展,低价位指纹采集仪的出现以及高可靠算法的实现,更使得自动指纹识 别技术越来越多地进入到人们的生活和工作中,自动指纹识别系统的研究和开发正在成为国内外学术 界和商业界的热点。相对于其他生物特征鉴别技术例如语音识别及虹膜识别,指纹识别具有许多独到 的优点,更重要的是它具有很高的实用性和可行性,已经被认为是一种理想的身份认证技术,有着十分 广泛的应用前景,是将来生物特征识别技术的主流。 1 指纹取像 图 1 是一个自动指纹识别系统AFIS(Automated Fingerprint Identification System) 的简单流程。 → → → ↓ ↑ ———— 将一个人的指纹采集下来输入计算机进行处理是指纹自动识别的首要步骤。指纹图像的获取主要利用设备取像,方便实用,比较适合AFIS 。利用设备取像的主要方法又利用光学设备、晶体传感器和超声波来进行。光学取像设备是根据光的全反射原理来设计的。晶体传感器取像是根据谷线和脊线皮肤与传感器之间距离不同而产生的电容不同来设计的。超声波设备取像也是采用光波来取像,但由于超声波波长较短,抗干扰能力较强,所以成像的质量非常好。 2 图像的预处理与特征提取 无论采取哪种方法提取指纹,总会给指纹图像带来各种噪声。预处理的目的就是去除图像中的噪 音,把它变成一幅清晰的点线图,以便于提取正确的指纹特征。预处理是指纹自动识别过程的第一步, 它的好坏直接影响着指纹识别的效果。常用的预处理与特征提取( Image Preprocessing and Feature Ex 2 t raction) 方法的主要步骤包括方向图计算、图像滤波、二值化、细化、提取特征和后处理。当然这些步骤 可以根据系统和应用的具体情况再进行适当变化。文献[ 1 ]提出了基于脊线跟踪的方法能够指纹取像 图像预处理 特征提取 指纹识别 数据库管理

中英文文献翻译-加工中心数控技术

加工中心数控技术 出处:数控加工中心的分类以及各自特点 出版社:化学工业出版社; 第1版 (2009年3月16日) 作者:徐衡、段晓旭 加工中心是典型的集高技术于一体的机械加工设备,它的发展代表了一个国家设计制造的水平也大大提高了劳动生产率,降低了劳动成本,改善了工人的工作环境,降低了工人的劳动强度。本文经过对不同运动方案和各部件的设计方案的定性分析比较确定该教立式加工中心的进给传动方案为:采用固定床身,电主轴通过安装座安装在床身导轨的滑座上,床身导轨采用滚动导轨,可以实现Y 方向的进给运动。由X-Y双向精密数控工作台带动工件完成X,Y两个方向的进给运动;X,Y,Z三个方向的进给运动均滚珠丝杠,并由交流伺服电机驱动。导轨、滚珠丝杠有相应的润滑、防护等装置。 加工中心(英文缩写为CNC 全称为Computerized Numerical Control):是带有刀库和自动换刀装置的一种高度自动化的多功能数控机床。在中国香港,台湾及广东一代也有很多人叫它电脑锣。 工件在加工中心上经一次装夹后,数字控制系统能控制机床按不同工序,自动选择和更换刀具,自动改变机床主轴转速、进给量和刀具相对工件的运动轨迹及其他辅助机能,依次完成工件几个面上多工序的加工。并且有多种换刀或选刀功能,从而使生产效率大大提高。 加工中心数控机床是一种装有计算机数字控制系统的机床,数控系统能够处理加工程序,控制机床完成各种动作。与普通机床相比,数控机床能够完成平面曲线和空间曲面的加工,加工精度和生产效率都比较高,因而应用日益广泛。 数控机床的组成 一般来说,数控机床由机械部分、数字控制计算机、伺服系统、PC控制部分、液压气压传动系统、冷却润滑和排泄装置组成。数控机床是由程序控制的,零件的编程工作是数控机床加工的重要组成部分。伺服系统是数控机床的驱动部分,计算机输出的控制命令是通过伺服系统产生坐标移动的。普通的立式加工中心有三个伺服电机,分别驱动纵向工作台、横向工作台、主轴箱沿X向、Y向、Z向运动。X、Y、Z是互相垂直的坐标轴,因而当机床三坐标联动时可以加工空

旅游管理专业论文外文文献翻译

外文资料译文及原文 译文(一) 消费者体验旅游和品牌的结合 米契尔罗伯特 定义消费者体验旅游 制造工厂参观,公司博物馆和公司访客中心表现为被不同名字已知的观光事业片段:制造业观光事业,工业的吸引、工业的观光事业和工业的遗产观光事业。在每一个描述性的长期的共同目标是在消费者学习品牌,其运作,生产过程,历史和历史意义的时候建立一个消费者和品牌之间的纽带。有人建议在这里CET代表一个统一的主题的旅游。这个术语捕捉消费者的消费能力发现更多关于他们所消费的品牌,而制造商可以在与该工厂的客人接触的30-120分钟时间里建立与这些消费者更密切的关系。 参与的品牌 品牌经理寻求解决在三个层次消费者的需求: (1)功能(对消费者提供解决问题的办法); (2)符号(提供心理欲望满意度); (3)经历(提供感官快乐,品种,认知,刺激) CET可以通过视觉地介绍品牌,运作,生产工艺,历史和历史意义加强消费者和品牌之间的纽带。这种纽带可以被看作是个人品牌参与和品牌忠诚度的提高。认知参与反映了消费者对产品的兴趣(或学习更多)。CET可以通过刺激消费者对于品牌和生产过程的想象提高消费者的认知水平。此外,积极口碑沟通刺激满足旅客可能会比其他形式的促销更可信。 缺乏现有的直接研究关注 迄今为止,CET已经在行销文学中受到一点注意。米契尔和米契尔(2001年)对此内容这种的旅游网站进行了评估。此外,这些相同的作者已经评估食物和饮料工业中的现象(米契尔和米契尔,2000年),非营利部门(米契尔和米契尔,2001年b),和整体经济(米契尔等, 2001)。米契尔和米契尔(2002)为学者提出了格式,用来评估在当地的服务领域这些设施的地方利益。该主题通常包括对整合营销的简要讨论,但已收到直接研究的关注很有限。

通信工程项目毕业材料外文翻译

用于多跳认知无线电网络的分布式网络编码控制信道 Alfred Asterjadhi等著 1 前言 大多数电磁频谱由政府机构长期指定给公司或机构专门用于区域或国家地区。由于这种资源的静态分配,许可频谱的许多部分在许多时间和/或位置未使用或未被充分利用。另一方面,几种最近的无线技术在诸如IEEE802.11,蓝牙,Zigbee之类的非许可频段中运行,并且在一定程度上对WiMAX进行操作;这些技术已经看到这样的成功和扩散,他们正在访问的频谱- 主要是2.4 GHz ISM频段- 已经过度拥挤。为了为这些现有技术提供更多的频谱资源,并且允许替代和创新技术的潜在开发,最近已经提出允许被许可的设备(称为次要用户)访问那些许可的频谱资源,主要用户未被使用或零星地使用。这种方法通常被称为动态频谱接入(DSA),无线电设备发现和机会性利用未使用或未充分利用的频谱带的能力通常称为认知无线电(CR)技术。 DSA和CR最近都引起了无线通信和网络界的极大关注。通常设想两种主要应用。第一个是认知无线接入(CW A),根据该认知接入点,认知接入点负责识别未使用的许可频谱,并使用它来提供对次用户的接入。第二个应用是我们在这个技术中研究的应用,它是认知自组织网络(CAN),也就是使用 用于二级用户本身之间通信的无许可频谱,用于诸如点对点内容分发,环境监控,安全性等目的,灾难恢复情景通信,军事通信等等。 设计CAN系统比CW A有更多困难,主要有两个原因。第一是识别未使用的频谱。在CW A中,接入点的作用是连接到互联网,因此可以使用简单的策略来推断频谱可用性,例如查询频谱调节器在其地理位置的频谱可用性或直接与主用户协商频谱可用性或一些中间频谱经纪人另一方面,在CAN中,与频谱调节器或主要用户的缺乏直接通信需要二级用户能够使用检测技术自己识别未使用的频谱。第二个困难是辅助用户协调媒体访问目的。在CW A中存在接入点和通常所有二级用户直接与之通信(即,网络是单跳)的事实使得直接使用集中式媒体接入控制(MAC)解决方案,如时分多址(TDMA)或正交频分多址(OFDMA)。相反,预计CAN将跨越多跳,缺少集中控制器;而对于传统的单通道多跳自组织网络而言,这个问题的几个解决方案是已知的,因为假设我们处理允许设备访问的具有成

毕业论文外文翻译-图像分割

图像分割 前一章的资料使我们所研究的图像处理方法开始发生了转变。从输人输出均为图像的处理方法转变为输人为图像而输出为从这些图像中提取出来的属性的处理方法〔这方面在1.1节中定义过)。图像分割是这一方向的另一主要步骤。 分割将图像细分为构成它的子区域或对象。分割的程度取决于要解决的问题。就是说当感兴趣的对象已经被分离出来时就停止分割。例如,在电子元件的自动检测方面,我们关注的是分析产品的图像,检测是否存在特定的异常状态,比如,缺失的元件或断裂的连接线路。超过识别这此元件所需的分割是没有意义的。 异常图像的分割是图像处理中最困难的任务之一。精确的分割决定着计算分析过程的成败。因此,应该特别的关注分割的稳定性。在某些情况下,比如工业检测应用,至少有可能对环境进行适度控制的检测。有经验的图像处理系统设计师总是将相当大的注意力放在这类可能性上。在其他应用方面,比如自动目标采集,系统设计者无法对环境进行控制。所以,通常的方法是将注意力集中于传感器类型的选择上,这样可以增强获取所关注对象的能力,从而减少图像无关细节的影响。一个很好的例子就是,军方利用红外线图像发现有很强热信号的目标,比如移动中的装备和部队。 图像分割算法一般是基于亮度值的不连续性和相似性两个基本特性之一。第一类性质的应用途径是基于亮度的不连续变化分割图像,比如图像的边缘。第二类的主要应用途径是依据事先制定的准则将图像分割为相似的区域,门限处理、区域生长、区域分离和聚合都是这类方法的实例。 本章中,我们将对刚刚提到的两类特性各讨论一些方法。我们先从适合于检测灰度级的不连续性的方法展开,如点、线和边缘。特别是边缘检测近年来已经成为分割算法的主题。除了边缘检测本身,我们还会讨论一些连接边缘线段和把边缘“组装”为边界的方法。关于边缘检测的讨论将在介绍了各种门限处理技术之后进行。门限处理也是一种人们普遍关注的用于分割处理的基础性方法,特别是在速度因素占重要地位的应用中。关于门限处理的讨论将在几种面向区域的分割方法展开的讨论之后进行。之后,我们将讨论一种称为分水岭分割法的形态学

文献综述_人工智能

人工智能的形成及其发展现状分析 冯海东 (长江大学管理学院荆州434023) 摘要:人工智能的历史并不久远,故将从人工智能的出现、形成、发展现 状及前景几个方面对其进行分析,总结其发展过程中所出现的问题,以及发展现状中的不足之处,分析其今后的发展方向。 关键词:人工智能,发展过程,现状分析,前景。 一.引言 人工智能最早是在1936年被英国的科学家图灵提出,并不为多数人所认知。 当时,他编写了一个下象棋的程序,这就是最早期的人工智能的应用。也有著名的“图灵测试”,这也是最初判断是否是人工智能的方案,因此,图灵被尊称为“人工智能之父”。人工智能从产生到发展经历了一个起伏跌宕的过程,直到目前为止,人工智能的应用技术也不是很成熟,而且存在相当的缺陷。 通过搜集的资料,将详细的介绍人工智能这个领域的具体情况,剖析其面临的挑战和未来的前景。 二.人工智能的发展历程 1. 1956年前的孕育期 (1) 从公元前伟大的哲学家亚里斯多德(Aristotle)到16世纪英国哲学家培根(F. Bacon),他们提出的形式逻辑的三段论、归纳法以及“知识就是力量”的警句,都对人类思维过程的研究产生了重要影响。 (2)17世纪德国数学家莱布尼兹(G..Leibniz)提出了万能符号和推理计算思想,为数理逻辑的产生和发展奠定了基础,播下了现代机器思维设计思想的种子。而19世纪的英国逻辑学家布尔(G. Boole)创立的布尔代数,实现了用符号语言描述人类思维活动的基本推理法则。 (3) 20世纪30年代迅速发展的数学逻辑和关于计算的新思想,使人们在计算机出现之前,就建立了计算与智能关系的概念。被誉为人工智能之父的英国天才的数学家图灵(A. Tur-ing)在1936年提出了一种理想计算机的数学模型,即图灵机之后,1946年就由美国数学家莫克利(J. Mauchly)和埃柯特(J. Echert)研制出了世界上第一台数字计算机,它为人工智能的研究奠定了不可缺少的物质基础。1950年图灵又发表了“计算机与智能”的论文,提出了著名的“图灵测试”,形象地指出什么是人工智能以及机器具有智能的标准,对人工智能的发展产生了极其深远的影响。 (4) 1934年美国神经生理学家麦克洛奇(W. McCulloch) 和匹兹(W. Pitts )建立了第一个神经网络模型,为以后的人工神经网络研究奠定了基础。 2. 1956年至1969年的诞生发育期 (1)1956年夏季,麻省理工学院(MIT)的麦卡锡(J.McCarthy)、明斯基(M. Minshy)、塞尔夫里奇(O. Selfridge)与索罗门夫(R. Solomonff)、 IBM的洛

虹膜识别外文翻译文献

虹膜识别外文翻译文献 虹膜识别外文翻译文献 (文档含中英文对照即英文原文和中文翻译) 外文: The first chapter 1.1 The research background of iris recognition Biometrics is a technology for personal identification using physiological characteristics and behavior characteristics inherent in the human body. Can be used for the biological characteristics of biological recognition, fingerprint, hand type face, iris, retina, pulse, ear etc.. Behavior has the following characteristics: signature, voice, gait, etc.. Based on these characteristics, it has been the development of hand shape recognition, fingerprint recognition, facial recognition, iris recognition, signature recognition and other biometric technology, many techniques have been formed and mature to application of. Biological recognition technology in a , has a long history, the ancient Egyptians through identification of each part of the body size measure to carry out identity may be the earliest human based on the earliest history of biometrics. But the modern biological recognition technology began in twentieth Century 70 time metaphase, as biometric devices early is relatively expensive, so only a higher security level atomic test, production base.due to declining cost of microprocessor and various electronic components, precision gradually improve, control device of a biological recognition technology has been gradually applied to commerce authorized, such as access control, attendance management, management system, safety certification field etc.. All biometric technology, iris recognition is currently used as a convenient and accurate.

机械类数控车床外文翻译外文文献英文文献车床.doc

Lathes Lathes are machine tools designed primarily to do turning, facing and boring, Very little turning is done on other types of machine tools, and none can do it with equal facility. Because lathes also can do drilling and reaming, their versatility permits several operations to be done with a single setup of the work piece. Consequently, more lathes of various types are used in manufacturing than any other machine tool. The essential components of a lathe are the bed, headstock assembly, tailstock assembly, and the leads crew and feed rod. The bed is the backbone of a lathe. It usually is made of well normalized or aged gray or nodular cast iron and provides s heavy, rigid frame on which all the other basic components are mounted. Two sets of parallel, longitudinal ways, inner and outer, are contained on the bed, usually on the upper side. Some makers use an inverted V-shape for all four ways, whereas others utilize one inverted V and one flat way in one or both sets, They are precision-machined to assure accuracy of alignment. On most modern lathes the way are surface-hardened to resist wear and abrasion, but precaution should be taken in operating a lathe to assure that the ways are not damaged. Any inaccuracy in them usually means that the accuracy of the entire lathe is destroyed. The headstock is mounted in a foxed position on the inner ways, usually at the left end of the bed. It provides a powered means of rotating the word at various speeds . Essentially, it consists of a hollow spindle, mounted in accurate bearings, and a set of transmission gears-similar to a truck transmission—through which the spindle can be rotated at a number of speeds. Most lathes provide from 8 to 18 speeds, usually in a geometric ratio, and on modern lathes all the speeds can be obtained merely by moving from two to four levers. An increasing trend is to provide a continuously variable speed range through electrical or mechanical drives. Because the accuracy of a lathe is greatly dependent on the spindle, it is of heavy construction and mounted in heavy bearings, usually preloaded tapered roller or ball types. The spindle has a hole extending through its length, through which long bar stock can be fed. The size of maximum size of bar stock that can be machined when the material must be fed through spindle. The tailsticd assembly consists, essentially, of three parts. A lower casting fits on the inner ways of the bed and can slide longitudinally thereon, with a means for clamping the entire assembly in any desired location, An upper casting fits on the lower one and can be moved transversely upon it, on some type of keyed ways, to permit aligning the assembly is the tailstock quill. This is a hollow steel cylinder, usually about 51 to 76mm(2to 3 inches) in diameter, that can be moved several inches longitudinally in and out of the upper casting by means of a hand wheel and screw. The size of a lathe is designated by two dimensions. The first is known as the swing. This is the maximum diameter of work that can be rotated on a lathe. It is approximately twice the distance between the line connecting the lathe centers and the nearest point on the ways, The second size dimension is the maximum distance between centers. The swing thus indicates the maximum work piece diameter that can be turned in the lathe, while the distance between centers indicates the maximum length of work piece that can be mounted between centers. Engine lathes are the type most frequently used in manufacturing. They are heavy-duty machine tools with all the components described previously and have power drive for all tool movements except on the compound rest. They commonly range in size from 305 to 610 mm(12 to 24 inches)swing and from 610 to 1219 mm(24 to 48 inches) center distances, but swings up to 1270 mm(50 inches) and center distances up

旅游品牌定位外文翻译文献

旅游品牌定位外文翻译文献(文档含英文原文和中文翻译)

原文: Destination brand positions of a competitive set of near-home destinations Abstract: Although the branding literature commenced during the 1940s, the first publications related to destination branding did not emerge until half a century later. A review of 74 destination branding publications by 102 authors from the first 10 years of destination branding literature (1998–2007) found at least nine potential research gaps warranting attention by researchers. In particular, there has been a lack of research examining the extent to which brand positioning campaigns have been successful in enhancing brand equity in the manner intended in the brand identity. The purpose of this paper is to report the results of an investigation of brand equity tracking for a competitive set of destinations in Queensland, Australia between 2003 and 2007. A hierarchy of consumer-based brand equity (CBBE) provided an effective means to monitor destination brand positions over time. A key implication of the results was the finding that there was no change in brand positions for any of the five destinations over the four year period. This leads to the proposition that destination position change within a competitive set will only

5G无线通信网络中英文对照外文翻译文献

5G无线通信网络中英文对照外文翻译文献(文档含英文原文和中文翻译)

翻译: 5G无线通信网络的蜂窝结构和关键技术 摘要 第四代无线通信系统已经或者即将在许多国家部署。然而,随着无线移动设备和服务的激增,仍然有一些挑战尤其是4G所不能容纳的,例如像频谱危机和高能量消耗。无线系统设计师们面临着满足新型无线应用对高数据速率和机动性要求的持续性增长的需求,因此他们已经开始研究被期望于2020年后就能部署的第五代无线系统。在这篇文章里面,我们提出一个有内门和外门情景之分的潜在的蜂窝结构,并且讨论了多种可行性关于5G无线通信系统的技术,比如大量的MIMO技术,节能通信,认知的广播网络和可见光通信。面临潜在技术的未知挑战也被讨论了。 介绍 信息通信技术(ICT)创新合理的使用对世界经济的提高变得越来越重要。无线通信网络在全球ICT战略中也许是最挑剔的元素,并且支撑着很多其他的行业,它是世界上成长最快最有活力的行业之一。欧洲移动天文台(EMO)报道2010年移动通信业总计税收1740亿欧元,从而超过了航空航天业和制药业。无线技术的发展大大提高了人们在商业运作和社交功能方面通信和生活的能力无线移动通信的显著成就表现在技术创新的快速步伐。从1991年二代移动通信系统(2G)的初次登场到2001年三代系统(3G)的首次起飞,无线移动网络已经实现了从一个纯粹的技术系统到一个能承载大量多媒体内容网络的转变。4G无线系统被设计出来用来满足IMT-A技术使用IP面向所有服务的需求。在4G系统中,先进的无线接口被用于正交频分复用技术(OFDM),多输入多输出系统(MIMO)和链路自适应技术。4G无线网络可支持数据速率可达1Gb/s的低流度,比如流动局域无线访问,还有速率高达100M/s的高流速,例如像移动访问。LTE系统和它的延伸系统LTE-A,作为实用的4G系统已经在全球于最近期或不久的将来部署。 然而,每年仍然有戏剧性增长数量的用户支持移动宽频带系统。越来越多的

外文翻译---图像的边缘检测

附:英文资料翻译 图像的边缘检测 To image edge examination algorithm research academic report Abstract Digital image processing took a relative quite young discipline, is following the computer technology rapid development, day by day obtains the widespread application.The edge took the image one kind of basic characteristic, in the pattern recognition, the image division, the image intensification as well as the image compression and so on in the domain has a more widespread application.Image edge detection method many and varied, in which based on brightness algorithm, is studies the time to be most long, the theory develops the maturest method, it mainly is through some difference operator, calculates its gradient based on image brightness the change, thus examines the edge, mainly has Robert, Laplacian, Sobel, Canny, operators and so on LOG. First as a whole introduced digital image processing and the edge detection survey, has enumerated several kind of at present commonly used edge detection technology and the algorithm, and selects two kinds to use Visual the C language programming realization, through withdraws the image result to two algorithms the comparison, the research discusses their good and bad points. 对图像边缘检测算法的研究学术报告摘要 数字图像处理作为一门相对比较年轻的学科, 伴随着计算机技术的飞速发展, 日益得到广泛的应用. 边缘作为图像的一种基本特征, 在图像识别,图像分割,图像增强以及图像压缩等的领域中有较为广泛的应用.图像边缘提取的手段多种多样,其中基于亮度的算法,是研究时间最久,理论发展最成熟的方法, 它主要是通过一些差分算子, 由图像的亮度计算其梯度的变化, 从而检测出边缘, 主要有Robert, Laplacian, Sobel, Canny, LOG 等算子. 首先从总体上介绍了数字图像处理及边缘提取的概况, 列举了几种目前常用的边缘提取技术和算法,并选取其中两种使用Visual C++语言编程实现,通过对两种算法所提取图像结果的比较,研究探讨它们的优缺点. First chapter introduction §1.1 image edge examination introduction The image edge is one of image most basic characteristics, often is carrying image majority of informations.But the edge exists in the image irregular structure and in

相关文档
最新文档