喜讯是什么意思| 梦见手机丢了又找到了是什么意思| 男人吃女人有什么好处| 田五行属什么| 梅毒早期什么症状| 日加一笔变成什么字| 手书是什么意思| 人缺钾有什么症状| 什么时候跑步减肥效果最好| 大道无为是什么意思| 美国为什么叫鹰酱| 胃炎吃什么药| 集体户口和个人户口有什么区别| 马兰头是什么菜| 长可以加什么偏旁| 生离死别是什么生肖| 羽立念什么| 风吹动窗吹动夜声响是什么歌| 说梦话是什么原因引起的| 今天的日子适合做什么| 心火旺喝什么茶| 一个虫一个夫念什么| 长期腹泻是什么病| 苏州为什么叫姑苏| 补气补血吃什么好| 出尔反尔是什么意思| 右下腹疼痛什么原因| 张学良为什么叫小六子| 血压低头晕吃什么药| 狗到家里是什么预兆| 朋友越来越少暗示什么| 孕妇吃什么好| 唯我独尊是什么意思| 宫颈炎和阴道炎有什么区别| 突然晕倒是什么原因造成的| 为什么喝中药越来越胖| 红萝卜什么时候种| 阴道炎要用什么药| 血沉是什么| 鹅蛋孕妇吃有什么好处| 9月17号是什么星座的| 窝边草是什么意思| 杨柳是什么生肖| 过敏性鼻炎挂什么科| 上火什么症状| 什么是深海鱼| 高笋和茭白有什么区别| 人的心脏在什么位置| 华西医院院长什么级别| 北方的木瓜叫什么| 口字五行属什么| 几何图形是什么| 天花板是什么意思| 化骨龙是什么意思| 来例假头晕是什么原因| 肝功能七项是检查什么| 脚底长鸡眼是什么原因| 妊娠纹什么时候开始长| 如火如荼是什么意思| 跪乳的动物是什么生肖| 支配是什么意思| 喝酒不能吃什么东西| 扒灰什么意思| 孤寡老人国家有什么政策| 石光荣是什么军衔| rh阳性是什么意思| 爱出汗是什么原因女人| 十月五号是什么星座| 命硬的人有什么特征| 自然色是什么颜色| 昆明飞机场叫什么名字| 霸王龙的后代是什么| 今天什么节日| 无厘头什么意思| 大便羊粪状吃什么药| ru是什么意思| 为什么便秘| 不可漂白是什么意思| 办身份证的地方叫什么| 剖腹产可以吃什么| 睡不着吃什么| 三级医院是什么意思| 扁桃体为什么会发炎| 三点水加个真念什么| 转氨酶偏高是什么意思| 审计署是什么级别| 精致是什么意思| hm什么牌子| 之际是什么意思| 什么东西含铅量高| 脚背痒是什么原因| 有缘人什么意思| 火车头是什么意思| 生死离别代表什么生肖| 21度穿什么衣服| 武则天原名叫什么| charging是什么意思| 挂帅是什么意思| 决定的近义词是什么| 碱性磷酸酶偏高吃什么能降下来呢| 阴道发臭是什么原因| 嗓子疼吃什么药| 日行一善下一句是什么| 拉拉裤后面的胶片是做什么用的| 打胰岛素是什么病| 做梦梦到自己拉屎是什么意思| 梦见下大雪是什么意思| 耳朵长疙瘩是什么原因| 东是什么生肖| 鸽子吃什么食物| 什么程度要做肾穿刺| 1880年是什么朝代| 劳改是什么意思| 无精是什么原因造成的| 赡养什么意思| 世界上最长的河流是什么| 前面有个豹子是什么车| 7月7日是什么纪念日| 旗人是什么意思| ep是什么| 端庄的意思是什么| 世界七大奇迹分别是什么| 觉悟高是什么意思| 风湿吃什么药好| 4月份有什么节日| 夜尿多是什么原因引起的| 初级会计考什么科目| 绿豆有什么功效| 市盈率和市净率是什么意思| 属马是什么星座| 隔空打牛是什么意思| 出血热是什么病| 肝火旺盛吃什么药效果最好| 武则天墓为什么不敢挖| 农历5月是什么月| 左舌根疼痛是什么情况| 总是想吐是什么原因| 汗水多吃什么药| 茉莉茶叶属于什么茶| 吃什么东西可以减肥| 什么水果含维生素c最多| 郁结是什么意思| 经常口臭的人是什么原因引起的| 查艾滋挂什么科| 日光性皮炎用什么药膏| 歹人是什么意思| 一什么影子| 10月25日什么星座| 小三阳和大三阳有什么区别| 甘油三酯指的是什么| 羊日冲牛是什么意思| 家里有蚂蚁是什么原因| 得了阴虱用什么药能除根| 肝火旺盛是什么意思| 报销凭证是什么| 五彩斑斓的意思是什么| 心五行属什么| 大什么大| 膂力是什么意思| 怎么看自己五行属什么| 治疗狐臭最好的方法是什么| 丰的部首是什么偏旁| 什么什么入胜| 心肌缺血吃什么食物| 12月6日什么星座| 成人高考是什么| 大公鸡衣服是什么牌子| 交是什么结构的字| 做胃镜前喝的那个液体是什么| 今日是什么生肖日| 今年是什么年| 养尊处优什么意思| 胆囊结石会引起身体什么症状| 希五行属什么| 甲泼尼龙主要治什么| 金益什么字| 洁面膏和洗面奶有什么区别| 过渡句的作用是什么| 最贵的烟是什么| 什么都有| 牟作为姓氏时读什么| 百字五行属什么| 毛周角化症是什么原因引起的| 心脏疼是什么原因| 甲功三项是检查什么| 接触性皮炎用什么药| 男人要吃什么才能壮阳| 仙草是什么| 七月什么星座| 长期便秘是什么原因引起的| 上海话小赤佬是什么意思| 六月十号是什么星座| 除颤是什么意思| 为什么家里有蟑螂| 甲状腺五类是什么意思| cancer是什么意思| 庆幸是什么意思| 玉米是什么时候传入中国的| 安道尔微信暗示什么| 阴部痒是什么原因| 介入室是干什么的| 4岁小孩流鼻血是什么原因| 湖北人喜欢吃什么菜| 脱水什么意思| 物极必反什么意思| 抗酸杆菌是什么| 出生日期查五行缺什么| 柠檬泡水喝有什么作用| 一边脸大一边脸小是什么原因| 走胎是什么意思| 维生素d什么牌子的好| 1884年属什么生肖| lively是什么意思| 杠杠的是什么意思| 狼狗是什么品种| 婴儿蚊虫叮咬红肿用什么药| 女人大腿内侧黑是什么原因引起的| nothomme什么牌子| 鼻子发酸是什么原因| 做梦梦见牛是什么意思| 好运是什么意思| 发改委是做什么的| 燕子吃什么| 男人吃什么补肾壮阳效果最好| 四件套包括什么| 艾灸肚脐眼有什么好处| ra什么意思| 抠是什么意思| 日本的町是什么意思| 超纤皮是什么| 什么是心理健康| ib是什么单位| sp是什么的缩写| 吉利丁片是什么做的| 翘嘴鱼吃什么食物| 梦见做春梦是什么意思| 寒是什么意思| 怀孕吃鹅蛋有什么好处| 女主是什么意思| 胃溃疡什么症状| 女性腋臭什么年龄消失| 狗狗吐黄水是什么原因| 头响脑鸣是什么原因引起的| 农历今天属什么生肖| 螺旋杆菌感染吃什么药| 什么感觉| 什么品牌的洗衣机最好| xy什么意思| 什么天长地久| 南北朝后面是什么朝代| 西瓜吃多了有什么坏处| fpd是什么意思| 椰青是什么| 整天放屁是什么原因| 湿热泄泻是什么意思| 甲亢可以吃什么水果| 榴莲不能与什么食物一起吃| 什么是五险一金| 敬谢不敏是什么意思| 调羹是什么意思| 绿壳鸡蛋是什么鸡生的| 梅子是什么水果| 方法是什么意思| 小老头是什么意思| 百度

好站推荐:必盛互联 一站式建站服务销售平台

百度 张大千常以画论吃,以吃论画。

In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping between two vector spaces that preserves the operations of vector addition and scalar multiplication. The same names and the same definition are also used for the more general case of modules over a ring; see Module homomorphism.

If a linear map is a bijection then it is called a linear isomorphism. In the case where , a linear map is called a linear endomorphism. Sometimes the term linear operator refers to this case,[1] but the term "linear operator" can have different meanings for different conventions: for example, it can be used to emphasize that and are real vector spaces (not necessarily with ),[citation needed] or it can be used to emphasize that is a function space, which is a common convention in functional analysis.[2] Sometimes the term linear function has the same meaning as linear map, while in analysis it does not.

A linear map from to always maps the origin of to the origin of . Moreover, it maps linear subspaces in onto linear subspaces in (possibly of a lower dimension);[3] for example, it maps a plane through the origin in to either a plane through the origin in , a line through the origin in , or just the origin in . Linear maps can often be represented as matrices, and simple examples include rotation and reflection linear transformations.

In the language of category theory, linear maps are the morphisms of vector spaces, and they form a category equivalent to the one of matrices.

Definition and first consequences

edit

Let   and   be vector spaces over the same field  . A function   is said to be a linear map if for any two vectors   and any scalar   the following two conditions are satisfied:

  • Additivity / operation of addition  
  • Homogeneity of degree 1 / operation of scalar multiplication  

Thus, a linear map is said to be operation preserving. In other words, it does not matter whether the linear map is applied before (the right hand sides of the above examples) or after (the left hand sides of the examples) the operations of addition and scalar multiplication.

By the associativity of the addition operation denoted as +, for any vectors   and scalars   the following equality holds:[4][5]   Thus a linear map is one which preserves linear combinations.

Denoting the zero elements of the vector spaces   and   by   and   respectively, it follows that   Let   and   in the equation for homogeneity of degree 1:  

A linear map   with   viewed as a one-dimensional vector space over itself is called a linear functional.[6]

These statements generalize to any left-module   over a ring   without modification, and to any right-module upon reversing of the scalar multiplication.

Examples

edit
  • A prototypical example that gives linear maps their name is a function  , of which the graph is a line through the origin.[7]
  • More generally, any homothety   centered in the origin of a vector space is a linear map (here c is a scalar).
  • The zero map   between two vector spaces (over the same field) is linear.
  • The identity map on any module is a linear operator.
  • For real numbers, the map   is not linear.
  • For real numbers, the map   is not linear (but is an affine transformation).
  • If   is a   real matrix, then   defines a linear map from   to   by sending a column vector   to the column vector  . Conversely, any linear map between finite-dimensional vector spaces can be represented in this manner; see the § Matrices, below.
  • If   is an isometry between real normed spaces such that   then   is a linear map. This result is not necessarily true for complex normed space.[8]
  • Differentiation defines a linear map from the space of all differentiable functions to the space of all functions. It also defines a linear operator on the space of all smooth functions (a linear operator is a linear endomorphism, that is, a linear map with the same domain and codomain). Indeed,  
  • A definite integral over some interval I is a linear map from the space of all real-valued integrable functions on I to  . Indeed,  
  • An indefinite integral (or antiderivative) with a fixed integration starting point defines a linear map from the space of all real-valued integrable functions on   to the space of all real-valued, differentiable functions on  . Without a fixed starting point, the antiderivative maps to the quotient space of the differentiable functions by the linear space of constant functions.
  • If   and   are finite-dimensional vector spaces over a field F, of respective dimensions m and n, then the function that maps linear maps   to n × m matrices in the way described in § Matrices (below) is a linear map, and even a linear isomorphism.
  • The expected value of a random variable (which is in fact a function, and as such an element of a vector space) is linear, as for random variables   and   we have   and  , but the variance of a random variable is not linear.

Linear extensions

edit

Often, a linear map is constructed by defining it on a subset of a vector space and then extending by linearity to the linear span of the domain. Suppose   and   are vector spaces and   is a function defined on some subset   Then a linear extension of   to   if it exists, is a linear map   defined on   that extends  [note 1] (meaning that   for all  ) and takes its values from the codomain of  [9] When the subset   is a vector subspace of   then a ( -valued) linear extension of   to all of   is guaranteed to exist if (and only if)   is a linear map.[9] In particular, if   has a linear extension to   then it has a linear extension to all of  

The map   can be extended to a linear map   if and only if whenever   is an integer,   are scalars, and   are vectors such that   then necessarily  [10] If a linear extension of   exists then the linear extension   is unique and   holds for all   and   as above.[10] If   is linearly independent then every function   into any vector space has a linear extension to a (linear) map   (the converse is also true).

For example, if   and   then the assignment   and   can be linearly extended from the linearly independent set of vectors   to a linear map on   The unique linear extension   is the map that sends   to  

Every (scalar-valued) linear functional   defined on a vector subspace of a real or complex vector space   has a linear extension to all of   Indeed, the Hahn–Banach dominated extension theorem even guarantees that when this linear functional   is dominated by some given seminorm   (meaning that   holds for all   in the domain of  ) then there exists a linear extension to   that is also dominated by  

Matrices

edit

If   and   are finite-dimensional vector spaces and a basis is defined for each vector space, then every linear map from   to   can be represented by a matrix.[11] This is useful because it allows concrete calculations. Matrices yield examples of linear maps: if   is a real   matrix, then   describes a linear map   (see Euclidean space).

Let   be a basis for  . Then every vector   is uniquely determined by the coefficients   in the field  :  

If   is a linear map,  

which implies that the function f is entirely determined by the vectors  . Now let   be a basis for  . Then we can represent each vector   as  

Thus, the function   is entirely determined by the values of  . If we put these values into an   matrix  , then we can conveniently use it to compute the vector output of   for any vector in  . To get  , every column   of   is a vector   corresponding to   as defined above. To define it more clearly, for some column   that corresponds to the mapping  ,   where   is the matrix of  . In other words, every column   has a corresponding vector   whose coordinates   are the elements of column  . A single linear map may be represented by many matrices. This is because the values of the elements of a matrix depend on the bases chosen.

The matrices of a linear transformation can be represented visually:

  1. Matrix for   relative to  :  
  2. Matrix for   relative to  :  
  3. Transition matrix from   to  :  
  4. Transition matrix from   to  :  
 
The relationship between matrices in a linear transformation

Such that starting in the bottom left corner   and looking for the bottom right corner  , one would left-multiply—that is,  . The equivalent method would be the "longer" method going clockwise from the same point such that   is left-multiplied with  , or  .

Examples in two dimensions

edit

In two-dimensional space R2 linear maps are described by 2 × 2 matrices. These are some examples:

  • rotation
    • by 90 degrees counterclockwise:  
    • by an angle θ counterclockwise:  
  • reflection
    • through the x axis:  
    • through the y axis:  
    • through a line making an angle θ with the origin:  
  • scaling by 2 in all directions:  
  • horizontal shear mapping:  
  • skew of the y axis by an angle θ:  
  • squeeze mapping:  
  • projection onto the y axis:  

If a linear map is only composed of rotation, reflection, and/or uniform scaling, then the linear map is a conformal linear transformation.

Vector space of linear maps

edit

The composition of linear maps is linear: if   and   are linear, then so is their composition  . It follows from this that the class of all vector spaces over a given field K, together with K-linear maps as morphisms, forms a category.

The inverse of a linear map, when defined, is again a linear map.

If   and   are linear, then so is their pointwise sum  , which is defined by  .

If   is linear and   is an element of the ground field  , then the map  , defined by  , is also linear.

Thus the set   of linear maps from   to   itself forms a vector space over  ,[12] sometimes denoted  .[13] Furthermore, in the case that  , this vector space, denoted  , is an associative algebra under composition of maps, since the composition of two linear maps is again a linear map, and the composition of maps is always associative. This case is discussed in more detail below.

Given again the finite-dimensional case, if bases have been chosen, then the composition of linear maps corresponds to the matrix multiplication, the addition of linear maps corresponds to the matrix addition, and the multiplication of linear maps with scalars corresponds to the multiplication of matrices with scalars.

Endomorphisms and automorphisms

edit

A linear transformation   is an endomorphism of  ; the set of all such endomorphisms   together with addition, composition and scalar multiplication as defined above forms an associative algebra with identity element over the field   (and in particular a ring). The multiplicative identity element of this algebra is the identity map  .

An endomorphism of   that is also an isomorphism is called an automorphism of  . The composition of two automorphisms is again an automorphism, and the set of all automorphisms of   forms a group, the automorphism group of   which is denoted by   or  . Since the automorphisms are precisely those endomorphisms which possess inverses under composition,   is the group of units in the ring  .

If   has finite dimension  , then   is isomorphic to the associative algebra of all   matrices with entries in  . The automorphism group of   is isomorphic to the general linear group   of all   invertible matrices with entries in  .

Kernel, image and the rank–nullity theorem

edit

If   is linear, we define the kernel and the image or range of   by  

  is a subspace of   and   is a subspace of  . The following dimension formula is known as the rank–nullity theorem:[14]  

The number   is also called the rank of   and written as  , or sometimes,  ;[15][16] the number   is called the nullity of   and written as   or  .[15][16] If   and   are finite-dimensional, bases have been chosen and   is represented by the matrix  , then the rank and nullity of   are equal to the rank and nullity of the matrix  , respectively.

Cokernel

edit

A subtler invariant of a linear transformation   is the cokernel, which is defined as  

This is the dual notion to the kernel: just as the kernel is a subspace of the domain, the co-kernel is a quotient space of the target. Formally, one has the exact sequence  

These can be interpreted thus: given a linear equation f(v) = w to solve,

  • the kernel is the space of solutions to the homogeneous equation f(v) = 0, and its dimension is the number of degrees of freedom in the space of solutions, if it is not empty;
  • the co-kernel is the space of constraints that the solutions must satisfy, and its dimension is the maximal number of independent constraints.

The dimension of the co-kernel and the dimension of the image (the rank) add up to the dimension of the target space. For finite dimensions, this means that the dimension of the quotient space W/f(V) is the dimension of the target space minus the dimension of the image.

As a simple example, consider the map f: R2R2, given by f(x, y) = (0, y). Then for an equation f(x, y) = (a, b) to have a solution, we must have a = 0 (one constraint), and in that case the solution space is (x, b) or equivalently stated, (0, b) + (x, 0), (one degree of freedom). The kernel may be expressed as the subspace (x, 0) < V: the value of x is the freedom in a solution – while the cokernel may be expressed via the map WR,  : given a vector (a, b), the value of a is the obstruction to there being a solution.

An example illustrating the infinite-dimensional case is afforded by the map f: RR,   with b1 = 0 and bn + 1 = an for n > 0. Its image consists of all sequences with first element 0, and thus its cokernel consists of the classes of sequences with identical first element. Thus, whereas its kernel has dimension 0 (it maps only the zero sequence to the zero sequence), its co-kernel has dimension 1. Since the domain and the target space are the same, the rank and the dimension of the kernel add up to the same sum as the rank and the dimension of the co-kernel ( ), but in the infinite-dimensional case it cannot be inferred that the kernel and the co-kernel of an endomorphism have the same dimension (0 ≠ 1). The reverse situation obtains for the map h: RR,   with cn = an + 1. Its image is the entire target space, and hence its co-kernel has dimension 0, but since it maps all sequences in which only the first element is non-zero to the zero sequence, its kernel has dimension 1.

Index

edit

For a linear operator with finite-dimensional kernel and co-kernel, one may define index as:   namely the degrees of freedom minus the number of constraints.

For a transformation between finite-dimensional vector spaces, this is just the difference dim(V) ? dim(W), by rank–nullity. This gives an indication of how many solutions or how many constraints one has: if mapping from a larger space to a smaller one, the map may be onto, and thus will have degrees of freedom even without constraints. Conversely, if mapping from a smaller space to a larger one, the map cannot be onto, and thus one will have constraints even without degrees of freedom.

The index of an operator is precisely the Euler characteristic of the 2-term complex 0 → VW → 0. In operator theory, the index of Fredholm operators is an object of study, with a major result being the Atiyah–Singer index theorem.[17]

Algebraic classifications of linear transformations

edit

No classification of linear maps could be exhaustive. The following incomplete list enumerates some important classifications that do not require any additional structure on the vector space.

Let V and W denote vector spaces over a field F and let T: VW be a linear map.

Monomorphism

edit

T is said to be injective or a monomorphism if any of the following equivalent conditions are true:

  1. T is one-to-one as a map of sets.
  2. ker T = {0V}
  3. dim(ker T) = 0
  4. T is monic or left-cancellable, which is to say, for any vector space U and any pair of linear maps R: UV and S: UV, the equation TR = TS implies R = S.
  5. T is left-invertible, which is to say there exists a linear map S: WV such that ST is the identity map on V.

Epimorphism

edit

T is said to be surjective or an epimorphism if any of the following equivalent conditions are true:

  1. T is onto as a map of sets.
  2. coker T = {0W}
  3. T is epic or right-cancellable, which is to say, for any vector space U and any pair of linear maps R: WU and S: WU, the equation RT = ST implies R = S.
  4. T is right-invertible, which is to say there exists a linear map S: WV such that TS is the identity map on W.

Isomorphism

edit

T is said to be an isomorphism if it is both left- and right-invertible. This is equivalent to T being both one-to-one and onto (a bijection of sets) or also to T being both epic and monic, and so being a bimorphism.

If T: VV is an endomorphism, then:

  • If, for some positive integer n, the n-th iterate of T, Tn, is identically zero, then T is said to be nilpotent.
  • If T2 = T, then T is said to be idempotent
  • If T = kI, where k is some scalar, then T is said to be a scaling transformation or scalar multiplication map; see scalar matrix.

Change of basis

edit

Given a linear map which is an endomorphism whose matrix is A, in the basis B of the space it transforms vector coordinates [u] as [v] = A[u]. As vectors change with the inverse of B (vectors coordinates are contravariant) its inverse transformation is [v] = B[v'].

Substituting this in the first expression   hence  

Therefore, the matrix in the new basis is A′ = B?1AB, being B the matrix of the given basis.

Therefore, linear maps are said to be 1-co- 1-contra-variant objects, or type (1, 1) tensors.

Continuity

edit

A linear transformation between topological vector spaces, for example normed spaces, may be continuous. If its domain and codomain are the same, it will then be a continuous linear operator. A linear operator on a normed linear space is continuous if and only if it is bounded, for example, when the domain is finite-dimensional.[18] An infinite-dimensional domain may have discontinuous linear operators.

An example of an unbounded, hence discontinuous, linear transformation is differentiation on the space of smooth functions equipped with the supremum norm (a function with small values can have a derivative with large values, while the derivative of 0 is 0). For a specific example, sin(nx)/n converges to 0, but its derivative cos(nx) does not, so differentiation is not continuous at 0 (and by a variation of this argument, it is not continuous anywhere).

Applications

edit

A specific application of linear maps is for geometric transformations, such as those performed in computer graphics, where the translation, rotation and scaling of 2D or 3D objects is performed by the use of a transformation matrix. Linear mappings also are used as a mechanism for describing change: for example in calculus correspond to derivatives; or in relativity, used as a device to keep track of the local transformations of reference frames.

Another application of these transformations is in compiler optimizations of nested-loop code, and in parallelizing compiler techniques.

See also

edit

Notes

edit
  1. ^ "Linear transformations of V into V are often called linear operators on V." Rudin 1976, p. 207
  2. ^ Let V and W be two real vector spaces. A mapping a from V into W Is called a 'linear mapping' or 'linear transformation' or 'linear operator' [...] from V into W, if
      for all  ,
      for all   and all real λ. Bronshtein & Semendyayev 2004, p. 316
  3. ^ Rudin 1991, p. 14
    Here are some properties of linear mappings   whose proofs are so easy that we omit them; it is assumed that   and  :
    1.  
    2. If A is a subspace (or a convex set, or a balanced set) the same is true of  
    3. If B is a subspace (or a convex set, or a balanced set) the same is true of  
    4. In particular, the set:   is a subspace of X, called the null space of  .
  4. ^ Rudin 1991, p. 14. Suppose now that X and Y are vector spaces over the same scalar field. A mapping   is said to be linear if   for all   and all scalars   and  . Note that one often writes  , rather than  , when   is linear.
  5. ^ Rudin 1976, p. 206. A mapping A of a vector space X into a vector space Y is said to be a linear transformation if:   for all   and all scalars c. Note that one often writes   instead of   if A is linear.
  6. ^ Rudin 1991, p. 14. Linear mappings of X onto its scalar field are called linear functionals.
  7. ^ "terminology - What does 'linear' mean in Linear Algebra?". Mathematics Stack Exchange. Retrieved 2025-08-06.
  8. ^ Wilansky 2013, pp. 21–26.
  9. ^ a b Kubrusly 2001, p. 57.
  10. ^ a b Schechter 1996, pp. 277–280.
  11. ^ Rudin 1976, p. 210 Suppose   and   are bases of vector spaces X and Y, respectively. Then every   determines a set of numbers   such that   It is convenient to represent these numbers in a rectangular array of m rows and n columns, called an m by n matrix:   Observe that the coordinates   of the vector   (with respect to the basis  ) appear in the jth column of  . The vectors   are therefore sometimes called the column vectors of  . With this terminology, the range of A is spanned by the column vectors of  .
  12. ^ Axler (2015) p. 52, § 3.3
  13. ^ Tu (2011), p. 19, § 3.1
  14. ^ Horn & Johnson 2013, 0.2.3 Vector spaces associated with a matrix or linear transformation, p. 6
  15. ^ a b Katznelson & Katznelson (2008) p. 52, § 2.5.1
  16. ^ a b Halmos (1974) p. 90, § 50
  17. ^ Nistor, Victor (2001) [1994], "Index theory", Encyclopedia of Mathematics, EMS Press: "The main question in index theory is to provide index formulas for classes of Fredholm operators ... Index theory has become a subject on its own only after M. F. Atiyah and I. Singer published their index theorems"
  18. ^ Rudin 1991, p. 15 1.18 Theorem Let   be a linear functional on a topological vector space X. Assume   for some  . Then each of the following four properties implies the other three:
    1.   is continuous
    2. The null space   is closed.
    3.   is not dense in X.
    4.   is bounded in some neighbourhood V of 0.
  1. ^ One map   is said to extend another map   if when   is defined at a point   then so is   and  

Bibliography

edit
甲沟炎涂什么药膏 精液发黄是什么原因引起的 英语专八是什么水平 呸是什么意思 尿酸高可以喝什么饮料
品牌是什么 怀孕40天有什么症状 什么教无类 脚底褪皮是什么原因 云服务是什么
喜乐是什么意思 盆腔炎吃什么药效果最好 雍正叫什么 紫色和蓝色混合是什么颜色 蔡徐坤粉丝名叫什么
M3什么意思 同舟共济是什么意思 yishion是什么牌子 江字五行属什么 白带多用什么药
蒲公英和什么相克致死clwhiglsz.com 黄疸高是什么原因引起的hcv8jop5ns2r.cn 生殖感染用什么消炎药效果好hcv7jop9ns5r.cn 床头上面挂什么画好hcv8jop6ns1r.cn 艳阳高照是什么生肖hcv8jop6ns0r.cn
熬是什么意思hcv9jop3ns4r.cn 为什么无缘无故流鼻血hcv9jop6ns1r.cn sds是什么意思zhongyiyatai.com 女性膀胱炎吃什么药hcv9jop5ns6r.cn ysl是什么意思hcv8jop8ns8r.cn
1993属什么生肖hcv8jop1ns6r.cn 食字五行属什么hcv8jop5ns0r.cn 交叉感染是什么意思hcv8jop2ns1r.cn 睡眠障碍挂什么科hcv7jop4ns6r.cn 金牛座后面是什么星座hcv8jop1ns9r.cn
农历七月是什么月份hcv8jop9ns7r.cn 糖化血红蛋白是什么意思hcv8jop2ns0r.cn 血热是什么症状hcv8jop3ns4r.cn 皮质醇是什么意思hcv8jop1ns2r.cn 脂肪瘤是什么原因引起的fenrenren.com
百度