Multi-dimensional Arrays

Julia, like most technical computing languages, provides a first-class array implementation. Most technical computing languages pay a lot of attention to their array implementation at the expense of other containers. Julia does not treat arrays in any special way. The array library is implemented almost completely in Julia itself, and derives its performance from the compiler, just like any other code written in Julia. As such, it’s also possible to define custom array types by inheriting from AbstractArray. See the manual section on the AbstractArray interface for more details on implementing a custom array type.

An array is a collection of objects stored in a multi-dimensional grid. In the most general case, an array may contain objects of type Any. For most computational purposes, arrays should contain objects of a more specific type, such as Float64 or Int32.

In general, unlike many other technical computing languages, Julia does not expect programs to be written in a vectorized style for performance. Julia’s compiler uses type inference and generates optimized code for scalar array indexing, allowing programs to be written in a style that is convenient and readable, without sacrificing performance, and using less memory at times.

In Julia, all arguments to functions are passed by reference. Some technical computing languages pass arrays by value, and this is convenient in many cases. In Julia, modifications made to input arrays within a function will be visible in the parent function. The entire Julia array library ensures that inputs are not modified by library functions. User code, if it needs to exhibit similar behavior, should take care to create a copy of inputs that it may modify.


Basic Functions

eltype(A)the type of the elements contained in A
length(A)the number of elements in A
ndims(A)the number of dimensions of A
size(A)a tuple containing the dimensions of A
size(A,n)the size of A along a particular dimension
indices(A)a tuple containing the valid indices of A
indices(A,n)a range expressing the valid indices along dimension n
eachindex(A)an efficient iterator for visiting each position in A
stride(A,k)the stride (linear index distance between adjacent elements) along dimension k
strides(A)a tuple of the strides in each dimension

Construction and Initialization

Many functions for constructing and initializing arrays are provided. In the following list of such functions, calls with a dims... argument can either take a single tuple of dimension sizes or a series of dimension sizes passed as a variable number of arguments.

Array{type}(dims...)an uninitialized dense array
zeros(type,dims...)an array of all zeros of specified type, defaults to Float64 if type not specified
zeros(A)an array of all zeros of same element type and shape of A
ones(type,dims...)an array of all ones of specified type, defaults to Float64 if type not specified
ones(A)an array of all ones of same element type and shape of A
trues(dims...)a Bool array with all values true
trues(A)a Bool array with all values true and the shape of A
falses(dims...)a Bool array with all values false
falses(A)a Bool array with all values false and the shape of A
reshape(A,dims...)an array with the same data as the given array, but with different dimensions.
copy(A)copy A
deepcopy(A)copy A, recursively copying its elements
similar(A,element_type,dims...)an uninitialized array of the same type as the given array (dense, sparse, etc.), but with the specified element type and dimensions. The second and third arguments are both optional, defaulting to the element type and dimensions of A if omitted.
reinterpret(type,A)an array with the same binary data as the given array, but with the specified element type
rand(dims)Array of Float64s with random, iid [1] and uniformly distributed values in the half-open interval \([0, 1)\)
randn(dims)Array of Float64s with random, iid and standard normally distributed random values
eye(n)n-by-n identity matrix
eye(m,n)m-by-n identity matrix
linspace(start,stop,n)range of n linearly spaced elements from start to stop
fill!(A,x)fill the array A with the value x
fill(x,dims)create an array filled with the value x
[1]iid, independently and identically distributed.

The syntax [A,B,C,...] constructs a 1-d array (vector) of its arguments.


Arrays can be constructed and also concatenated using the following functions:

cat(k,A...)concatenate input n-d arrays along the dimension k
vcat(A...)shorthand for cat(1,A...)
hcat(A...)shorthand for cat(2,A...)

Scalar values passed to these functions are treated as 1-element arrays.

The concatenation functions are used so often that they have special syntax:


hvcat() concatenates in both dimension 1 (with semicolons) and dimension 2 (with spaces).

Typed array initializers

An array with a specific element type can be constructed using the syntax T[A,B,C,...]. This will construct a 1-d array with element type T, initialized to contain elements A, B, C, etc. For example Any[x,y,z] constructs a heterogeneous array that can contain any values.

Concatenation syntax can similarly be prefixed with a type to specify the element type of the result.



Comprehensions provide a general and powerful way to construct arrays. Comprehension syntax is similar to set construction notation in mathematics:


The meaning of this form is that F(x,y,...) is evaluated with the variables x, y, etc. taking on each value in their given list of values. Values can be specified as any iterable object, but will commonly be ranges like 1:n or 2:(n-1), or explicit arrays of values like [1.2,3.4,5.7]. The result is an N-d dense array with dimensions that are the concatenation of the dimensions of the variable ranges rx, ry, etc. and each F(x,y,...) evaluation returns a scalar.

The following example computes a weighted average of the current element and its left and right neighbor along a 1-d grid. :


The resulting array type depends on the types of the computed elements. In order to control the type explicitly, a type can be prepended to the comprehension. For example, we could have requested the result in single precision by writing:


Generator Expressions

Comprehensions can also be written without the enclosing square brackets, producing an object known as a generator. This object can be iterated to produce values on demand, instead of allocating an array and storing them in advance (see Iteration). For example, the following expression sums a series without allocating memory:


When writing a generator expression with multiple dimensions inside an argument list, parentheses are needed to separate the generator from subsequent arguments:


All comma-separated expressions after for are interpreted as ranges. Adding parentheses lets us add a third argument to map:


Ranges in generators and comprehensions can depend on previous ranges by writing multiple for keywords:


In such cases, the result is always 1-d.

Generated values can be filtered using the if keyword:



The general syntax for indexing into an n-dimensional array A is:


where each I_k may be:

  1. A scalar integer
  2. A Range of the form a:b, or a:b:c
  3. A : or Colon() to select entire dimensions
  4. An arbitrary integer array, including the empty array []
  5. A boolean array to select a vector of elements at its true indices

If all the indices are scalars, then the result X is a single element from the array A. Otherwise, X is an array with the same number of dimensions as the sum of the dimensionalities of all the indices.

If all indices are vectors, for example, then the shape of X would be (length(I_1),length(I_2),...,length(I_n)), with location (i_1,i_2,...,i_n) of X containing the value A[I_1[i_1],I_2[i_2],...,I_n[i_n]]. If I_1 is changed to a two-dimensional matrix, then X becomes an n+1-dimensional array of shape (size(I_1,1),size(I_1,2),length(I_2),...,length(I_n)). The matrix adds a dimension. The location (i_1,i_2,i_3,...,i_{n+1}) contains the value at A[I_1[i_1,i_2],I_2[i_3],...,I_n[i_{n+1}]]. All dimensions indexed with scalars are dropped. For example, the result of A[2,I,3] is an array with size size(I). Its ith element is populated by A[2,I[i],3].

Indexing by a boolean array B is effectively the same as indexing by the vector that is returned by find(B). Often referred to as logical indexing, this selects elements at the indices where the values are true, akin to a mask. A logical index must be a vector of the same length as the dimension it indexes into, or it must be the only index provided and match the size and dimensionality of the array it indexes into. It is generally more efficient to use boolean arrays as indices directly instead of first calling find().

Additionally, single elements of a multidimensional array can be indexed as x=A[I], where I is a CartesianIndex. It effectively behaves like an n-tuple of integers spanning multiple dimensions of A. See Iteration below.

As a special part of this syntax, the end keyword may be used to represent the last index of each dimension within the indexing brackets, as determined by the size of the innermost array being indexed. Indexing syntax without the end keyword is equivalent to a call to getindex:




Empty ranges of the form n:n-1 are sometimes used to indicate the inter-index location between n-1 and n. For example, the searchsorted() function uses this convention to indicate the insertion point of a value not found in a sorted array:



The general syntax for assigning values in an n-dimensional array A is:


where each I_k may be:

  1. A scalar integer
  2. A Range of the form a:b, or a:b:c
  3. A : or Colon() to select entire dimensions
  4. An arbitrary integer array, including the empty array []
  5. A boolean array to select elements at its true indices

If X is an array, it must have the same number of elements as the product of the lengths of the indices: prod(length(I_1),length(I_2),...,length(I_n)). The value in location I_1[i_1],I_2[i_2],...,I_n[i_n] of A is overwritten with the value X[i_1,i_2,...,i_n]. If X is not an array, its value is written to all referenced locations of A.

A boolean array used as an index behaves as in getindex(), behaving as though it is first transformed with find().

Index assignment syntax is equivalent to a call to setindex!():





The recommended ways to iterate over a whole array are

forainA# Do something with the element aendforiineachindex(A)# Do something with i and/or A[i]end

The first construct is used when you need the value, but not index, of each element. In the second construct, i will be an Int if A is an array type with fast linear indexing; otherwise, it will be a CartesianIndex:


In contrast with fori=1:length(A), iterating with eachindex provides an efficient way to iterate over any array type.

Array traits

If you write a custom AbstractArray type, you can specify that it has fast linear indexing using


This setting will cause eachindex iteration over a MyArray to use integers. If you don’t specify this trait, the default value LinearSlow() is used.

Vectorized Operators and Functions

The following operators are supported for arrays. The dot version of a binary operator should be used for elementwise operations.

  1. Unary arithmetic — -, +, !
  2. Binary arithmetic — +, -, *, .*, /, ./, \, .\, ^, .^, div, mod
  3. Comparison — .==, .!=, .<, .<=, .>, .>=
  4. Unary Boolean or bitwise — ~
  5. Binary Boolean or bitwise — &, |, $

Some operators without dots operate elementwise anyway when one argument is a scalar. These operators are *, +, -, and the bitwise operators. The operators / and \ operate elementwise when the denominator is a scalar.

Note that comparisons such as == operate on whole arrays, giving a single boolean answer. Use dot operators for elementwise comparisons.

The following built-in functions are also vectorized, whereby the functions act elementwise:


Note that there is a difference between min() and max(), which operate elementwise over multiple array arguments, and minimum() and maximum(), which find the smallest and largest values within an array.

Julia provides the @vectorize_1arg() and @vectorize_2arg() macros to automatically vectorize any function of one or two arguments respectively. Each of these takes two arguments, namely the Type of argument (which is usually chosen to be the most general possible) and the name of the function to vectorize. Here is a simple example:

julia>square(x)=x^2square(genericfunction with1method)julia>@vectorize_1argNumbersquaresquare(genericfunction with2methods)julia>methods(square)# 2 methods for generic function "square":square{T<:Number}(x::AbstractArray{T,N<:Any})atoperators.jl:555square(x)atnone:1julia>square([124;567])2×3Array{Int64,2}:1416253649


It is sometimes useful to perform element-by-element binary operations on arrays of different sizes, such as adding a vector to each column of a matrix. An inefficient way to do this would be to replicate the vector to the size of the matrix:


This is wasteful when dimensions get large, so Julia offers broadcast(), which expands singleton dimensions in array arguments to match the corresponding dimension in the other array without using extra memory, and applies the given function elementwise:


Elementwise operators such as .+ and .* perform broadcasting if necessary. There is also a broadcast!() function to specify an explicit destination, and broadcast_getindex() and broadcast_setindex!() that broadcast the indices before indexing. Moreover, f.(args...) is equivalent to broadcast(f,args...), providing a convenient syntax to broadcast any function (Dot Syntax for Vectorizing Functions).


The base array type in Julia is the abstract type AbstractArray{T,N}. It is parametrized by the number of dimensions N and the element type T. AbstractVector and AbstractMatrix are aliases for the 1-d and 2-d cases. Operations on AbstractArray objects are defined using higher level operators and functions, in a way that is independent of the underlying storage. These operations generally work correctly as a fallback for any specific array implementation.

The AbstractArray type includes anything vaguely array-like, and implementations of it might be quite different from conventional arrays. For example, elements might be computed on request rather than stored. However, any concrete AbstractArray{T,N} type should generally implement at least size(A) (returning an Int tuple), getindex(A,i) and getindex(A,i1,...,iN); mutable arrays should also implement setindex!(). It is recommended that these operations have nearly constant time complexity, or technically Õ(1) complexity, as otherwise some array functions may be unexpectedly slow. Concrete types should also typically provide a similar(A,T=eltype(A),dims=size(A)) method, which is used to allocate a similar array for copy() and other out-of-place operations. No matter how an AbstractArray{T,N} is represented internally, T is the type of object returned by integer indexing (A[1,...,1], when A is not empty) and N should be the length of the tuple returned by size().

DenseArray is an abstract subtype of AbstractArray intended to include all arrays that are laid out at regular offsets in memory, and which can therefore be passed to external C and Fortran functions expecting this memory layout. Subtypes should provide a method stride(A,k) that returns the “stride” of dimension k: increasing the index of dimension k by 1 should increase the index i of getindex(A,i) by stride(A,k). If a pointer conversion method Base.unsafe_convert(Ptr{T},A) is provided, the memory layout should correspond in the same way to these strides.

The Array type is a specific instance of DenseArray where elements are stored in column-major order (see additional notes in Performance Tips). Vector and Matrix are aliases for the 1-d and 2-d cases. Specific operations such as scalar indexing, assignment, and a few other basic storage-specific operations are all that have to be implemented for Array, so that the rest of the array library can be implemented in a generic manner.

SubArray is a specialization of AbstractArray that performs indexing by reference rather than by copying. A SubArray is created with the view() function, which is called the same way as getindex() (with an array and a series of index arguments). The result of view() looks the same as the result of getindex(), except the data is left in place. view() stores the input index vectors in a SubArray object, which can later be used to index the original array indirectly.

StridedVector and StridedMatrix are convenient aliases defined to make it possible for Julia to call a wider range of BLAS and LAPACK functions by passing them either Array or SubArray objects, and thus saving inefficiencies from memory allocation and copying.

The following example computes the QR decomposition of a small section of a larger array, without creating any temporaries, and by calling the appropriate LAPACK function with the right leading dimension size and stride parameters.


Sparse Matrices

Sparse matrices are matrices that contain enough zeros that storing them in a special data structure leads to savings in space and execution time. Sparse matrices may be used when operations on the sparse representation of a matrix lead to considerable gains in either time or space when compared to performing the same operations on a dense matrix.

Compressed Sparse Column (CSC) Storage

In Julia, sparse matrices are stored in the Compressed Sparse Column (CSC) format. Julia sparse matrices have the type SparseMatrixCSC{Tv,Ti}, where Tv is the type of the nonzero values, and Ti is the integer type for storing column pointers and row indices.:

type SparseMatrixCSC{Tv,Ti<:Integer}<:AbstractSparseMatrix{Tv,Ti}m::Int# Number of rowsn::Int# Number of columnscolptr::Vector{Ti}# Column i is in colptr[i]:(colptr[i+1]-1)rowval::Vector{Ti}# Row values of nonzerosnzval::Vector{Tv}# Nonzero valuesend

The compressed sparse column storage makes it easy and quick to access the elements in the column of a sparse matrix, whereas accessing the sparse matrix by rows is considerably slower. Operations such as insertion of nonzero values one at a time in the CSC structure tend to be slow. This is because all elements of the sparse matrix that are beyond the point of insertion have to be moved one place over.

All operations on sparse matrices are carefully implemented to exploit the CSC data structure for performance, and to avoid expensive operations.

If you have data in CSC format from a different application or library, and wish to import it in Julia, make sure that you use 1-based indexing. The row indices in every column need to be sorted. If your SparseMatrixCSC object contains unsorted row indices, one quick way to sort them is by doing a double transpose.

In some applications, it is convenient to store explicit zero values in a SparseMatrixCSC. These are accepted by functions in Base (but there is no guarantee that they will be preserved in mutating operations). Such explicitly stored zeros are treated as structural nonzeros by many routines. The nnz() function returns the number of elements explicitly stored in the sparse data structure, including structural nonzeros. In order to count the exact number of actual values that are nonzero, use countnz(), which inspects every stored element of a sparse matrix.

Sparse matrix constructors

The simplest way to create sparse matrices is to use functions equivalent to the zeros() and eye() functions that Julia provides for working with dense matrices. To produce sparse matrices instead, you can use the same names with an sp prefix:


The sparse() function is often a handy way to construct sparse matrices. It takes as its input a vector I of row indices, a vector J of column indices, and a vector V of nonzero values. sparse(I,J,V) constructs a sparse matrix such that S[I[k],J[k]]=V[k].


The inverse of the sparse() function is findn(), which retrieves the inputs used to create the sparse matrix.


Another way to create sparse matrices is to convert a dense matrix into a sparse matrix using the sparse() function:


You can go in the other direction using the full() function. The issparse() function can be used to query if a matrix is sparse.


Sparse matrix operations

Arithmetic operations on sparse matrices also work as they do on dense matrices. Indexing of, assignment into, and concatenation of sparse matrices work in the same way as dense matrices. Indexing operations, especially assignment, are expensive, when carried out one element at a time. In many cases it may be better to convert the sparse matrix into (I,J,V) format using findnz(), manipulate the non-zeroes or the structure in the dense vectors (I,J,V), and then reconstruct the sparse matrix.

Correspondence of dense and sparse methods

The following table gives a correspondence between built-in methods on sparse matrices and their corresponding methods on dense matrix types. In general, methods that generate sparse matrices differ from their dense counterparts in that the resulting matrix follows the same sparsity pattern as a given sparse matrix S, or that the resulting sparse matrix has density d, i.e. each matrix element has a probability d of being non-zero.

Details can be found in the Sparse Vectors and Matrices section of the standard library reference.

spzeros(m,n)zeros(m,n)Creates a m-by-n matrix of zeros. (spzeros(m,n) is empty.)
spones(S)ones(m,n)Creates a matrix filled with ones. Unlike the dense version, spones() has the same sparsity pattern as S.
speye(n)eye(n)Creates a n-by-n identity matrix.
full(S)sparse(A)Interconverts between dense and sparse formats.
sprand(m,n,d)rand(m,n)Creates a m-by-n random matrix (of density d) with iid non-zero elements distributed uniformly on the half-open interval \([0, 1)\).
sprandn(m,n,d)randn(m,n)Creates a m-by-n random matrix (of density d) with iid non-zero elements distributed according to the standard normal (Gaussian) distribution.
sprandn(m,n,d,X)randn(m,n,X)Creates a m-by-n random matrix (of density d) with iid non-zero elements distributed according to the X distribution. (Requires the Distributions package.)