Edited, memorised or added to reading queue

on 27-Mar-2018 (Tue)

Do you want BuboFlash to help you learning these things? Click here to log in or create user.

#matrix-inversion
A common use of the pseudoinverse is to compute a 'best fit' (least squares) solution to a system of linear equations that lacks a unique solution
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Moore–Penrose inverse - Wikipedia
tegral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse. <span>A common use of the pseudoinverse is to compute a 'best fit' (least squares) solution to a system of linear equations that lacks a unique solution (see below under § Applications). Another use is to find the minimum (Euclidean) norm solution to a system of linear equations with multiple solutions. The pseudoinverse facilitates the




#matrix-inversion

The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers. It can be computed using the singular value decomposition.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Moore–Penrose inverse - Wikipedia
tion (see below under § Applications). Another use is to find the minimum (Euclidean) norm solution to a system of linear equations with multiple solutions. The pseudoinverse facilitates the statement and proof of results in linear algebra. <span>The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers. It can be computed using the singular value decomposition. Contents [hide] 1 Notation 2 Definition 3 Properties 3.1 Existence and uniqueness 3.2 Basic properties 3.2.1 Identities 3.3 Reduction to Hermitian case 3.4 Products 3.5




#matrix-inversion

For , a pseudoinverse of is defined as a matrix satisfying all of the following four criteria:

  1. ( AA+ need not be the general identity matrix, but it maps all column vectors of A to themselves);
  2. ( A+ is a weak inverse for the multiplicative semigroup);
  3. ( AA+ is Hermitian); and
  4. ( A+A is also Hermitian).

Moore-Penrose Pseudo-inverse

exists for any matrix , but when the latter has full rank, can be expressed as a simple algebraic formula.

In particular, when has linearly independent columns (and thus matrix is invertible), can be computed as:

...
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Moore–Penrose inverse - Wikipedia
; K ) {\displaystyle I_{n}\in \mathrm {M} (n,n;K)} denotes the n × n {\displaystyle n\times n} identity matrix. Definition[edit source] <span>For A ∈ M ( m , n ; K ) {\displaystyle A\in \mathrm {M} (m,n;K)} , a pseudoinverse of A {\displaystyle A} is defined as a matrix A + ∈ M ( n , m ; K ) {\displaystyle A^{+}\in \mathrm {M} (n,m;K)} satisfying all of the following four criteria: [8] [9] A A + A = A {\displaystyle AA^{+}A=A\,\!} (AA + need not be the general identity matrix, but it maps all column vectors of A to themselves); A + A A + = A + {\displaystyle A^{+}AA^{+}=A^{+}\,\!} (A + is a weak inverse for the multiplicative semigroup); ( A A + ) ∗ = A A + {\displaystyle (AA^{+})^{*}=AA^{+}\,\!} (AA + is Hermitian); and ( A + A ) ∗ = A + A {\displaystyle (A^{+}A)^{*}=A^{+}A\,\!} (A + A is also Hermitian). A + {\displaystyle A^{+}} exists for any matrix A {\displaystyle A} , but when the latter has full rank, A + {\displaystyle A^{+}} can be expressed as a simple algebraic formula. In particular, when A {\displaystyle A} has linearly independent columns (and thus matrix A ∗ A {\displaystyle A^{*}A} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = ( A ∗ A ) − 1 A ∗ . {\displaystyle A^{+}=(A^{*}A)^{-1}A^{*}\,.} This particular pseudoinverse constitutes a left inverse, since, in this case, A + A = I {\displaystyle A^{+}A=I} . When A {\displaystyle A} has linearly independent rows (matrix A A ∗ {\displaystyle AA^{*}} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = A ∗ ( A A ∗ ) − 1 . {\displaystyle A^{+}=A^{*}(AA^{*})^{-1}\,.} This is a right inverse, as A A + = I {\displaystyle AA^{+}=I} . Properties[edit source] Proofs for some of these facts may be found on a separate page Proofs involving the Moore–Penrose inverse. Existence and uniqueness[edit source] The pseu




#linear-algebra #matrix-decomposition

In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Eigendecomposition of a matrix - Wikipedia
| ocultar ahora Eigendecomposition of a matrix From Wikipedia, the free encyclopedia (Redirected from Eigendecomposition) Jump to: navigation, search <span>In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. Contents [hide] 1 Fundamental theory of matrix eigenvectors and eigenvalues 2 Eigendecomposition of a matrix 2.1 Example 2.2 Matrix inverse via eigendecomposition 2.2.1 Pr




#linear-algebra #matrix-decomposition
The eigendecomposition can be derived from the fundamental property of eigenvectors: and thus which yields .
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Eigendecomposition of a matrix - Wikipedia
, {\displaystyle v_{i}\,\,(i=1,\dots ,N),} can also be used as the columns of Q. That can be understood by noting that the magnitude of the eigenvectors in Q gets canceled in the decomposition by the presence of Q −1 . <span>The decomposition can be derived from the fundamental property of eigenvectors: A v = λ v {\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {v} } and thus A Q = Q Λ {\displaystyle \mathbf {A} \mathbf {Q} =\mathbf {Q} \mathbf {\Lambda } } which yields A = Q Λ Q − 1 {\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}} . Example[edit source] Taking a 2 × 2 real matrix A = [




#gaussian-process
Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
f them is normally distributed. The distribution of a Gaussian process is the joint distribution of all those (infinitely many) random variables, and as such, it is a distribution over functions with a continuous domain, e.g. time or space. <span>Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data. The prediction is not just an estimate for that point, but also has uncertainty information—it is a one-dimensional Gaussian distribution (which is the marginal distribution at that poi




#gaussian-process
A key fact of Gaussian processes is that they can be completely defined by their second-order statistics.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
μ ℓ {\displaystyle \mu _{\ell }} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] <span>A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of t




#gaussian-process
if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, <span>if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] Stationarity refers to the process' beha




#gaussian-process
Stationarity refers to the process' behaviour regarding the separation of any two points x and x' . If the process is stationary, it depends on their separation, xx', while if non-stationary it depends on the actual position of the points x and x'.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
initeness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] <span>Stationarity refers to the process' behaviour regarding the separation of any two points x and x' . If the process is stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. If the process depends only on |x − x'|, the Euclidean distance (not the dire




#gaussian-process

If the process depends only on |xx'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous;[7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne




#gaussian-process
If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of continuity is present.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
r the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] <span>If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of continuity is present. If we wish to allow for significant displacement then we might choose a rougher covariance function. Extreme examples of the behaviour is the Ornstein–Uhlenbeck covariance function and




#gaussian-process

Periodicity refers to inducing periodic patterns within the behaviour of the process. Formally, this is achieved by mapping the input x to a two dimensional vector u(x) = (cos(x), sin(x)).

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
en we might choose a rougher covariance function. Extreme examples of the behaviour is the Ornstein–Uhlenbeck covariance function and the squared exponential where the former is never differentiable and the latter infinitely differentiable. <span>Periodicity refers to inducing periodic patterns within the behaviour of the process. Formally, this is achieved by mapping the input x to a two dimensional vector u(x) = (cos(x), sin(x)). Usual covariance functions[edit source] [imagelink] The effect of choosing different kernels on the prior function distribution of the Gaussian process. Left is a squared expon




#gaussian-process
Importantly, a complicated covariance function can be defined as a linear combination of other simpler covariance functions in order to incorporate different insights about the data-set at hand.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gaussian process - Wikipedia
{\displaystyle \nu } and Γ ( ν ) {\displaystyle \Gamma (\nu )} is the gamma function evaluated at ν {\displaystyle \nu } . <span>Importantly, a complicated covariance function can be defined as a linear combination of other simpler covariance functions in order to incorporate different insights about the data-set at hand. Clearly, the inferential results are dependent on the values of the hyperparameters θ (e.g. ℓ and σ) defining the model's behaviour. A popular choice for θ is to provide maximum a pos




Flashcard 1729609469196

Tags
#gaussian-process
Question
Importantly, a complicated covariance function can be defined as a [...] of other simpler covariance functions in order to incorporate different insights about the data-set at hand.
Answer
linear combination

Perhaps more than linear combination.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Importantly, a complicated covariance function can be defined as a linear combination of other simpler covariance functions in order to incorporate different insights about the data-set at hand.

Original toplevel document

Gaussian process - Wikipedia
{\displaystyle \nu } and Γ ( ν ) {\displaystyle \Gamma (\nu )} is the gamma function evaluated at ν {\displaystyle \nu } . <span>Importantly, a complicated covariance function can be defined as a linear combination of other simpler covariance functions in order to incorporate different insights about the data-set at hand. Clearly, the inferential results are dependent on the values of the hyperparameters θ (e.g. ℓ and σ) defining the model's behaviour. A popular choice for θ is to provide maximum a pos







Flashcard 1729611042060

Tags
#gaussian-process
Question

Periodicity maps the input x to a two dimensional vector [...]

Answer
u(x) = (cos(x), sin(x)).

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Periodicity refers to inducing periodic patterns within the behaviour of the process. Formally, this is achieved by mapping the input x to a two dimensional vector u(x) = (cos(x), sin(x)).

Original toplevel document

Gaussian process - Wikipedia
en we might choose a rougher covariance function. Extreme examples of the behaviour is the Ornstein–Uhlenbeck covariance function and the squared exponential where the former is never differentiable and the latter infinitely differentiable. <span>Periodicity refers to inducing periodic patterns within the behaviour of the process. Formally, this is achieved by mapping the input x to a two dimensional vector u(x) = (cos(x), sin(x)). Usual covariance functions[edit source] [imagelink] The effect of choosing different kernels on the prior function distribution of the Gaussian process. Left is a squared expon







Flashcard 1729612614924

Tags
#gaussian-process
Question
If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of [...] is present.
Answer
continuity

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of continuity is present.

Original toplevel document

Gaussian process - Wikipedia
r the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] <span>If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of continuity is present. If we wish to allow for significant displacement then we might choose a rougher covariance function. Extreme examples of the behaviour is the Ornstein–Uhlenbeck covariance function and







#gaussian-process
If the process depends only on |xx'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the be

Original toplevel document

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne




Flashcard 1729615760652

Tags
#gaussian-process
Question
isotropic process depend only on distance, not [...]
Answer
direction

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic.

Original toplevel document

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne







Flashcard 1729617333516

Tags
#gaussian-process
Question
a stochastic process is called [...] if it depends only on distance but not the direction
Answer
isotropic

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic.

Original toplevel document

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne







#gaussian-process
A process that is concurrently stationary and isotropic is considered to be homogeneous;[7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer.

Original toplevel document

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne




Flashcard 1729620479244

Tags
#gaussian-process
Question
homogeneous process behaves the same regardless the location of [...].
Answer
the observer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of <span>the observer. <span><body><html>

Original toplevel document

Gaussian process - Wikipedia
stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. <span>If the process depends only on |x − x'|, the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be homogeneous; [7] in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. [5] If we expect that for "ne







Flashcard 1729622052108

Tags
#gaussian-process
Question
Stationarity process' behaviour depends on the distance between points, not their [...].
Answer
the actual positions

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
y> Stationarity refers to the process' behaviour regarding the separation of any two points x and x' . If the process is stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. <body><html>

Original toplevel document

Gaussian process - Wikipedia
initeness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] <span>Stationarity refers to the process' behaviour regarding the separation of any two points x and x' . If the process is stationary, it depends on their separation, x − x', while if non-stationary it depends on the actual position of the points x and x'. For example, the special case of an Ornstein–Uhlenbeck process, a Brownian motion process, is stationary. If the process depends only on |x − x'|, the Euclidean distance (not the dire







#gaussian-process
if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion.

Original toplevel document

Gaussian process - Wikipedia
} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, <span>if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] Stationarity refers to the process' beha




Flashcard 1729625197836

Tags
#gaussian-process
Question
if a Gaussian process is assumed to have mean zero, defining [...] completely defines the process' behaviour.
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour.

Original toplevel document

Gaussian process - Wikipedia
} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, <span>if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] Stationarity refers to the process' beha







Flashcard 1729626770700

Tags
#gaussian-process
Question
In Gaussian process, the non-negative definiteness of the covariance function enables its [...] using the Karhunen–Loeve expansion.
Answer
spectral decomposition

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
head> if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. <html>

Original toplevel document

Gaussian process - Wikipedia
} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, <span>if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loeve expansion. Basic aspects that can be defined through the covariance function are the process' stationarity, isotropy, smoothness and periodicity. [5] [6] Stationarity refers to the process' beha







Flashcard 1729628343564

Tags
#gaussian-process
Question
Gaussian processes can be completely defined by their [...].
Answer
second-order statistics

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A key fact of Gaussian processes is that they can be completely defined by their second-order statistics.

Original toplevel document

Gaussian process - Wikipedia
μ ℓ {\displaystyle \mu _{\ell }} can be shown to be the covariances and means of the variables in the process. [3] Covariance functions[edit source] <span>A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. [4] Thus, if a Gaussian process is assumed to have mean zero, defining the covariance function completely defines the process' behaviour. Importantly the non-negative definiteness of t







Flashcard 1729634897164

Tags
#gaussian-process
Question
Viewed as a machine-learning algorithm, a Gaussian process uses [...] and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data.
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data.

Original toplevel document

Gaussian process - Wikipedia
f them is normally distributed. The distribution of a Gaussian process is the joint distribution of all those (infinitely many) random variables, and as such, it is a distribution over functions with a continuous domain, e.g. time or space. <span>Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data. The prediction is not just an estimate for that point, but also has uncertainty information—it is a one-dimensional Gaussian distribution (which is the marginal distribution at that poi







Flashcard 1729636470028

Tags
#gaussian-process
Question
Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of [...] to predict the value for an unseen point from training data.
Answer
the similarity between points

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data.

Original toplevel document

Gaussian process - Wikipedia
f them is normally distributed. The distribution of a Gaussian process is the joint distribution of all those (infinitely many) random variables, and as such, it is a distribution over functions with a continuous domain, e.g. time or space. <span>Viewed as a machine-learning algorithm, a Gaussian process uses lazy learning and a measure of the similarity between points (the kernel function) to predict the value for an unseen point from training data. The prediction is not just an estimate for that point, but also has uncertainty information—it is a one-dimensional Gaussian distribution (which is the marginal distribution at that poi







Flashcard 1729641712908

Tags
#linear-algebra #matrix-decomposition
Question
The eigendecomposition can be derived from [...] and thus which yields .
Answer
the fundamental property of eigenvectors:

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The eigendecomposition can be derived from the fundamental property of eigenvectors: and thus which yields .

Original toplevel document

Eigendecomposition of a matrix - Wikipedia
, {\displaystyle v_{i}\,\,(i=1,\dots ,N),} can also be used as the columns of Q. That can be understood by noting that the magnitude of the eigenvectors in Q gets canceled in the decomposition by the presence of Q −1 . <span>The decomposition can be derived from the fundamental property of eigenvectors: A v = λ v {\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {v} } and thus A Q = Q Λ {\displaystyle \mathbf {A} \mathbf {Q} =\mathbf {Q} \mathbf {\Lambda } } which yields A = Q Λ Q − 1 {\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}} . Example[edit source] Taking a 2 × 2 real matrix A = [







Flashcard 1729643285772

Tags
#linear-algebra #matrix-decomposition
Question
The eigendecomposition decomposes matrix A to [...]
Answer
.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The eigendecomposition can be derived from the fundamental property of eigenvectors: and thus which yields .

Original toplevel document

Eigendecomposition of a matrix - Wikipedia
, {\displaystyle v_{i}\,\,(i=1,\dots ,N),} can also be used as the columns of Q. That can be understood by noting that the magnitude of the eigenvectors in Q gets canceled in the decomposition by the presence of Q −1 . <span>The decomposition can be derived from the fundamental property of eigenvectors: A v = λ v {\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {v} } and thus A Q = Q Λ {\displaystyle \mathbf {A} \mathbf {Q} =\mathbf {Q} \mathbf {\Lambda } } which yields A = Q Λ Q − 1 {\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}} . Example[edit source] Taking a 2 × 2 real matrix A = [







Flashcard 1729646955788

Tags
#linear-algebra #matrix-decomposition
Question

[...] is the factorization of a matrix into a canonical form

Answer
eigendecomposition

Also called spectral decomposition

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonal

Original toplevel document

Eigendecomposition of a matrix - Wikipedia
| ocultar ahora Eigendecomposition of a matrix From Wikipedia, the free encyclopedia (Redirected from Eigendecomposition) Jump to: navigation, search <span>In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. Contents [hide] 1 Fundamental theory of matrix eigenvectors and eigenvalues 2 Eigendecomposition of a matrix 2.1 Example 2.2 Matrix inverse via eigendecomposition 2.2.1 Pr







Flashcard 1729648528652

Tags
#linear-algebra #matrix-decomposition
Question

eigendecomposition is sometimes also called [...]


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in

Original toplevel document

Eigendecomposition of a matrix - Wikipedia
| ocultar ahora Eigendecomposition of a matrix From Wikipedia, the free encyclopedia (Redirected from Eigendecomposition) Jump to: navigation, search <span>In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. Contents [hide] 1 Fundamental theory of matrix eigenvectors and eigenvalues 2 Eigendecomposition of a matrix 2.1 Example 2.2 Matrix inverse via eigendecomposition 2.2.1 Pr







Flashcard 1729650101516

Tags
#linear-algebra #matrix-decomposition
Question

eigendecomposition factorises a matrix into a [...] form

Answer
canonical

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way.

Original toplevel document

Eigendecomposition of a matrix - Wikipedia
| ocultar ahora Eigendecomposition of a matrix From Wikipedia, the free encyclopedia (Redirected from Eigendecomposition) Jump to: navigation, search <span>In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. Contents [hide] 1 Fundamental theory of matrix eigenvectors and eigenvalues 2 Eigendecomposition of a matrix 2.1 Example 2.2 Matrix inverse via eigendecomposition 2.2.1 Pr







Flashcard 1729683393804

Tags
#matrix-inversion
Question

The pseudoinverse can be computed using [...].

Answer
singular value decomposition

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers. It can be computed using the singular value decomposition.

Original toplevel document

Moore–Penrose inverse - Wikipedia
tion (see below under § Applications). Another use is to find the minimum (Euclidean) norm solution to a system of linear equations with multiple solutions. The pseudoinverse facilitates the statement and proof of results in linear algebra. <span>The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers. It can be computed using the singular value decomposition. Contents [hide] 1 Notation 2 Definition 3 Properties 3.1 Existence and uniqueness 3.2 Basic properties 3.2.1 Identities 3.3 Reduction to Hermitian case 3.4 Products 3.5







Flashcard 1729684966668

Tags
#matrix-inversion
Question
A common use of the pseudoinverse is to compute a [...] to a system of linear equations that lacks a unique solution
Answer
'best fit' (least squares) solution

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A common use of the pseudoinverse is to compute a 'best fit' (least squares) solution to a system of linear equations that lacks a unique solution

Original toplevel document

Moore–Penrose inverse - Wikipedia
tegral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse. <span>A common use of the pseudoinverse is to compute a 'best fit' (least squares) solution to a system of linear equations that lacks a unique solution (see below under § Applications). Another use is to find the minimum (Euclidean) norm solution to a system of linear equations with multiple solutions. The pseudoinverse facilitates the







#singular-value-decomposition
In linear algebra, the singular-value decomposition (SVD) generalises the eigendecomposition of a positive semidefinite normal matrix (for example, a symmetric matrix with positive eigenvalues) to any matrix via an extension of the polar decomposition.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Singular-value decomposition - Wikipedia
nto three simple transformations: an initial rotation V ∗ , a scaling Σ along the coordinate axes, and a final rotation U. The lengths σ 1 and σ 2 of the semi-axes of the ellipse are the singular values of M, namely Σ 1,1 and Σ 2,2 . <span>In linear algebra, the singular-value decomposition (SVD) is a factorization of a real or complex matrix. It is the generalization of the eigendecomposition of a positive semidefinite normal matrix (for example, a symmetric matrix with positive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. Formally, the singular-value decomposition of an m × n {\d




#singular-value-decomposition
Formally, the singular-value decomposition of an real or complex matrix is a factorization of the form , where is an real or complex unitary matrix, is a rectangular diagonal matrix with non-negative real numbers on the diagonal, and is an real or complex unitary matrix.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Singular-value decomposition - Wikipedia
ositive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. <span>Formally, the singular-value decomposition of an m × n {\displaystyle m\times n} real or complex matrix M {\displaystyle \mathbf {M} } is a factorization of the form U Σ V ∗ {\displaystyle \mathbf {U\Sigma V^{*}} } , where U {\displaystyle \mathbf {U} } is an m × m {\displaystyle m\times m} real or complex unitary matrix, Σ {\displaystyle \mathbf {\Sigma } } is a m × n {\displaystyle m\times n} rectangular diagonal matrix with non-negative real numbers on the diagonal, and V {\displaystyle \mathbf {V} } is an n × n {\displaystyle n\times n} real or complex unitary matrix. The diagonal entries σ i {\displaystyle \sigma _{i}} of




Flashcard 1729714326796

Tags
#singular-value-decomposition
Question
[...] generalises eigendecomposition of a positive semidefinite normal matrix to any matrix
Answer
singular-value decomposition (SVD)

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In linear algebra, the singular-value decomposition (SVD) generalises the eigendecomposition of a positive semidefinite normal matrix (for example, a symmetric matrix with positive eigenvalues) to any matrix via an extension of the polar deco

Original toplevel document

Singular-value decomposition - Wikipedia
nto three simple transformations: an initial rotation V ∗ , a scaling Σ along the coordinate axes, and a final rotation U. The lengths σ 1 and σ 2 of the semi-axes of the ellipse are the singular values of M, namely Σ 1,1 and Σ 2,2 . <span>In linear algebra, the singular-value decomposition (SVD) is a factorization of a real or complex matrix. It is the generalization of the eigendecomposition of a positive semidefinite normal matrix (for example, a symmetric matrix with positive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. Formally, the singular-value decomposition of an m × n {\d







#singular-value-decomposition
Formally, the singular-value decomposition of an real or complex matrix is a factorization of the form
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
Formally, the singular-value decomposition of an real or complex matrix is a factorization of the form , where is an real or complex unitary matrix, is a rectangular diagonal matrix with non-negative real numbers on the diagonal, and is an real or complex unitary matrix.

Original toplevel document

Singular-value decomposition - Wikipedia
ositive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. <span>Formally, the singular-value decomposition of an m × n {\displaystyle m\times n} real or complex matrix M {\displaystyle \mathbf {M} } is a factorization of the form U Σ V ∗ {\displaystyle \mathbf {U\Sigma V^{*}} } , where U {\displaystyle \mathbf {U} } is an m × m {\displaystyle m\times m} real or complex unitary matrix, Σ {\displaystyle \mathbf {\Sigma } } is a m × n {\displaystyle m\times n} rectangular diagonal matrix with non-negative real numbers on the diagonal, and V {\displaystyle \mathbf {V} } is an n × n {\displaystyle n\times n} real or complex unitary matrix. The diagonal entries σ i {\displaystyle \sigma _{i}} of




Flashcard 1729718258956

Tags
#singular-value-decomposition
Question
singular-value decomposition factorises an matrix M to the form [...]
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Formally, the singular-value decomposition of an real or complex matrix is a factorization of the form

Original toplevel document

Singular-value decomposition - Wikipedia
ositive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. <span>Formally, the singular-value decomposition of an m × n {\displaystyle m\times n} real or complex matrix M {\displaystyle \mathbf {M} } is a factorization of the form U Σ V ∗ {\displaystyle \mathbf {U\Sigma V^{*}} } , where U {\displaystyle \mathbf {U} } is an m × m {\displaystyle m\times m} real or complex unitary matrix, Σ {\displaystyle \mathbf {\Sigma } } is a m × n {\displaystyle m\times n} rectangular diagonal matrix with non-negative real numbers on the diagonal, and V {\displaystyle \mathbf {V} } is an n × n {\displaystyle n\times n} real or complex unitary matrix. The diagonal entries σ i {\displaystyle \sigma _{i}} of







Flashcard 1729720618252

Tags
#singular-value-decomposition
Question
With a factorization of the form , , , represent [...]
Answer
real or complex unitary matrix
real or complex unitary matrix.
rectangular diagonal matrix

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Formally, the singular-value decomposition of an real or complex matrix is a factorization of the form , where is an real or complex unitary matrix, is a rectangular diagonal matrix with non-negative real numbers on the diagonal, and is an real or complex unitary matrix.

Original toplevel document

Singular-value decomposition - Wikipedia
ositive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics. <span>Formally, the singular-value decomposition of an m × n {\displaystyle m\times n} real or complex matrix M {\displaystyle \mathbf {M} } is a factorization of the form U Σ V ∗ {\displaystyle \mathbf {U\Sigma V^{*}} } , where U {\displaystyle \mathbf {U} } is an m × m {\displaystyle m\times m} real or complex unitary matrix, Σ {\displaystyle \mathbf {\Sigma } } is a m × n {\displaystyle m\times n} rectangular diagonal matrix with non-negative real numbers on the diagonal, and V {\displaystyle \mathbf {V} } is an n × n {\displaystyle n\times n} real or complex unitary matrix. The diagonal entries σ i {\displaystyle \sigma _{i}} of







#variational-inference
Variational Bayes can be seen as an extension of the EM (expectation-maximization) algorithm from maximum a posteriori estimation (MAP estimation) of the single most probable value of each parameter to fully Bayesian estimation which computes (an approximation to) the entire posterior distribution of the parameters and latent variables.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Variational Bayesian methods - Wikipedia
om. In particular, whereas Monte Carlo techniques provide a numerical approximation to the exact posterior using a set of samples, Variational Bayes provides a locally-optimal, exact analytical solution to an approximation of the posterior. <span>Variational Bayes can be seen as an extension of the EM (expectation-maximization) algorithm from maximum a posteriori estimation (MAP estimation) of the single most probable value of each parameter to fully Bayesian estimation which computes (an approximation to) the entire posterior distribution of the parameters and latent variables. As in EM, it finds a set of optimal parameter values, and it has the same alternating structure as does EM, based on a set of interlocked (mutually dependent) equations that cannot be s




Flashcard 1730164690188

Tags
#variational-inference
Question
[...] can be seen as an extension of the expectation-maximization algorithm
Answer
Variational inference

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Variational Bayes can be seen as an extension of the EM (expectation-maximization) algorithm from maximum a posteriori estimation (MAP estimation) of the single most probable value of each parameter to f

Original toplevel document

Variational Bayesian methods - Wikipedia
om. In particular, whereas Monte Carlo techniques provide a numerical approximation to the exact posterior using a set of samples, Variational Bayes provides a locally-optimal, exact analytical solution to an approximation of the posterior. <span>Variational Bayes can be seen as an extension of the EM (expectation-maximization) algorithm from maximum a posteriori estimation (MAP estimation) of the single most probable value of each parameter to fully Bayesian estimation which computes (an approximation to) the entire posterior distribution of the parameters and latent variables. As in EM, it finds a set of optimal parameter values, and it has the same alternating structure as does EM, based on a set of interlocked (mutually dependent) equations that cannot be s







#variational-inference
Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Variational Bayesian methods - Wikipedia
f references, but its sources remain unclear because it has insufficient inline citations. Please help to improve this article by introducing more precise citations. (September 2010) (Learn how and when to remove this template message) <span>Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables (usually termed "data") as well as unknown parameters and latent variables, with various




#variational-inference

Variational Bayesian methods are primarily used for two purposes:

  1. To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
  2. To derive a lower bound for the marginal likelihood (sometimes called the "evidence") of the observed data (i.e. the marginal probability of the data given the model, with marginalization performed over unobserved variables). This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data. (See also the Bayes factor article.)
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Variational Bayesian methods - Wikipedia
s sorts of relationships among the three types of random variables, as might be described by a graphical model. As is typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". <span>Variational Bayesian methods are primarily used for two purposes: To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables. To derive a lower bound for the marginal likelihood (sometimes called the "evidence") of the observed data (i.e. the marginal probability of the data given the model, with marginalization performed over unobserved variables). This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data. (See also the Bayes factor article.) In the former purpose (that of approximating a posterior probability), variational Bayes is an alternative to Monte Carlo sampling methods — particularly, Markov chain Monte Carlo met




#deep-gaussian-process
DGPs can perform input warping or dimensionality compression or expansion, and automatically learn to construct a kernel that works well for the data at hand. As a result, learning in this model provides a flexible form of Bayesian kernel design.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

pdf

cannot see any pdfs




#variational-inference
For many applications, variational Bayes produces solutions of comparable accuracy to Gibbs sampling at greater speed.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Variational Bayesian methods - Wikipedia
he parameters and latent variables. As in EM, it finds a set of optimal parameter values, and it has the same alternating structure as does EM, based on a set of interlocked (mutually dependent) equations that cannot be solved analytically. <span>For many applications, variational Bayes produces solutions of comparable accuracy to Gibbs sampling at greater speed. However, deriving the set of equations used to iteratively update the parameters often requires a large amount of work compared with deriving the comparable Gibbs sampling equations. Th




#singular-value-decomposition
SVD as change of coordinates

The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : KnKm one can find orthonormal bases of Kn and Km such that T maps the i-th basis vector of Kn to a non-negative multiple of the i-th basis vector of Km , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Singular-value decomposition - Wikipedia
m , n ) , {\displaystyle T(\mathbf {V} _{i})=\sigma _{i}\mathbf {U} _{i},\qquad i=1,\ldots ,\min(m,n),} where σ i is the i-th diagonal entry of Σ, and T(V i ) = 0 for i > min(m,n). <span>The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries. To get a more visual flavour of singular values and SVD factorization — at least when working on real vector spaces — consider the sphere S of radius one in R n . The linear map T map




#matrix-inversion
A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition.[1][9][15] If is the singular value decomposition of A , then . For a rectangular diagonal matrix such as Σ {\displaystyle \Sigma } , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in place, and then transposing the matrix. In numerical computation, only elements larger than some small tolerance are taken to be nonzero, and the others are replaced by zeros. For example, in the MATLAB, GNU Octave, or NumPy function pinv , the tolerance is taken to be t = ε⋅max(m,n)⋅max(Σ) , where ε is the machine epsilon.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Moore–Penrose inverse - Wikipedia
A {\displaystyle A} and A ∗ {\displaystyle A^{*}} . Singular value decomposition (SVD)[edit source] <span>A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition. [1] [9] [15] If A = U Σ V ∗ {\displaystyle A=U\Sigma V^{*}} is the singular value decomposition of A, then A + = V Σ + U ∗ {\displaystyle A^{+}=V\Sigma ^{+}U^{*}} . For a rectangular diagonal matrix such as Σ {\displaystyle \Sigma } , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in place, and then transposing the matrix. In numerical computation, only elements larger than some small tolerance are taken to be nonzero, and the others are replaced by zeros. For example, in the MATLAB, GNU Octave, or NumPy function pinv , the tolerance is taken to be t = ε⋅max(m,n)⋅max(Σ), where ε is the machine epsilon. The computational cost of this method is dominated by the cost of computing the SVD, which is several times higher than matrix–matrix multiplication, even if a state-of-the art implem




#matrix-inversion

Moore-Penrose Pseudo-inverse

exists for any matrix , but when the latter has full rank, can be expressed as a simple algebraic formula.

In particular, when has linearly independent columns (and thus matrix is invertible), can be computed as:

This particular pseudoinverse constitutes a left inverse, since, in this case, .

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
3; ( AA + need not be the general identity matrix, but it maps all column vectors of A to themselves); ( A + is a weak inverse for the multiplicative semigroup); ( AA + is Hermitian); and ( A + A is also Hermitian). <span>Moore-Penrose Pseudo-inverse exists for any matrix , but when the latter has full rank, can be expressed as a simple algebraic formula. In particular, when has linearly independent columns (and thus matrix is invertible), can be computed as: This particular pseudoinverse constitutes a left inverse, since, in this case, . When has linearly independent rows (matrix is invertible), can be computed as: This is a right inverse, as . <span><body><html>

Original toplevel document

Moore–Penrose inverse - Wikipedia
; K ) {\displaystyle I_{n}\in \mathrm {M} (n,n;K)} denotes the n × n {\displaystyle n\times n} identity matrix. Definition[edit source] <span>For A ∈ M ( m , n ; K ) {\displaystyle A\in \mathrm {M} (m,n;K)} , a pseudoinverse of A {\displaystyle A} is defined as a matrix A + ∈ M ( n , m ; K ) {\displaystyle A^{+}\in \mathrm {M} (n,m;K)} satisfying all of the following four criteria: [8] [9] A A + A = A {\displaystyle AA^{+}A=A\,\!} (AA + need not be the general identity matrix, but it maps all column vectors of A to themselves); A + A A + = A + {\displaystyle A^{+}AA^{+}=A^{+}\,\!} (A + is a weak inverse for the multiplicative semigroup); ( A A + ) ∗ = A A + {\displaystyle (AA^{+})^{*}=AA^{+}\,\!} (AA + is Hermitian); and ( A + A ) ∗ = A + A {\displaystyle (A^{+}A)^{*}=A^{+}A\,\!} (A + A is also Hermitian). A + {\displaystyle A^{+}} exists for any matrix A {\displaystyle A} , but when the latter has full rank, A + {\displaystyle A^{+}} can be expressed as a simple algebraic formula. In particular, when A {\displaystyle A} has linearly independent columns (and thus matrix A ∗ A {\displaystyle A^{*}A} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = ( A ∗ A ) − 1 A ∗ . {\displaystyle A^{+}=(A^{*}A)^{-1}A^{*}\,.} This particular pseudoinverse constitutes a left inverse, since, in this case, A + A = I {\displaystyle A^{+}A=I} . When A {\displaystyle A} has linearly independent rows (matrix A A ∗ {\displaystyle AA^{*}} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = A ∗ ( A A ∗ ) − 1 . {\displaystyle A^{+}=A^{*}(AA^{*})^{-1}\,.} This is a right inverse, as A A + = I {\displaystyle AA^{+}=I} . Properties[edit source] Proofs for some of these facts may be found on a separate page Proofs involving the Moore–Penrose inverse. Existence and uniqueness[edit source] The pseu




Flashcard 1731444477196

Tags
#matrix-inversion
Question

when has [...] the Moore-Penrose inverse is a left inverse

Answer
linearly independent columns

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ml> Moore-Penrose Pseudo-inverse exists for any matrix , but when the latter has full rank, can be expressed as a simple algebraic formula. In particular, when has linearly independent columns (and thus matrix is invertible), can be computed as: This particular pseudoinverse constitutes a left inverse, since, in this case, . <html>

Original toplevel document

Moore–Penrose inverse - Wikipedia
; K ) {\displaystyle I_{n}\in \mathrm {M} (n,n;K)} denotes the n × n {\displaystyle n\times n} identity matrix. Definition[edit source] <span>For A ∈ M ( m , n ; K ) {\displaystyle A\in \mathrm {M} (m,n;K)} , a pseudoinverse of A {\displaystyle A} is defined as a matrix A + ∈ M ( n , m ; K ) {\displaystyle A^{+}\in \mathrm {M} (n,m;K)} satisfying all of the following four criteria: [8] [9] A A + A = A {\displaystyle AA^{+}A=A\,\!} (AA + need not be the general identity matrix, but it maps all column vectors of A to themselves); A + A A + = A + {\displaystyle A^{+}AA^{+}=A^{+}\,\!} (A + is a weak inverse for the multiplicative semigroup); ( A A + ) ∗ = A A + {\displaystyle (AA^{+})^{*}=AA^{+}\,\!} (AA + is Hermitian); and ( A + A ) ∗ = A + A {\displaystyle (A^{+}A)^{*}=A^{+}A\,\!} (A + A is also Hermitian). A + {\displaystyle A^{+}} exists for any matrix A {\displaystyle A} , but when the latter has full rank, A + {\displaystyle A^{+}} can be expressed as a simple algebraic formula. In particular, when A {\displaystyle A} has linearly independent columns (and thus matrix A ∗ A {\displaystyle A^{*}A} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = ( A ∗ A ) − 1 A ∗ . {\displaystyle A^{+}=(A^{*}A)^{-1}A^{*}\,.} This particular pseudoinverse constitutes a left inverse, since, in this case, A + A = I {\displaystyle A^{+}A=I} . When A {\displaystyle A} has linearly independent rows (matrix A A ∗ {\displaystyle AA^{*}} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = A ∗ ( A A ∗ ) − 1 . {\displaystyle A^{+}=A^{*}(AA^{*})^{-1}\,.} This is a right inverse, as A A + = I {\displaystyle AA^{+}=I} . Properties[edit source] Proofs for some of these facts may be found on a separate page Proofs involving the Moore–Penrose inverse. Existence and uniqueness[edit source] The pseu







Flashcard 1731448147212

Tags
#matrix-inversion
Question

The left Moore-Penrose Pseudo-inverse is [...]

Answer

This is the one for linear models


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
e exists for any matrix , but when the latter has full rank, can be expressed as a simple algebraic formula. In particular, when has linearly independent columns (and thus matrix is invertible), can be computed as<span>: This particular pseudoinverse constitutes a left inverse, since, in this case, . <span><body><html>

Original toplevel document

Moore–Penrose inverse - Wikipedia
; K ) {\displaystyle I_{n}\in \mathrm {M} (n,n;K)} denotes the n × n {\displaystyle n\times n} identity matrix. Definition[edit source] <span>For A ∈ M ( m , n ; K ) {\displaystyle A\in \mathrm {M} (m,n;K)} , a pseudoinverse of A {\displaystyle A} is defined as a matrix A + ∈ M ( n , m ; K ) {\displaystyle A^{+}\in \mathrm {M} (n,m;K)} satisfying all of the following four criteria: [8] [9] A A + A = A {\displaystyle AA^{+}A=A\,\!} (AA + need not be the general identity matrix, but it maps all column vectors of A to themselves); A + A A + = A + {\displaystyle A^{+}AA^{+}=A^{+}\,\!} (A + is a weak inverse for the multiplicative semigroup); ( A A + ) ∗ = A A + {\displaystyle (AA^{+})^{*}=AA^{+}\,\!} (AA + is Hermitian); and ( A + A ) ∗ = A + A {\displaystyle (A^{+}A)^{*}=A^{+}A\,\!} (A + A is also Hermitian). A + {\displaystyle A^{+}} exists for any matrix A {\displaystyle A} , but when the latter has full rank, A + {\displaystyle A^{+}} can be expressed as a simple algebraic formula. In particular, when A {\displaystyle A} has linearly independent columns (and thus matrix A ∗ A {\displaystyle A^{*}A} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = ( A ∗ A ) − 1 A ∗ . {\displaystyle A^{+}=(A^{*}A)^{-1}A^{*}\,.} This particular pseudoinverse constitutes a left inverse, since, in this case, A + A = I {\displaystyle A^{+}A=I} . When A {\displaystyle A} has linearly independent rows (matrix A A ∗ {\displaystyle AA^{*}} is invertible), A + {\displaystyle A^{+}} can be computed as: A + = A ∗ ( A A ∗ ) − 1 . {\displaystyle A^{+}=A^{*}(AA^{*})^{-1}\,.} This is a right inverse, as A A + = I {\displaystyle AA^{+}=I} . Properties[edit source] Proofs for some of these facts may be found on a separate page Proofs involving the Moore–Penrose inverse. Existence and uniqueness[edit source] The pseu







Flashcard 1731451292940

Tags
#singular-value-decomposition
Question
geometrically SVD finds [...] for every linear map T : KnKm
Answer
orthonormal bases of Kn and Km

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
SVD as change of coordinates The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases,

Original toplevel document

Singular-value decomposition - Wikipedia
m , n ) , {\displaystyle T(\mathbf {V} _{i})=\sigma _{i}\mathbf {U} _{i},\qquad i=1,\ldots ,\min(m,n),} where σ i is the i-th diagonal entry of Σ, and T(V i ) = 0 for i > min(m,n). <span>The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries. To get a more visual flavour of singular values and SVD factorization — at least when working on real vector spaces — consider the sphere S of radius one in R n . The linear map T map







Flashcard 1731453127948

Tags
#singular-value-decomposition
Question
Geometrically SVD finds orthonormal bases of Kn and Km for every linear map T : KnKm such that T maps the i-th basis vector of Kn to a non-negative multiple of the i-th basis vector of Km , and sends the left-over basis vectors to [...].
Answer
zero

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ollows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to <span>zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries. <span><body><html>

Original toplevel document

Singular-value decomposition - Wikipedia
m , n ) , {\displaystyle T(\mathbf {V} _{i})=\sigma _{i}\mathbf {U} _{i},\qquad i=1,\ldots ,\min(m,n),} where σ i is the i-th diagonal entry of Σ, and T(V i ) = 0 for i > min(m,n). <span>The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries. To get a more visual flavour of singular values and SVD factorization — at least when working on real vector spaces — consider the sphere S of radius one in R n . The linear map T map







Flashcard 1731454700812

Tags
#singular-value-decomposition
Question
With SVD geometrically every linear map T : KnKm is represented by a diagonal matrix with [...] entries.
Answer
non-negative real diagonal

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
h that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with <span>non-negative real diagonal entries. <span><body><html>

Original toplevel document

Singular-value decomposition - Wikipedia
m , n ) , {\displaystyle T(\mathbf {V} _{i})=\sigma _{i}\mathbf {U} _{i},\qquad i=1,\ldots ,\min(m,n),} where σ i is the i-th diagonal entry of Σ, and T(V i ) = 0 for i > min(m,n). <span>The geometric content of the SVD theorem can thus be summarized as follows: for every linear map T : K n → K m one can find orthonormal bases of K n and K m such that T maps the i-th basis vector of K n to a non-negative multiple of the i-th basis vector of K m , and sends the left-over basis vectors to zero. With respect to these bases, the map T is therefore represented by a diagonal matrix with non-negative real diagonal entries. To get a more visual flavour of singular values and SVD factorization — at least when working on real vector spaces — consider the sphere S of radius one in R n . The linear map T map







#logic
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara




Flashcard 1731479866636

Tags
#spanish
Question
use conditional simple to [...]:

¿Podrías pasarme ese plato, por favor?
Answer
ask politely

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Verbs: The Conditional Simple Usage: To ask politely: ¿Podrías pasarme ese plato, por favor? (Could you pass me that plate, please?).

Original toplevel document

Open it
Verbs: The Conditional Simple Usage: To ask politely: ¿Podrías pasarme ese plato, por favor? (Could you pass me that plate, please?). To express wishes: ¡Me encantaría ir de viaje a Australia! (I would love to go on a trip to Australia!). To suggest: Creo que deberías ir al médico a verte ese dolor de espalda (I think







Flashcard 1731508440332

Tags
#logic
Question
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were [...], the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.
Answer
the ancient Greek period

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara







Flashcard 1731510013196

Tags
#logic
Question
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, [...], and the mathematical period of the 19th and 20th centuries.
Answer
the medieval scholastic period

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara







Flashcard 1731511586060

Tags
#logic
Question
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and [...] of the 19th and 20th centuries.
Answer
the mathematical period

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara







#matrix-inversion
A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition.[1][9][15] If is the singular value decomposition of A , then .
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition. [1] [9] [15] If is the singular value decomposition of A , then . For a rectangular diagonal matrix such as Σ {\displaystyle \Sigma } , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in p

Original toplevel document

Moore–Penrose inverse - Wikipedia
A {\displaystyle A} and A ∗ {\displaystyle A^{*}} . Singular value decomposition (SVD)[edit source] <span>A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition. [1] [9] [15] If A = U Σ V ∗ {\displaystyle A=U\Sigma V^{*}} is the singular value decomposition of A, then A + = V Σ + U ∗ {\displaystyle A^{+}=V\Sigma ^{+}U^{*}} . For a rectangular diagonal matrix such as Σ {\displaystyle \Sigma } , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in place, and then transposing the matrix. In numerical computation, only elements larger than some small tolerance are taken to be nonzero, and the others are replaced by zeros. For example, in the MATLAB, GNU Octave, or NumPy function pinv , the tolerance is taken to be t = ε⋅max(m,n)⋅max(Σ), where ε is the machine epsilon. The computational cost of this method is dominated by the cost of computing the SVD, which is several times higher than matrix–matrix multiplication, even if a state-of-the art implem




Flashcard 1731520498956

Tags
#matrix-inversion
Question
If is the singular value decomposition of A , then the pseudoinverse of A is [...]
Answer
.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition. [1] [9] [15] If is the singular value decomposition of A , then .

Original toplevel document

Moore–Penrose inverse - Wikipedia
A {\displaystyle A} and A ∗ {\displaystyle A^{*}} . Singular value decomposition (SVD)[edit source] <span>A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition. [1] [9] [15] If A = U Σ V ∗ {\displaystyle A=U\Sigma V^{*}} is the singular value decomposition of A, then A + = V Σ + U ∗ {\displaystyle A^{+}=V\Sigma ^{+}U^{*}} . For a rectangular diagonal matrix such as Σ {\displaystyle \Sigma } , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in place, and then transposing the matrix. In numerical computation, only elements larger than some small tolerance are taken to be nonzero, and the others are replaced by zeros. For example, in the MATLAB, GNU Octave, or NumPy function pinv , the tolerance is taken to be t = ε⋅max(m,n)⋅max(Σ), where ε is the machine epsilon. The computational cost of this method is dominated by the cost of computing the SVD, which is several times higher than matrix–matrix multiplication, even if a state-of-the art implem







Flashcard 1731525217548

Tags
#deep-gaussian-process
Question
pseudo datapoint based approximation methods for DGPs has a computational complexity of [...]
Answer
\(O(NLM^ 2 ) \)

where N is the number of datapoints, L is the number of layers, and M is the number of pseudo datapoints.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
pseudo datapoint based approximation methods for DGPs trade model complexity for a lower computational complexity of \(O(NLM^ 2 ) \) where N is the number of datapoints, L is the number of layers, and M is the number of pseudo datapoints.

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 1731526790412

Tags
#deep-gaussian-process
Question
DGPs can perform [...] or dimensionality compression or expansion
Answer
input warping

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
DGPs can perform input warping or dimensionality compression or expansion, and automatically learn to construct a kernel that works well for the data at hand. As a result, learning in this model provides a flexible f

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 1731528363276

Tags
#deep-gaussian-process
Question
DGPs can perform input warping or [...]
Answer
dimensionality compression or expansion

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
DGPs can perform input warping or dimensionality compression or expansion, and automatically learn to construct a kernel that works well for the data at hand. As a result, learning in this model provides a flexible form of Bayesian kernel design. </sp

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 1731529936140

Tags
#deep-gaussian-process
Question
DGPs can automatically learn to [...] that works well for the data at hand.
Answer
construct a kernel

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
DGPs can perform input warping or dimensionality compression or expansion, and automatically learn to construct a kernel that works well for the data at hand. As a result, learning in this model provides a flexible form of Bayesian kernel design.

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 1731535703308

Tags
#variational-inference
Question
Variational Bayesian methods are a family of techniques for approximating [...] arising in Bayesian inference and machine learning.
Answer
intractable integrals

In Bayesian inference this manifests as calculating marginal posteriors

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning.

Original toplevel document

Variational Bayesian methods - Wikipedia
f references, but its sources remain unclear because it has insufficient inline citations. Please help to improve this article by introducing more precise citations. (September 2010) (Learn how and when to remove this template message) <span>Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables (usually termed "data") as well as unknown parameters and latent variables, with various







#dynamic-programming
In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by
  1. breaking it down into a collection of simpler subproblems,
  2. solving each of those subproblems just once, and
  3. storing their solutions.
The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space.
(Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.)
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna




Flashcard 1731673328908

Tags
#dynamic-programming
Question
The first step in dynamic programming it to [...]
Answer
break the problem down into a collection of simpler subproblems

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
head><head> In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the p

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731674901772

Tags
#dynamic-programming
Question
dynamic programming aims to solve each subproblems [...] and storing their solutions.
Answer
just once

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ter science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, <span>solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving c

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731676474636

Tags
#dynamic-programming
Question
dynamic programming solves each of the subproblems just once, and [...].
Answer
storing their solutions

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
nomics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and <span>storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expens

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731678047500

Tags
#dynamic-programming
Question
In dynamic programming, when the same subproblem occurs, instead of [...], one simply looks up the previously computed solution
Answer
recomputing its solution

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of <span>recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731679620364

Tags
#dynamic-programming
Question
In dynamic programming, the next time when the same subproblem occurs, instead of recomputing its solution, one simply [...]
Answer
looks up the previously computed solution

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
blem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply <span>looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on t

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731681455372

Tags
#dynamic-programming
Question
The purpose of dynamic programming is to [...].
Answer
trade space for time

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby <span>saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) <span><body><html>

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731683028236

Tags
#dynamic-programming
Question
In dynamic programming each of the subproblem solutions is [...], typically based on the values of its input parameters, so as to facilitate its lookup.
Answer
indexed in some way

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is <span>indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) <span><body><html>

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







Flashcard 1731684601100

Tags
#dynamic-programming
Question
subproblem solutions are typically indexed by [...] to facilitate lookup.
Answer
input values

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
tion, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on <span>the values of its input parameters, so as to facilitate its lookup.) <span><body><html>

Original toplevel document

Dynamic programming - Wikipedia
This article's factual accuracy is disputed. Relevant discussion may be found on the talk page. Please help to ensure that disputed statements are reliably sourced. (November 2015) (Learn how and when to remove this template message) <span>In computer science, mathematics, management science, economics and bioinformatics, dynamic programming (also known as dynamic optimization) is a method for solving a complex problem by breaking it down into a collection of simpler subproblems, solving each of those subproblems just once, and storing their solutions. The next time the same subproblem occurs, instead of recomputing its solution, one simply looks up the previously computed solution, thereby saving computation time at the expense of a (hopefully) modest expenditure in storage space. (Each of the subproblem solutions is indexed in some way, typically based on the values of its input parameters, so as to facilitate its lookup.) The technique of storing solutions to subproblems instead of recomputing them is called "memoization". Dynamic programming algorithms are often used for optimization. A dyna







#finance
In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Option (finance) - Wikipedia
y Real estate Reinsurance Over-the-counter (off-exchange) Forwards Options Spot market Swaps Trading Participants Regulation Clearing Related areas Banks and banking Finance corporate personal public v t e <span>In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option. The strike price may be set by reference to the spot price (market price) of the underlying security or commodity on the day an option is taken out, or it may be fixed at a discount or




#finance
An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Option (finance) - Wikipedia
or commodity on the day an option is taken out, or it may be fixed at a discount or at a premium. The seller has the corresponding obligation to fulfill the transaction – to sell or buy – if the buyer (owner) "exercises" the option. <span>An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. The seller may grant an option to a buyer as part of another transaction, such as a share issue or as part of an employee incentive scheme, otherwise a buyer would pay a premium to th




Flashcard 1731692465420

Tags
#finance
Question
an option gives the buyer [...], to buy or sell
Answer
the right, but not the obligation

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option.

Original toplevel document

Option (finance) - Wikipedia
y Real estate Reinsurance Over-the-counter (off-exchange) Forwards Options Spot market Swaps Trading Participants Regulation Clearing Related areas Banks and banking Finance corporate personal public v t e <span>In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option. The strike price may be set by reference to the spot price (market price) of the underlying security or commodity on the day an option is taken out, or it may be fixed at a discount or







Flashcard 1731694038284

Tags
#finance
Question
an option gives the buyer the right to [...] at a specified strike price on a specified date, depending on the form of the option.
Answer
buy or sell an underlying asset or instrument

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option.

Original toplevel document

Option (finance) - Wikipedia
y Real estate Reinsurance Over-the-counter (off-exchange) Forwards Options Spot market Swaps Trading Participants Regulation Clearing Related areas Banks and banking Finance corporate personal public v t e <span>In finance, an option is a contract which gives the buyer (the owner or holder of the option) the right, but not the obligation, to buy or sell an underlying asset or instrument at a specified strike price on a specified date, depending on the form of the option. The strike price may be set by reference to the spot price (market price) of the underlying security or commodity on the day an option is taken out, or it may be fixed at a discount or







Flashcard 1731695611148

Tags
#finance
Question
[...] is an option to buy
Answer
a call

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. <

Original toplevel document

Option (finance) - Wikipedia
or commodity on the day an option is taken out, or it may be fixed at a discount or at a premium. The seller has the corresponding obligation to fulfill the transaction – to sell or buy – if the buyer (owner) "exercises" the option. <span>An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. The seller may grant an option to a buyer as part of another transaction, such as a share issue or as part of an employee incentive scheme, otherwise a buyer would pay a premium to th







Flashcard 1731697184012

Tags
#finance
Question
[...] is an option to sell
Answer
a put

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. <html>

Original toplevel document

Option (finance) - Wikipedia
or commodity on the day an option is taken out, or it may be fixed at a discount or at a premium. The seller has the corresponding obligation to fulfill the transaction – to sell or buy – if the buyer (owner) "exercises" the option. <span>An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. The seller may grant an option to a buyer as part of another transaction, such as a share issue or as part of an employee incentive scheme, otherwise a buyer would pay a premium to th







Flashcard 1731698756876

Tags
#finance
Question
the [...] option is more frequently discussed.
Answer
call

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but <span>the call option is more frequently discussed. <span><body><html>

Original toplevel document

Option (finance) - Wikipedia
or commodity on the day an option is taken out, or it may be fixed at a discount or at a premium. The seller has the corresponding obligation to fulfill the transaction – to sell or buy – if the buyer (owner) "exercises" the option. <span>An option that conveys to the owner the right to buy at a specific price is referred to as a call; an option that conveys the right of the owner to sell at a specific price is referred to as a put. Both are commonly traded, but the call option is more frequently discussed. The seller may grant an option to a buyer as part of another transaction, such as a share issue or as part of an employee incentive scheme, otherwise a buyer would pay a premium to th







#finance
The Black–Scholes / ˌ b l æ k ˈ ʃ oʊ l z / [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments.
  1. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce
  2. the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that
  3. the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate).
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]




Flashcard 1731703737612

Tags
#finance
Question
The Black–Scholes model is a mathematical model of a financial market containing [...] instruments
Answer
derivative investment

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Black–Scholes / ˌ b l æ k ˈ ʃ oʊ l z / [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of E

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







Flashcard 1731706096908

Tags
#finance
Question
mathematically, the Black–Scholes equation is a [...]

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Black–Scholes / ˌ b l æ k ˈ ʃ oʊ l z / [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







Flashcard 1731707669772

Tags
#finance
Question
the Black–Scholes formula estimates the price of [...]

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives <span>a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate)

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







Flashcard 1731709242636

Tags
#finance
Question
under the Black–Scholes–Merton model, the option [...] regardless of the risk of the security and its expected return.
Answer
the option has a unique price

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
nt instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that <span>the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). <span><body><html>

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







Flashcard 1731710815500

Tags
#finance
Question
In the Black–Scholes model the option has a unique price regardless of [...]
Answer
the risk of the security and its expected return

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
rential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of <span>the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). <span><body><html>

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







Flashcard 1731712388364

Tags
#finance
Question
Under the Black–Scholes–Merton model, the option's unique price is decided by [...]
Answer
the risk-neutral rate

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
mula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with <span>the risk-neutral rate). <span><body><html>

Original toplevel document

Black–Scholes model - Wikipedia
Black–Scholes model - Wikipedia Black–Scholes model From Wikipedia, the free encyclopedia (Redirected from Black–Scholes) Jump to: navigation, search The Black–Scholes /ˌblæk ˈʃoʊlz/ [1] or Black–Scholes–Merton model is a mathematical model of a financial market containing derivative investment instruments. From the partial differential equation in the model, known as the Black–Scholes equation, one can deduce the Black–Scholes formula, which gives a theoretical estimate of the price of European-style options and shows that the option has a unique price regardless of the risk of the security and its expected return (instead replacing the security's expected return with the risk-neutral rate). The formula led to a boom in options trading and provided mathematical legitimacy to the activities of the Chicago Board Options Exchange and other options markets around the world. [2]







#stochastics
In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Martingale (probability theory) - Wikipedia
h For the martingale betting strategy, see martingale (betting system). [imagelink] Stopped Brownian motion is an example of a martingale. It can model an even coin-toss betting game with the possibility of bankruptcy. <span>In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. Contents [hide] 1 History 2 Definitions 2.1 Martingale sequences with respect to another sequence 2.2 General definition 3 Examples of martingales 4 Submartingales, super




Flashcard 1731718679820

Tags
#stochastics
Question
for a [...], at a particular time in the realized sequence, the expectation of the next value is equal to the present observed value, even given knowledge of all prior observed values.
Answer
martingale

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to th

Original toplevel document

Martingale (probability theory) - Wikipedia
h For the martingale betting strategy, see martingale (betting system). [imagelink] Stopped Brownian motion is an example of a martingale. It can model an even coin-toss betting game with the possibility of bankruptcy. <span>In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. Contents [hide] 1 History 2 Definitions 2.1 Martingale sequences with respect to another sequence 2.2 General definition 3 Examples of martingales 4 Submartingales, super







Flashcard 1731720252684

Tags
#stochastics
Question
a martingale is a [...]

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior obse

Original toplevel document

Martingale (probability theory) - Wikipedia
h For the martingale betting strategy, see martingale (betting system). [imagelink] Stopped Brownian motion is an example of a martingale. It can model an even coin-toss betting game with the possibility of bankruptcy. <span>In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. Contents [hide] 1 History 2 Definitions 2.1 Martingale sequences with respect to another sequence 2.2 General definition 3 Examples of martingales 4 Submartingales, super







Flashcard 1731721825548

Tags
#stochastics
Question
In a martingale [...] is equal to the present observed value even given knowledge of all prior observed values.
Answer
the expectation of the next value in the sequence

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values.

Original toplevel document

Martingale (probability theory) - Wikipedia
h For the martingale betting strategy, see martingale (betting system). [imagelink] Stopped Brownian motion is an example of a martingale. It can model an even coin-toss betting game with the possibility of bankruptcy. <span>In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. Contents [hide] 1 History 2 Definitions 2.1 Martingale sequences with respect to another sequence 2.2 General definition 3 Examples of martingales 4 Submartingales, super







Flashcard 1731723398412

Tags
#stochastics
Question
In a martingale, the expectation of the next value in the sequence equals to [...]
Answer
the present observed value

even given knowledge of all prior observed values.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. <body><html>

Original toplevel document

Martingale (probability theory) - Wikipedia
h For the martingale betting strategy, see martingale (betting system). [imagelink] Stopped Brownian motion is an example of a martingale. It can model an even coin-toss betting game with the possibility of bankruptcy. <span>In probability theory, a martingale is a sequence of random variables (i.e., a stochastic process) for which, at a particular time in the realized sequence, the expectation of the next value in the sequence is equal to the present observed value even given knowledge of all prior observed values. Contents [hide] 1 History 2 Definitions 2.1 Martingale sequences with respect to another sequence 2.2 General definition 3 Examples of martingales 4 Submartingales, super







#gauss-markov-process

Gauss–Markov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes.[1][2] The stationary Gauss–Markov process (also known as a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Gauss–Markov process - Wikipedia
translations!] Gauss–Markov process From Wikipedia, the free encyclopedia Jump to: navigation, search Not to be confused with the Gauss–Markov theorem of mathematical statistics. <span>Gauss–Markov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes. [1] [2] The stationary Gauss–Markov process (also known as a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions. Every Gauss–Markov process X(t) possesses the three following properties: If h(t) is a non-zero scalar function of t, then Z(t) = h(t)X(t) is also a Gauss–Markov process If f(t) is




Flashcard 1731745680652

Tags
#gauss-markov-process
Question

Gauss–Markov stochastic processes satisfy the requirements for [...]


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Gauss–Markov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes. [1] [2] The stationary Gauss–Markov process (also known as a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions. </sp

Original toplevel document

Gauss–Markov process - Wikipedia
translations!] Gauss–Markov process From Wikipedia, the free encyclopedia Jump to: navigation, search Not to be confused with the Gauss–Markov theorem of mathematical statistics. <span>Gauss–Markov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes. [1] [2] The stationary Gauss–Markov process (also known as a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions. Every Gauss–Markov process X(t) possesses the three following properties: If h(t) is a non-zero scalar function of t, then Z(t) = h(t)X(t) is also a Gauss–Markov process If f(t) is







Flashcard 1731747253516

Tags
#gauss-markov-process
Question

The stationary Gauss–Markov process is known as [...]


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
kov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes. [1] [2] The stationary Gauss–Markov process (also known as <span>a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions. <span><body><html>

Original toplevel document

Gauss–Markov process - Wikipedia
translations!] Gauss–Markov process From Wikipedia, the free encyclopedia Jump to: navigation, search Not to be confused with the Gauss–Markov theorem of mathematical statistics. <span>Gauss–Markov stochastic processes (named after Carl Friedrich Gauss and Andrey Markov) are stochastic processes that satisfy the requirements for both Gaussian processes and Markov processes. [1] [2] The stationary Gauss–Markov process (also known as a Ornstein–Uhlenbeck process) is a very special case because it is unique, except for some trivial exceptions. Every Gauss–Markov process X(t) possesses the three following properties: If h(t) is a non-zero scalar function of t, then Z(t) = h(t)X(t) is also a Gauss–Markov process If f(t) is







Flashcard 1731762982156

Tags
#spanish
Question
[...] significa ‘en la parte anterior’ y [...] equivale a ‘hacia allá’
Answer
Delante Adelante

Examples
Voy yo delante, que sé el camino.
Sigamos adelante.


La forma alante, por otro lado, es incorrecta.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Delante significa ‘en la parte anterior’, ‘en frente’ o ‘ante alguien’, se usa por lo general para indicar la situación de alguien o algo. Adelante , por su parte, equivale a ‘más allá’, ‘hacia allá’, o ‘hacia enfrente’, y se emplea para indicar la existencia de un movimiento, sea real o figurado. La forma alante , por otro lado, es incorrecta.

Original toplevel document

Delante o adelante - Diccionario de Dudas
amp;cj=1"> [imagelink] Palabras Homófonas Palabras Parónimas Fonética y fonología Uso Grafía Léxicas Ver más Latinismos Extranjerismos Barbarismos Ultracorrecciones Dudas de uso Delante o adelante Delante significa ‘en la parte anterior’, ‘en frente’ o ‘ante alguien’, se usa por lo general para indicar la situación de alguien o algo. Adelante , por su parte, equivale a ‘más allá’, ‘hacia allá’, o ‘hacia enfrente’, y se emplea para indicar la existencia de un movimiento, sea real o figurado. La forma alante , por otro lado, es incorrecta. Cuándo usar delante Delante es un adverbio de lugar; se emplea con el significado de ‘en la parte anterior’, ‘en frente’ o ‘en presencia de alguien’. Por lo general, es un adverbio que







Flashcard 1731771632908

Tags
#gaussian-process
Question
Ornstein–Uhlenbeck covariance function: [...]
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Gaussian process - Wikipedia
2 ) {\displaystyle K_{\operatorname {SE} }(x,x')=\exp {\Big (}-{\frac {\|d\|^{2}}{2\ell ^{2}}}{\Big )}} <span>Ornstein–Uhlenbeck: K OU ( x , x ′ ) = exp ⁡ ( − | d | ℓ ) {\displaystyle K_{\operatorname {OU} }(x,x')=\exp \left(-{\frac {|d|}{\ell }}\right)} Matérn: K Matern ( x , x ′ )







Flashcard 1731774516492

Tags
#gaussian-process
Question
Periodic covariance function: [...]
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Gaussian process - Wikipedia
{\displaystyle K_{\operatorname {Matern} }(x,x')={\frac {2^{1-\nu }}{\Gamma (\nu )}}{\Big (}{\frac {{\sqrt {2\nu }}|d|}{\ell }}{\Big )}^{\nu }K_{\nu }{\Big (}{\frac {{\sqrt {2\nu }}|d|}{\ell }}{\Big )}} <span>Periodic: K P ( x , x ′ ) = exp ⁡ ( − 2 sin 2 ⁡ ( d 2 ) ℓ 2 ) {\displaystyle K_{\operatorname {P} }(x,x')=\exp \left(-{\frac {2\sin ^{2}\left({\frac {d}{2}}\right)}{\ell ^{2}}}\right)} Rational quadratic: K RQ ( x , x ′







Flashcard 1731776875788

Tags
#gaussian-process
Question
Rational quadratic covariance function: [...]
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Gaussian process - Wikipedia
2 ) {\displaystyle K_{\operatorname {P} }(x,x')=\exp \left(-{\frac {2\sin ^{2}\left({\frac {d}{2}}\right)}{\ell ^{2}}}\right)} <span>Rational quadratic: K RQ ( x , x ′ ) = ( 1 + | d | 2 ) − α , α ≥ 0 {\displaystyle K_{\operatorname {RQ} }(x,x')=(1+|d|^{2})^{-\alpha },\quad \alpha \geq 0} Here d = x − x ′ {\displaystyle d=x-x'} . The parameter ℓ is the character







Flashcard 1732343631116

Tags
#logic
Question
J M Bocheński wrote [...] in 1961
Answer
A History of Formal Logic

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centurie

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara







#functional-analysis

Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Functional analysis - Wikipedia
analysis (psychology). [imagelink] One of the possible modes of vibration of an idealized circular drum head. These modes are eigenfunctions of a linear operator on a function space, a common construction in functional analysis. <span>Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations. The usage of the word functional goes back to the calculus of variations, implying a function whose argument is a function and the name was first used in Hadamard's 1910 book on that




#logic
The third of them, in contrast, exemplifies an entirely different rationale for logic, namely as a foundational branch of mathematics
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
grand periods in the history of logic, two of them, the ancient period and the medieval scholastic period, were closely connected to the idea that the primary application of logic is for practices of debating such as dialectical disputations. <span>The third of them, in contrast, exemplifies an entirely different rationale for logic, namely as a foundational branch of mathematics, not in any way connected to the ordinary languages in which debates are typically conducted. The hiatus between the second and third periods can be explained by the fall from grace of




#logic
I n the modern period, a number of philosophers came to see the nature of logic in terms of the faculties of mind.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
rving of the title ‘logic’ than the work done under this heading in the early modern period, given that it comes closer to the level of rigour and formal sophistication that came to be associated with logic from the late 19th century onwards. <span>In the modern period, a number of philosophers came to see the nature of logic in terms of the faculties of mind. To be sure, this is again a theme present in medieval scholastic thought (in the work of the 14th-century author Pierre d’Ailly, for example), but in the early modern period it became t




#logic
It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
ich is thoroughly disputational, with Meditations on First Philosophy (1641) by Descartes, a book argued through long paragraphs driven by the first-person singular. The nature of intellectual enquiry shifted with the downfall of disputation. <span>It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440. Before that, books were a rare commodity, and education was conducted almost exclusively by means of oral contact between masters and pupils in the form of expository lectures in which




#logic
The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
s Diafoirus resorts to disputational vocabulary to make a point about love: Distinguo, Mademoiselle; in all that does not concern the possession of the loved one, concedo, I grant it; but in what does regard that possession, nego, I deny it. <span>The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm. Despite all this, disputations continued to be practised in certain university contexts for some time – indeed, they live on in the ceremonial character of PhD defences. The point, thou




#logic
The tight connection between traditional logic and debating practices dates back to the classical Hellenistic period.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
cond intentions’, roughly what we call second-order concepts, or concepts of concepts. But as late as in the 16th century, the Spanish theologian Domingo de Soto could write with confidence that ‘dialectic is the art or science of disputing’. <span>The tight connection between traditional logic and debating practices dates back to the classical Hellenistic period. Intellectual activity then was quintessentially a dialogical affair, as registered in Plato’s dialogues. In these dialogues, Socrates regularly engages in the practice of refutation (el




#functional-analysis
Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as tr

Original toplevel document

Functional analysis - Wikipedia
analysis (psychology). [imagelink] One of the possible modes of vibration of an idealized circular drum head. These modes are eigenfunctions of a linear operator on a function space, a common construction in functional analysis. <span>Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations. The usage of the word functional goes back to the calculus of variations, implying a function whose argument is a function and the name was first used in Hadamard's 1910 book on that




Flashcard 1732490169612

Tags
#functional-analysis
Question
Functional analysis studies [...] endowed with limit-related structure
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitab

Original toplevel document

Functional analysis - Wikipedia
analysis (psychology). [imagelink] One of the possible modes of vibration of an idealized circular drum head. These modes are eigenfunctions of a linear operator on a function space, a common construction in functional analysis. <span>Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations. The usage of the word functional goes back to the calculus of variations, implying a function whose argument is a function and the name was first used in Hadamard's 1910 book on that







Flashcard 1732491742476

Tags
#functional-analysis
Question
In Functional analysis, the vector spaces are endowed with [...] structures
Answer
limit-related

Norms, inner products, etc.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense.

Original toplevel document

Functional analysis - Wikipedia
analysis (psychology). [imagelink] One of the possible modes of vibration of an idealized circular drum head. These modes are eigenfunctions of a linear operator on a function space, a common construction in functional analysis. <span>Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations. The usage of the word functional goes back to the calculus of variations, implying a function whose argument is a function and the name was first used in Hadamard's 1910 book on that







Flashcard 1732493315340

Tags
#functional-analysis
Question
Functional analysis study vector spaces endowed with limit-related structure and the [...] defined on them.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
span> Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the <span>linear functions defined on these spaces and respecting these structures in a suitable sense. <span><body><html>

Original toplevel document

Functional analysis - Wikipedia
analysis (psychology). [imagelink] One of the possible modes of vibration of an idealized circular drum head. These modes are eigenfunctions of a linear operator on a function space, a common construction in functional analysis. <span>Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (e.g. inner product, norm, topology, etc.) and the linear functions defined on these spaces and respecting these structures in a suitable sense. The historical roots of functional analysis lie in the study of spaces of functions and the formulation of properties of transformations of functions such as the Fourier transform as transformations defining continuous, unitary etc. operators between function spaces. This point of view turned out to be particularly useful for the study of differential and integral equations. The usage of the word functional goes back to the calculus of variations, implying a function whose argument is a function and the name was first used in Hadamard's 1910 book on that







Flashcard 1732605512972

Tags
#linear-algebra #matrix-decomposition
Question

If matrix A can be eigendecomposed and if none of its eigenvalues are zero, then A is [...] and its inverse is given by [...]

Answer
nonsingular

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Eigendecomposition of a matrix - Wikipedia
amp;0\\1&3\\\end{bmatrix}}{\begin{bmatrix}-2c&0\\c&d\\\end{bmatrix}}={\begin{bmatrix}1&0\\0&3\\\end{bmatrix}},[c,d]\in \mathbb {R} } Matrix inverse via eigendecomposition[edit source] Main article: Inverse matrix <span>If matrix A can be eigendecomposed and if none of its eigenvalues are zero, then A is nonsingular and its inverse is given by A − 1 = Q Λ − 1 Q − 1 {\displaystyle \mathbf {A} ^{-1}=\mathbf {Q} \mathbf {\Lambda } ^{-1}\mathbf {Q} ^{-1}} Furthermore, because Λ is a diagonal matrix, its inverse is easy to calculate: [ Λ







Flashcard 1732608658700

Tags
#linear-algebra #matrix-decomposition
Question

[...] has real eigenvalues and the eigenvectors can be chosen to be orthogonal

Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Eigendecomposition of a matrix - Wikipedia
{\displaystyle A=A^{*}} ), which implies that it is also complex normal, the diagonal matrix Λ has only real values, and if A is unitary, Λ takes all its values on the complex unit circle. Real symmetric matrices[edit source] <span>As a special case, for every N×N real symmetric matrix, the eigenvalues are real and the eigenvectors can be chosen such that they are orthogonal to each other. Thus a real symmetric matrix A can be decomposed as A = Q Λ Q T {\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{T}} where Q is an orthogonal matrix, and Λ is a diagonal matrix whose entries are the eigenvalues of A. Useful facts[edit source] Useful facts regarding eigenvalues[edit source] The product of the eigenvalues is equal to the determinant of A det







Flashcard 1735803669772

Tags
#logic
Question
A History of Formal Logic (1961) is written by the distinguished [...]
Answer
J M Bocheński

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
the hands of thinkers such as George Boole, Gottlob Frege, Bertrand Russell, Alfred Tarski and Kurt Gödel, it’s clear that Kant was dead wrong. But he was also wrong in thinking that there had been no progress since Aristotle up to his time. <span>According to A History of Formal Logic (1961) by the distinguished J M Bocheński, the golden periods for logic were the ancient Greek period, the medieval scholastic period, and the mathematical period of the 19th and 20th centuries. (Throughout this piece, the focus is on the logical traditions that emerged against the background of ancient Greek logic. So Indian and Chinese logic are not included, but medieval Ara







Flashcard 1735816514828

Tags
#logic
Question
The fall of [...] culture wasn’t the only cause for the demise of scholastic logic
Answer
disputational

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
s Diafoirus resorts to disputational vocabulary to make a point about love: Distinguo, Mademoiselle; in all that does not concern the possession of the loved one, concedo, I grant it; but in what does regard that possession, nego, I deny it. <span>The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm. Despite all this, disputations continued to be practised in certain university contexts for some time – indeed, they live on in the ceremonial character of PhD defences. The point, thou







Flashcard 1735818087692

Tags
#logic
Question
Scholastic logic was also viewed as being tied to Aristotelian conceptions of [...and...]
Answer
language and metaphysics

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm. <body><html>

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
s Diafoirus resorts to disputational vocabulary to make a point about love: Distinguo, Mademoiselle; in all that does not concern the possession of the loved one, concedo, I grant it; but in what does regard that possession, nego, I deny it. <span>The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm. Despite all this, disputations continued to be practised in certain university contexts for some time – indeed, they live on in the ceremonial character of PhD defences. The point, thou







Flashcard 1735819660556

Tags
#logic
Question
the introduction of [...] in Europe accelerated the downfall of the disputational culture.
Answer
new printing techniques

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
ich is thoroughly disputational, with Meditations on First Philosophy (1641) by Descartes, a book argued through long paragraphs driven by the first-person singular. The nature of intellectual enquiry shifted with the downfall of disputation. <span>It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440. Before that, books were a rare commodity, and education was conducted almost exclusively by means of oral contact between masters and pupils in the form of expository lectures in which







Flashcard 1735821233420

Tags
#history #logic
Question
Johannes Gutenberg introduced new printing techniques in Europe around [...].
Answer
1440

You can't terrorise Aristotle!

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
ich is thoroughly disputational, with Meditations on First Philosophy (1641) by Descartes, a book argued through long paragraphs driven by the first-person singular. The nature of intellectual enquiry shifted with the downfall of disputation. <span>It is also not happenstance that the downfall of the disputational culture roughly coincided with the introduction of new printing techniques in Europe by Johannes Gutenberg, around 1440. Before that, books were a rare commodity, and education was conducted almost exclusively by means of oral contact between masters and pupils in the form of expository lectures in which







#logic
Instead, early modern authors emphasise the role of novelty and individual discovery
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
tually unthinkable before the wide availability of printed books) was well-established. Moreover, as indicated by the passage from Descartes above, the very term ‘logic’ came to be used for something other than what the scholastics had meant. <span>Instead, early modern authors emphasise the role of novelty and individual discovery, as exemplified by the influential textbook Port-Royal Logic (1662), essentially, the logical version of Cartesianism, based on Descartes’s conception of mental operations and the prima




Flashcard 1735825165580

Tags
#logic
Question
Instead of justification of ideas, early modern authors emphasise the role of [...] and [...]
Answer
novelty and individual discovery

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Instead, early modern authors emphasise the role of novelty and individual discovery

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
tually unthinkable before the wide availability of printed books) was well-established. Moreover, as indicated by the passage from Descartes above, the very term ‘logic’ came to be used for something other than what the scholastics had meant. <span>Instead, early modern authors emphasise the role of novelty and individual discovery, as exemplified by the influential textbook Port-Royal Logic (1662), essentially, the logical version of Cartesianism, based on Descartes’s conception of mental operations and the prima







#logic
Descartes hits the nail on the head when he claims that the logic of the Schools (scholastic logic) is not really a logic of discovery. Its chief purpose is justification and exposition , which makes sense particularly against the background of dialectical practices, where interlocutors explain and debate what they themselves already know.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

The rise and fall and rise of logic | Aeon Essays
without judgment about things one does not know. Such logic corrupts good sense rather than increasing it. I mean instead the kind of logic which teaches us to direct our reason with a view to discovering the truths of which we are ignorant. <span>Descartes hits the nail on the head when he claims that the logic of the Schools (scholastic logic) is not really a logic of discovery. Its chief purpose is justification and exposition, which makes sense particularly against the background of dialectical practices, where interlocutors explain and debate what they themselves already know. Indeed, for much of the history of logic, both in ancient Greece and in the Latin medieval tradition, ‘dialectic’ and ‘logic’ were taken to be synonymous. Up to Descartes’s time, the ch




Flashcard 1735828311308

Tags
#logic
Question
Descartes claims the chief purpose of scholastic logic is [...]
Answer
justification and exposition

which makes sense particularly against the background of dialectical practices, where interlocutors explain and debate what they themselves already know.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Descartes hits the nail on the head when he claims that the logic of the Schools (scholastic logic) is not really a logic of discovery. Its chief purpose is justification and exposition , which makes sense particularly against the background of dialectical practices, where interlocutors explain and debate what they themselves already know.

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
without judgment about things one does not know. Such logic corrupts good sense rather than increasing it. I mean instead the kind of logic which teaches us to direct our reason with a view to discovering the truths of which we are ignorant. <span>Descartes hits the nail on the head when he claims that the logic of the Schools (scholastic logic) is not really a logic of discovery. Its chief purpose is justification and exposition, which makes sense particularly against the background of dialectical practices, where interlocutors explain and debate what they themselves already know. Indeed, for much of the history of logic, both in ancient Greece and in the Latin medieval tradition, ‘dialectic’ and ‘logic’ were taken to be synonymous. Up to Descartes’s time, the ch







Flashcard 1735995559180

Tags
#stochastics
Question
Random walks are usually defined as sums of [...] in Euclidean space
Answer
iid random variables or random vectors

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Random walks are stochastic processes that are usually defined as sums of iid random variables or random vectors in Euclidean space, so they are processes that change in discrete time.

Original toplevel document

Stochastic process - Wikipedia
one, while the value of a tail is zero. [61] In other words, a Bernoulli process is a sequence of iid Bernoulli random variables, [62] where each coin flip is a Bernoulli trial. [63] Random walk[edit source] Main article: Random walk <span>Random walks are stochastic processes that are usually defined as sums of iid random variables or random vectors in Euclidean space, so they are processes that change in discrete time. [64] [65] [66] [67] [68] But some also use the term to refer to processes that change in continuous time, [69] particularly the Wiener process used in finance, which has led to some c







#abstract-algebra
In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Algebraic structure - Wikipedia
Module-like[show] Module Group with operators Vector space Linear algebra Algebra-like[show] Algebra Associative Non-associative Composition algebra Lie algebra Graded Bialgebra v t e <span>In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets,




Flashcard 1736192953612

Tags
#abstract-algebra
Question
an [...] is a set with one or more operations defined on it that satisfies a list of axioms.
Answer
algebraic structure

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms.

Original toplevel document

Algebraic structure - Wikipedia
Module-like[show] Module Group with operators Vector space Linear algebra Algebra-like[show] Algebra Associative Non-associative Composition algebra Lie algebra Graded Bialgebra v t e <span>In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets,







Flashcard 1736195050764

Tags
#abstract-algebra
Question
The underlying set for an algebraic structure is called a [...]
Answer
carrier set

Think sigma-algebra and measurable space

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms.

Original toplevel document

Algebraic structure - Wikipedia
Module-like[show] Module Group with operators Vector space Linear algebra Algebra-like[show] Algebra Associative Non-associative Composition algebra Lie algebra Graded Bialgebra v t e <span>In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets,







Flashcard 1736196885772

Tags
#abstract-algebra
Question
an algebraic structure is a set with one or more [...] defined on it that satisfies a list of axioms.
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms.

Original toplevel document

Algebraic structure - Wikipedia
Module-like[show] Module Group with operators Vector space Linear algebra Algebra-like[show] Algebra Associative Non-associative Composition algebra Lie algebra Graded Bialgebra v t e <span>In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets,







Flashcard 1736198458636

Tags
#abstract-algebra
Question
an algebraic structure is a set with one or more operations defined on it that satisfies [...].
Answer
a list of axioms

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
tml> In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. <html>

Original toplevel document

Algebraic structure - Wikipedia
Module-like[show] Module Group with operators Vector space Linear algebra Algebra-like[show] Algebra Associative Non-associative Composition algebra Lie algebra Graded Bialgebra v t e <span>In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets,







#abstract-algebra

Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets, or by altering the defining axioms. Examples of more complex algebraic structures include vector spaces, modules, and algebras.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Algebraic structure - Wikipedia
algebra v t e In mathematics, and more specifically in abstract algebra, an algebraic structure is a set (called carrier set or underlying set) with one or more operations defined on it that satisfies a list of axioms. [1] <span>Examples of algebraic structures include groups, rings, fields, and lattices. More complex structures can be defined by introducing multiple operations, different underlying sets, or by altering the defining axioms. Examples of more complex algebraic structures include vector spaces, modules, and algebras. The properties of specific algebraic structures are studied in abstract algebra. The general theory of algebraic structures has been formalized in universal algebra. The language of c




#poisson-process #stochastics
Its name (Poisson Process) derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Poisson point process - Wikipedia
oint processes, some of which are constructed with the Poisson point process, that seek to capture such interaction. [22] The process is named after French mathematician Siméon Denis Poisson despite Poisson never having studied the process. <span>Its name derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution. The process was discovered independently and repeatedly in several settings, including experiments on radioactive decay, telephone call arrivals and insurance mathematics. [23] [24] T




#poisson-process #stochastics
The Poisson point process is often defined on the real line, where it can be considered as a stochastic process.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Poisson point process - Wikipedia
and used as a mathematical model for seemingly random processes in numerous disciplines such as astronomy, [3] biology, [4] ecology, [5] geology, [6] physics, [7] economics, [8] image processing, [9] and telecommunications. [10] [11] <span>The Poisson point process is often defined on the real line, where it can be considered as a stochastic process. In this setting, it is used, for example, in queueing theory [12] to model random events, such as the arrival of customers at a store or phone calls at an exchange, distributed in tim




#poisson-process #stochastics
In the plane, the point process, also known as a spatial Poisson process,[13] can represent the locations of scattered objects such as transmitters in a wireless network,[10][14][15][16] particles colliding into a detector, or trees in a forest.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Poisson point process - Wikipedia
e, where it can be considered as a stochastic process. In this setting, it is used, for example, in queueing theory [12] to model random events, such as the arrival of customers at a store or phone calls at an exchange, distributed in time. <span>In the plane, the point process, also known as a spatial Poisson process, [13] can represent the locations of scattered objects such as transmitters in a wireless network, [10] [14] [15] [16] particles colliding into a detector, or trees in a forest. [17] In this setting, the process is often used in mathematical models and in the related fields of spatial point processes, [18] stochastic geometry, [1] spatial statistics [18] [1




Flashcard 1736250101004

Tags
#poisson-process #stochastics
Question
if [...] in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution.
Answer
a collection of random points

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Its name (Poisson Process) derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution.

Original toplevel document

Poisson point process - Wikipedia
oint processes, some of which are constructed with the Poisson point process, that seek to capture such interaction. [22] The process is named after French mathematician Siméon Denis Poisson despite Poisson never having studied the process. <span>Its name derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution. The process was discovered independently and repeatedly in several settings, including experiments on radioactive decay, telephone call arrivals and insurance mathematics. [23] [24] T







Flashcard 1736251673868

Tags
#poisson-process #stochastics
Question
In a Poisson process, [...] is a random variable with a Poisson distribution.
Answer
the number of points in a region of finite size

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Its name (Poisson Process) derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution.

Original toplevel document

Poisson point process - Wikipedia
oint processes, some of which are constructed with the Poisson point process, that seek to capture such interaction. [22] The process is named after French mathematician Siméon Denis Poisson despite Poisson never having studied the process. <span>Its name derives from the fact that if a collection of random points in some space forms a Poisson process, then the number of points in a region of finite size is a random variable with a Poisson distribution. The process was discovered independently and repeatedly in several settings, including experiments on radioactive decay, telephone call arrivals and insurance mathematics. [23] [24] T







Flashcard 1736253246732

Tags
#poisson-process #stochastics
Question
The Poisson point process is often defined on [...], where it can be considered as a stochastic process.
Answer
the real line

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Poisson point process is often defined on the real line, where it can be considered as a stochastic process.

Original toplevel document

Poisson point process - Wikipedia
and used as a mathematical model for seemingly random processes in numerous disciplines such as astronomy, [3] biology, [4] ecology, [5] geology, [6] physics, [7] economics, [8] image processing, [9] and telecommunications. [10] [11] <span>The Poisson point process is often defined on the real line, where it can be considered as a stochastic process. In this setting, it is used, for example, in queueing theory [12] to model random events, such as the arrival of customers at a store or phone calls at an exchange, distributed in tim







Flashcard 1736254819596

Tags
#poisson-process #stochastics
Question
In the plane, the point process, also known as a spatial Poisson process,[13] can represent [...] such as transmitters in a wireless network,[10][14][15][16] particles colliding into a detector, or trees in a forest.
Answer
the locations of scattered objects

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In the plane, the point process, also known as a spatial Poisson process, [13] can represent the locations of scattered objects such as transmitters in a wireless network, [10] [14] [15] [16] particles colliding into a detector, or trees in a forest.

Original toplevel document

Poisson point process - Wikipedia
e, where it can be considered as a stochastic process. In this setting, it is used, for example, in queueing theory [12] to model random events, such as the arrival of customers at a store or phone calls at an exchange, distributed in time. <span>In the plane, the point process, also known as a spatial Poisson process, [13] can represent the locations of scattered objects such as transmitters in a wireless network, [10] [14] [15] [16] particles colliding into a detector, or trees in a forest. [17] In this setting, the process is often used in mathematical models and in the related fields of spatial point processes, [18] stochastic geometry, [1] spatial statistics [18] [1







#lists #python
An operation like this that combines a sequence of elements into a single value is sometimes called reduce .
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Lists
of the elements; a variable used this way is sometimes called an accumulator. Adding up the elements of a list is such a common operation that Python provides it as a built-in function, sum: >>> t = [1, 2, 3] >>> sum(t) 6 <span>An operation like this that combines a sequence of elements into a single value is sometimes called reduce. Sometimes you want to traverse one list while building another. For example, the following function takes a list of strings and returns a new list that contains capitalized strings:




#lists #python
An operation like capitalize_all is sometimes called a map because it “maps” a function (in this case the method capitalize ) onto each of the elements in a sequence.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Lists
ngs: def capitalize_all(t): res = [] for s in t: res.append(s.capitalize()) return res res is initialized with an empty list; each time through the loop, we append the next element. So res is another kind of accumulator. <span>An operation like capitalize_all is sometimes called a map because it “maps” a function (in this case the method capitalize) onto each of the elements in a sequence. Another common operation is to select some of the elements from a list and return a sublist. For example, the following function takes a list of strings and returns a list that cont




#lists #python
An operation like only_upper is called a filter because it selects some of the elements and filters out the others.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Lists
ontains only the uppercase strings: def only_upper(t): res = [] for s in t: if s.isupper(): res.append(s) return res isupper is a string method that returns True if the string contains only upper case letters. <span>An operation like only_upper is called a filter because it selects some of the elements and filters out the others. Most common list operations can be expressed as a combination of map, filter and reduce. 10.8 Deleting elements There are several ways to delete elements from a list. If you kno




#lists #python

Most common list operations can be expressed as a combination of map, filter and reduce.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Lists
res.append(s) return res isupper is a string method that returns True if the string contains only upper case letters. An operation like only_upper is called a filter because it selects some of the elements and filters out the others. <span>Most common list operations can be expressed as a combination of map, filter and reduce. 10.8 Deleting elements There are several ways to delete elements from a list. If you know the index of the element you want, you can use pop: >>> t = ['a', 'b', 'c']




Flashcard 1737334590732

Tags
#lists #python
Question
An operation like this that combines a sequence of elements into a single value is sometimes called [...] .
Answer
reduce

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An operation like this that combines a sequence of elements into a single value is sometimes called reduce .

Original toplevel document

Lists
of the elements; a variable used this way is sometimes called an accumulator. Adding up the elements of a list is such a common operation that Python provides it as a built-in function, sum: >>> t = [1, 2, 3] >>> sum(t) 6 <span>An operation like this that combines a sequence of elements into a single value is sometimes called reduce. Sometimes you want to traverse one list while building another. For example, the following function takes a list of strings and returns a new list that contains capitalized strings:







Flashcard 1737336950028

Tags
#lists #python
Question
An map operation “maps” a function onto [...] in a sequence.
Answer
each of the elements

like capitalize_all

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An operation like capitalize_all is sometimes called a map because it “maps” a function (in this case the method capitalize ) onto each of the elements in a sequence.

Original toplevel document

Lists
ngs: def capitalize_all(t): res = [] for s in t: res.append(s.capitalize()) return res res is initialized with an empty list; each time through the loop, we append the next element. So res is another kind of accumulator. <span>An operation like capitalize_all is sometimes called a map because it “maps” a function (in this case the method capitalize) onto each of the elements in a sequence. Another common operation is to select some of the elements from a list and return a sublist. For example, the following function takes a list of strings and returns a list that cont







Flashcard 1737338522892

Tags
#lists #python
Question
An operation like only_upper is called a [...] because it selects some of the elements and filters out the others.
Answer
filter

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
An operation like only_upper is called a filter because it selects some of the elements and filters out the others.

Original toplevel document

Lists
ontains only the uppercase strings: def only_upper(t): res = [] for s in t: if s.isupper(): res.append(s) return res isupper is a string method that returns True if the string contains only upper case letters. <span>An operation like only_upper is called a filter because it selects some of the elements and filters out the others. Most common list operations can be expressed as a combination of map, filter and reduce. 10.8 Deleting elements There are several ways to delete elements from a list. If you kno







Flashcard 1737340095756

Tags
#lists #python
Question

Most common list operations can be expressed as a combination of [...].

Answer
map, filter and reduce

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Most common list operations can be expressed as a combination of map, filter and reduce.

Original toplevel document

Lists
res.append(s) return res isupper is a string method that returns True if the string contains only upper case letters. An operation like only_upper is called a filter because it selects some of the elements and filters out the others. <span>Most common list operations can be expressed as a combination of map, filter and reduce. 10.8 Deleting elements There are several ways to delete elements from a list. If you know the index of the element you want, you can use pop: >>> t = ['a', 'b', 'c']







Flashcard 1737953512716

Tags
#state-space-models
Question
inference for [...] can be performed efficiently using the sum-product algorithm
Answer
directed graphs having a tree structure (no loops)

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
focus on the two most important examples of state space models, namely the hidden Markov model, in which the latent variables are discrete, and linear dynamical systems, in which the latent variables are Gaussian. Both models are described by <span>directed graphs having a tree structure (no loops) for which inference can be performed efficiently using the sum-product algorithm <span><body><html>

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 1737961377036

Tags
#state-space-models
Question
[...] assumes equal conditional distributions \(p(x_n |x_{n−1} ) \) for all n
Answer
stationary time series

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In most applications of such (Markovian) models, the conditional distributions p(xn|xn−1) that define the model will be constrained to be equal, corresponding to the assumption of a stationary time series. The model is then known as a homogeneous Markov chain.

Original toplevel document (pdf)

cannot see any pdfs







#measure-theory
The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, goes from 0 to 1 as goes from 0 to 1, and takes on every value in between.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Cantor function - Wikipedia
= 1/4. 200/243 becomes 0.21102 (or 0.211012222...) in base 3. The digits after the first 1 are replaced by 0s to produce 0.21. This is rewritten as 0.11. When read in base 2, this corresponds to 3/4, so c(200/243) = 3/4. Properties[edit] <span>The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, c ( x ) {\textstyle c(x)} goes from 0 to 1 as x {\textstyle x} goes from 0 to 1, and takes on every value in between. The Cantor function is the most frequently cited example of a real function that is uniformly continuous (precisely, it is Hölder continuous of exponent α = log 2/log 3) but not absolut




Flashcard 1738558541068

Tags
#measure-theory
Question
The [...] function challenges naive intuitions about continuity and measure
Answer
Cantor

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, goes from 0 to 1 as goes from 0 to 1, and take

Original toplevel document

Cantor function - Wikipedia
= 1/4. 200/243 becomes 0.21102 (or 0.211012222...) in base 3. The digits after the first 1 are replaced by 0s to produce 0.21. This is rewritten as 0.11. When read in base 2, this corresponds to 3/4, so c(200/243) = 3/4. Properties[edit] <span>The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, c ( x ) {\textstyle c(x)} goes from 0 to 1 as x {\textstyle x} goes from 0 to 1, and takes on every value in between. The Cantor function is the most frequently cited example of a real function that is uniformly continuous (precisely, it is Hölder continuous of exponent α = log 2/log 3) but not absolut







Flashcard 1738560113932

Tags
#measure-theory
Question
although the Cantor function is [...], goes from 0 to 1 as goes from 0 to 1, and takes on every value in between.
Answer
continuous everywhere and has zero derivative almost everywhere

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, goes from 0 to 1 as goes from 0 to 1, and takes on every value in between.

Original toplevel document

Cantor function - Wikipedia
= 1/4. 200/243 becomes 0.21102 (or 0.211012222...) in base 3. The digits after the first 1 are replaced by 0s to produce 0.21. This is rewritten as 0.11. When read in base 2, this corresponds to 3/4, so c(200/243) = 3/4. Properties[edit] <span>The Cantor function challenges naive intuitions about continuity and measure; though it is continuous everywhere and has zero derivative almost everywhere, c ( x ) {\textstyle c(x)} goes from 0 to 1 as x {\textstyle x} goes from 0 to 1, and takes on every value in between. The Cantor function is the most frequently cited example of a real function that is uniformly continuous (precisely, it is Hölder continuous of exponent α = log 2/log 3) but not absolut







Flashcard 1738581871884

Tags
#measure-theory #probability-theory
Question
Historically, the dividing line for probability is [...] when Andrey Kolmogorov published Foundations of the Theory of Probability
Answer
1933

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Historically, the dividing line is 1933 when Grundbegriffe der Wahrscheinlichkeitsrech- nung (Foundations of the Theory of Probability) by Andrey Kolmogorov was published

Original toplevel document (pdf)

cannot see any pdfs







#graphical-models
A graphical model or probabilistic graphical model (PGM) or structured probabilistic model is a probabilistic model for which a graph expresses the conditional dependence structure between random variables.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Graphical model - Wikipedia
list of references, but its sources remain unclear because it has insufficient inline citations. Please help to improve this article by introducing more precise citations. (May 2017) (Learn how and when to remove this template message) <span>A graphical model or probabilistic graphical model (PGM) or structured probabilistic model is a probabilistic model for which a graph expresses the conditional dependence structure between random variables. They are commonly used in probability theory, statistics—particularly Bayesian statistics—and machine learning. [imagelink] An example of a graphical model. Each arrow indicates




#graphical-models

In a Bayesian network, the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are then the joint probability satisfies

where is the set of parents of node . In other words, the joint distribution factors into a product of conditional distributions.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Graphical model - Wikipedia
the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce. [1] Bayesian network[edit source] Main article: Bayesian network <span>If the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are X 1 , … , X n {\displaystyle X_{1},\ldots ,X_{n}} then the joint probability satisfies P [ X 1 , … , X n ] = ∏ i = 1 n P [ X i | p a i ] {\displaystyle P[X_{1},\ldots ,X_{n}]=\prod _{i=1}^{n}P[X_{i}|pa_{i}]} where p a i {\displaystyle pa_{i}} is the set of parents of node X i {\displaystyle X_{i}} . In other words, the joint distribution factors into a product of conditional distributions. For example, the graphical model in the Figure shown above (which is actually not a directed acyclic graph, but an ancestral graph) consists of the random variables




#bayesian-network
Formally, Bayesian networks are DAGs whose:
  1. nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses.
  2. Edges represent conditional dependencies;
  3. nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other.
  4. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables




#graphical-models
A Markov random field, also known as a Markov network, is a model over an undirected graph.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Graphical model - Wikipedia
ne learning models like hidden Markov models, neural networks and newer models such as variable-order Markov models can be considered special cases of Bayesian networks. Markov random field[edit source] Main article: Markov random field <span>A Markov random field, also known as a Markov network, is a model over an undirected graph. A graphical model with many repeated subunits can be represented with plate notation. Other types[edit source] A factor graph is an undirected bipartite graph connecting variables a




Flashcard 1739036953868

Tags
#graphical-models
Question
In a graphical model, a graph expresses [...] between random variables.
Answer
the conditional dependence structure

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A graphical model or probabilistic graphical model (PGM) or structured probabilistic model is a probabilistic model for which a graph expresses the conditional dependence structure between random variables.

Original toplevel document

Graphical model - Wikipedia
list of references, but its sources remain unclear because it has insufficient inline citations. Please help to improve this article by introducing more precise citations. (May 2017) (Learn how and when to remove this template message) <span>A graphical model or probabilistic graphical model (PGM) or structured probabilistic model is a probabilistic model for which a graph expresses the conditional dependence structure between random variables. They are commonly used in probability theory, statistics—particularly Bayesian statistics—and machine learning. [imagelink] An example of a graphical model. Each arrow indicates







Flashcard 1739039313164

Tags
#graphical-models
Question

In a Bayesian network, the network structure of the model is a [...]


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In a Bayesian network, the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are then the joint probability satisfies where is

Original toplevel document

Graphical model - Wikipedia
the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce. [1] Bayesian network[edit source] Main article: Bayesian network <span>If the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are X 1 , … , X n {\displaystyle X_{1},\ldots ,X_{n}} then the joint probability satisfies P [ X 1 , … , X n ] = ∏ i = 1 n P [ X i | p a i ] {\displaystyle P[X_{1},\ldots ,X_{n}]=\prod _{i=1}^{n}P[X_{i}|pa_{i}]} where p a i {\displaystyle pa_{i}} is the set of parents of node X i {\displaystyle X_{i}} . In other words, the joint distribution factors into a product of conditional distributions. For example, the graphical model in the Figure shown above (which is actually not a directed acyclic graph, but an ancestral graph) consists of the random variables







Flashcard 1739041934604

Tags
#graphical-models
Question

if the events are the joint probability of a Bayesian network satisfies[...]

Answer

is the set of parents of node .


statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
In a Bayesian network, the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are then the joint probability <span>satisfies where is the set of parents of node . In other words, the joint distribution factors into a product of conditional distributions. <span><body><html>

Original toplevel document

Graphical model - Wikipedia
the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce. [1] Bayesian network[edit source] Main article: Bayesian network <span>If the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are X 1 , … , X n {\displaystyle X_{1},\ldots ,X_{n}} then the joint probability satisfies P [ X 1 , … , X n ] = ∏ i = 1 n P [ X i | p a i ] {\displaystyle P[X_{1},\ldots ,X_{n}]=\prod _{i=1}^{n}P[X_{i}|pa_{i}]} where p a i {\displaystyle pa_{i}} is the set of parents of node X i {\displaystyle X_{i}} . In other words, the joint distribution factors into a product of conditional distributions. For example, the graphical model in the Figure shown above (which is actually not a directed acyclic graph, but an ancestral graph) consists of the random variables







Flashcard 1739044556044

Tags
#graphical-models
Question

In a Bayesian network, the joint distribution factors into [...].

Answer
a product of conditional distributions

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
factorization of the joint probability of all random variables. More precisely, if the events are then the joint probability satisfies where is the set of parents of node . In other words, the joint distribution factors into <span>a product of conditional distributions. <span><body><html>

Original toplevel document

Graphical model - Wikipedia
the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce. [1] Bayesian network[edit source] Main article: Bayesian network <span>If the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are X 1 , … , X n {\displaystyle X_{1},\ldots ,X_{n}} then the joint probability satisfies P [ X 1 , … , X n ] = ∏ i = 1 n P [ X i | p a i ] {\displaystyle P[X_{1},\ldots ,X_{n}]=\prod _{i=1}^{n}P[X_{i}|pa_{i}]} where p a i {\displaystyle pa_{i}} is the set of parents of node X i {\displaystyle X_{i}} . In other words, the joint distribution factors into a product of conditional distributions. For example, the graphical model in the Figure shown above (which is actually not a directed acyclic graph, but an ancestral graph) consists of the random variables







Flashcard 1739046391052

Tags
#graphical-models
Question
A Markov random field is a model over [...].
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A Markov random field, also known as a Markov network, is a model over an undirected graph.

Original toplevel document

Graphical model - Wikipedia
ne learning models like hidden Markov models, neural networks and newer models such as variable-order Markov models can be considered special cases of Bayesian networks. Markov random field[edit source] Main article: Markov random field <span>A Markov random field, also known as a Markov network, is a model over an undirected graph. A graphical model with many repeated subunits can be represented with plate notation. Other types[edit source] A factor graph is an undirected bipartite graph connecting variables a







Flashcard 1739048488204

Tags
#graphical-models
Question
A [...] is a model over an undirected graph.
Answer
Markov random field

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
A Markov random field, also known as a Markov network, is a model over an undirected graph.

Original toplevel document

Graphical model - Wikipedia
ne learning models like hidden Markov models, neural networks and newer models such as variable-order Markov models can be considered special cases of Bayesian networks. Markov random field[edit source] Main article: Markov random field <span>A Markov random field, also known as a Markov network, is a model over an undirected graph. A graphical model with many repeated subunits can be represented with plate notation. Other types[edit source] A factor graph is an undirected bipartite graph connecting variables a







Flashcard 1739051896076

Tags
#bayesian-network
Question
In Bayesian network graphs [...] represent variables
Answer
nodes

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
Formally, Bayesian networks are DAGs whose: nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that ar

Original toplevel document

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables







Flashcard 1739053468940

Tags
#bayesian-network
Question
Graphically, Bayesian networks use [...] to represent conditional dependencies
Answer
Edges

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
html> Formally, Bayesian networks are DAGs whose: nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are condition

Original toplevel document

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables







Flashcard 1739055041804

Tags
#bayesian-network
Question
nodes that are not connected represent variables that are [...].
Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ntities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are <span>conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probabil

Original toplevel document

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables







Flashcard 1739056614668

Tags
#bayesian-network
Question
In a Bayesian network each node takes [...] as input
Answer
a set of values from parent nodes

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
re not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, <span>a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. <span><body><html>

Original toplevel document

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables







Flashcard 1739058187532

Tags
#bayesian-network
Question
In Bayesian networks each node represents a variable with [...]
Answer
a probability distribution

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) <span>the probability (or probability distribution, if applicable) of the variable represented by the node. <span><body><html>

Original toplevel document

Bayesian network - Wikipedia
ed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. <span>Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies; nodes that are not connected (there is no path from one of the variables to the other in the Bayesian network) represent variables that are conditionally independent of each other. Each node is associated with a probability function that takes, as input, a particular set of values for the node's parent variables, and gives (as output) the probability (or probability distribution, if applicable) of the variable represented by the node. For example, if m {\displaystyle m} parent nodes represent m {\displaystyle m} Boolean variables







#d-separation
d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Unknown title
CAUSALITY - Discussion d-SEPARATION WITHOUT TEARS (At the request of many readers) Introduction d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z. The idea is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness"




#d-separation
The idea (of d-separation) is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation".
The only twist on this simple idea is to define what we mean by "connecting path", given that
  1. we are dealing with a system of directed arrows
  2. in the graph some vertices correspond to measured variables, whose values are known precisely.
To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional").
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Unknown title
d-SEPARATION WITHOUT TEARS (At the request of many readers) Introduction d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z. <span>The idea is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation". The only twist on this simple idea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to measured variables, whose values are known precisely. To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"). We start by considering separation between two singleton variables, x and y; the extension to sets of variables is straightforward (i.e., two sets are separated if and only if each el




Flashcard 1739068411148

Tags
#d-separation
Question
The idea of d-separation is complicated by defining exactly what is a [...], given that the edges are directed and some variables are already measured.
Answer
connecting path

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
nce" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation". The only twist on this simple idea is to define what we mean by <span>"connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to measured variables, whose values are known precisely. To account f

Original toplevel document

Unknown title
d-SEPARATION WITHOUT TEARS (At the request of many readers) Introduction d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z. <span>The idea is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation". The only twist on this simple idea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to measured variables, whose values are known precisely. To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"). We start by considering separation between two singleton variables, x and y; the extension to sets of variables is straightforward (i.e., two sets are separated if and only if each el







Flashcard 1739069984012

Tags
#d-separation
Question
The only twist on this simple idea of "connecting path" is that we are dealing with a system of directed arrows in which some vertices correspond to [...], whose values are known precisely.
Answer
measured variables

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
or "separation". The only twist on this simple idea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to <span>measured variables, whose values are known precisely. To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"

Original toplevel document

Unknown title
d-SEPARATION WITHOUT TEARS (At the request of many readers) Introduction d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z. <span>The idea is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation". The only twist on this simple idea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to measured variables, whose values are known precisely. To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"). We start by considering separation between two singleton variables, x and y; the extension to sets of variables is straightforward (i.e., two sets are separated if and only if each el







#forward-backward-algorithm #hmm
The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Forward–backward algorithm - Wikipedia
ackward algorithm - Wikipedia Forward–backward algorithm From Wikipedia, the free encyclopedia (Redirected from Forward-backward algorithm) Jump to: navigation, search <span>The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations/emissions o 1 : t := o 1




#forward-backward-algorithm #hmm
The foreward-backward algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in time; hence the name forward–backward algorithm.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Forward–backward algorithm - Wikipedia
| o 1 : t ) {\displaystyle P(X_{k}\ |\ o_{1:t})} . This inference task is usually called smoothing. <span>The algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in time; hence the name forward–backward algorithm. The term forward–backward algorithm is also used to refer to any algorithm belonging to the general class of algorithms that operate on sequence models in a forward–backward manner. I




Flashcard 1739077061900

Tags
#forward-backward-algorithm #hmm
Question
The forward–backward algorithm is an inference algorithm for hidden Markov models which computes [...] given a sequence of observations
Answer
the posterior marginals of all hidden state variables

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations

Original toplevel document

Forward–backward algorithm - Wikipedia
ackward algorithm - Wikipedia Forward–backward algorithm From Wikipedia, the free encyclopedia (Redirected from Forward-backward algorithm) Jump to: navigation, search <span>The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations/emissions o 1 : t := o 1







Flashcard 1739078634764

Tags
#forward-backward-algorithm #hmm
Question
The forward–backward algorithm computes the posterior marginals of all hidden state variables given [...]
Answer
a sequence of observations

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations

Original toplevel document

Forward–backward algorithm - Wikipedia
ackward algorithm - Wikipedia Forward–backward algorithm From Wikipedia, the free encyclopedia (Redirected from Forward-backward algorithm) Jump to: navigation, search <span>The forward–backward algorithm is an inference algorithm for hidden Markov models which computes the posterior marginals of all hidden state variables given a sequence of observations/emissions o 1 : t := o 1







Flashcard 1739080994060

Tags
#forward-backward-algorithm #hmm
Question
The foreward-backward algorithm makes use of the principle of [...] in its two passes.
Answer
dynamic programming

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The foreward-backward algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in t

Original toplevel document

Forward–backward algorithm - Wikipedia
| o 1 : t ) {\displaystyle P(X_{k}\ |\ o_{1:t})} . This inference task is usually called smoothing. <span>The algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in time; hence the name forward–backward algorithm. The term forward–backward algorithm is also used to refer to any algorithm belonging to the general class of algorithms that operate on sequence models in a forward–backward manner. I







Flashcard 1739082566924

Tags
#forward-backward-algorithm #hmm
Question
The foreward-backward algorithm obtain [...] in two passes.
Answer
the posterior marginal distributions

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The foreward-backward algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in time; hence the name forward–backward algorithm.

Original toplevel document

Forward–backward algorithm - Wikipedia
| o 1 : t ) {\displaystyle P(X_{k}\ |\ o_{1:t})} . This inference task is usually called smoothing. <span>The algorithm makes use of the principle of dynamic programming to compute efficiently the values that are required to obtain the posterior marginal distributions in two passes. The first pass goes forward in time while the second goes backward in time; hence the name forward–backward algorithm. The term forward–backward algorithm is also used to refer to any algorithm belonging to the general class of algorithms that operate on sequence models in a forward–backward manner. I







Flashcard 1739212852492

Tags
#d-separation
Question
the "d" in terms "d-separated" and "d-connected" accounts for [...]
Answer
the orientations of the arrows

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
dea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in the graph some vertices correspond to measured variables, whose values are known precisely. To account for <span>the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"). <span><body><html>

Original toplevel document

Unknown title
d-SEPARATION WITHOUT TEARS (At the request of many readers) Introduction d-separation is a criterion for deciding, from a given a causal graph, whether a set X of variables is independent of another set Y, given a third set Z. <span>The idea is to associate "dependence" with "connectedness" (i.e., the existence of a connecting path) and "independence" with "unconnected-ness" or "separation". The only twist on this simple idea is to define what we mean by "connecting path", given that we are dealing with a system of directed arrows in which some vertices (those residing in Z) correspond to measured variables, whose values are known precisely. To account for the orientations of the arrows we use the terms "d-separated" and "d-connected" (d connotes "directional"). We start by considering separation between two singleton variables, x and y; the extension to sets of variables is straightforward (i.e., two sets are separated if and only if each el







#forward-backward-algorithm #hmm

The forward-backward algorithm

  1. In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all , the probability of ending up in any particular state given the first observations in the sequence, i.e. .
  2. In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point , i.e. .
  3. These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence:

The last step follows from an application of the Bayes' rule and the conditional independence of and given .

It remains to be seen, of course, how the forward and backward passes are actually calculated.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Forward–backward algorithm - Wikipedia
cific instance of this class. Contents [hide] 1 Overview 2 Forward probabilities 3 Backward probabilities 4 Example 5 Performance 6 Pseudocode 7 Python example 8 See also 9 References 10 External links Overview[edit source] <span>In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all k ∈ { 1 , … , t } {\displaystyle k\in \{1,\dots ,t\}} , the probability of ending up in any particular state given the first k {\displaystyle k} observations in the sequence, i.e. P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:k})} . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point k {\displaystyle k} , i.e. P ( o k + 1 : t | X k ) {\displaystyle P(o_{k+1:t}\ |\ X_{k})} . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence: P ( X k | o 1 : t ) = P ( X k | o 1 : k , o k + 1 : t ) ∝ P ( o k + 1 : t | X k ) P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:t})=P(X_{k}\ |\ o_{1:k},o_{k+1:t})\propto P(o_{k+1:t}\ |\ X_{k})P(X_{k}|o_{1:k})} The last step follows from an application of the Bayes' rule and the conditional independence of o k + 1 : t {\displaystyle o_{k+1:t}} and o 1 : k {\displaystyle o_{1:k}} given X k {\displaystyle X_{k}} . As outlined above, the algorithm involves three steps: computing forward probabilities computing backward probabilities computing smoothed values. The forward and backward steps m




Flashcard 1739928505612

Tags
#forward-backward-algorithm #hmm
Question

In the first pass, the forward–backward algorithm computes [...]

Answer
.

the distribution over hidden states given the observations up to the point.

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The forward-backward algorithm In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all , the probability of ending up in any particular state given the first observations in the sequence, i.e. . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point , i.e. . Thes

Original toplevel document

Forward–backward algorithm - Wikipedia
cific instance of this class. Contents [hide] 1 Overview 2 Forward probabilities 3 Backward probabilities 4 Example 5 Performance 6 Pseudocode 7 Python example 8 See also 9 References 10 External links Overview[edit source] <span>In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all k ∈ { 1 , … , t } {\displaystyle k\in \{1,\dots ,t\}} , the probability of ending up in any particular state given the first k {\displaystyle k} observations in the sequence, i.e. P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:k})} . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point k {\displaystyle k} , i.e. P ( o k + 1 : t | X k ) {\displaystyle P(o_{k+1:t}\ |\ X_{k})} . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence: P ( X k | o 1 : t ) = P ( X k | o 1 : k , o k + 1 : t ) ∝ P ( o k + 1 : t | X k ) P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:t})=P(X_{k}\ |\ o_{1:k},o_{k+1:t})\propto P(o_{k+1:t}\ |\ X_{k})P(X_{k}|o_{1:k})} The last step follows from an application of the Bayes' rule and the conditional independence of o k + 1 : t {\displaystyle o_{k+1:t}} and o 1 : k {\displaystyle o_{1:k}} given X k {\displaystyle X_{k}} . As outlined above, the algorithm involves three steps: computing forward probabilities computing backward probabilities computing smoothed values. The forward and backward steps m







Flashcard 1739930864908

Tags
#forward-backward-algorithm #hmm
Question

In the second pass, the forward-backward algorithm computes [...]

Answer
.

the probability of observing the remaining observations given any starting point

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
rward probabilities which provide, for all , the probability of ending up in any particular state given the first observations in the sequence, i.e. . In the second pass, the algorithm computes a set of backward probabilities which provide <span>the probability of observing the remaining observations given any starting point , i.e. . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence:

Original toplevel document

Forward–backward algorithm - Wikipedia
cific instance of this class. Contents [hide] 1 Overview 2 Forward probabilities 3 Backward probabilities 4 Example 5 Performance 6 Pseudocode 7 Python example 8 See also 9 References 10 External links Overview[edit source] <span>In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all k ∈ { 1 , … , t } {\displaystyle k\in \{1,\dots ,t\}} , the probability of ending up in any particular state given the first k {\displaystyle k} observations in the sequence, i.e. P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:k})} . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point k {\displaystyle k} , i.e. P ( o k + 1 : t | X k ) {\displaystyle P(o_{k+1:t}\ |\ X_{k})} . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence: P ( X k | o 1 : t ) = P ( X k | o 1 : k , o k + 1 : t ) ∝ P ( o k + 1 : t | X k ) P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:t})=P(X_{k}\ |\ o_{1:k},o_{k+1:t})\propto P(o_{k+1:t}\ |\ X_{k})P(X_{k}|o_{1:k})} The last step follows from an application of the Bayes' rule and the conditional independence of o k + 1 : t {\displaystyle o_{k+1:t}} and o 1 : k {\displaystyle o_{1:k}} given X k {\displaystyle X_{k}} . As outlined above, the algorithm involves three steps: computing forward probabilities computing backward probabilities computing smoothed values. The forward and backward steps m







Flashcard 1739933224204

Tags
#forward-backward-algorithm #hmm
Question

In the forward-backward algorithm, the forward and backward probability distributions are combined to obtain [...]

Answer
the distribution over states at any specific point in time given the entire observation sequence

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point , i.e. . These two sets of probability distributions can then be combined to obtain <span>the distribution over states at any specific point in time given the entire observation sequence: The last step follows from an application of the Bayes' rule and the conditional independence of and given . It remains to be seen, of course, how the forwa

Original toplevel document

Forward–backward algorithm - Wikipedia
cific instance of this class. Contents [hide] 1 Overview 2 Forward probabilities 3 Backward probabilities 4 Example 5 Performance 6 Pseudocode 7 Python example 8 See also 9 References 10 External links Overview[edit source] <span>In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all k ∈ { 1 , … , t } {\displaystyle k\in \{1,\dots ,t\}} , the probability of ending up in any particular state given the first k {\displaystyle k} observations in the sequence, i.e. P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:k})} . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point k {\displaystyle k} , i.e. P ( o k + 1 : t | X k ) {\displaystyle P(o_{k+1:t}\ |\ X_{k})} . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence: P ( X k | o 1 : t ) = P ( X k | o 1 : k , o k + 1 : t ) ∝ P ( o k + 1 : t | X k ) P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:t})=P(X_{k}\ |\ o_{1:k},o_{k+1:t})\propto P(o_{k+1:t}\ |\ X_{k})P(X_{k}|o_{1:k})} The last step follows from an application of the Bayes' rule and the conditional independence of o k + 1 : t {\displaystyle o_{k+1:t}} and o 1 : k {\displaystyle o_{1:k}} given X k {\displaystyle X_{k}} . As outlined above, the algorithm involves three steps: computing forward probabilities computing backward probabilities computing smoothed values. The forward and backward steps m







Flashcard 1739934797068

Tags
#forward-backward-algorithm #hmm
Question

In the forward-backward algorithm, the formula for posterior marginals is [...]

Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
bserving the remaining observations given any starting point , i.e. . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence<span>: The last step follows from an application of the Bayes' rule and the conditional independence of and given . It remains to be seen, of course, how the forwar

Original toplevel document

Forward–backward algorithm - Wikipedia
cific instance of this class. Contents [hide] 1 Overview 2 Forward probabilities 3 Backward probabilities 4 Example 5 Performance 6 Pseudocode 7 Python example 8 See also 9 References 10 External links Overview[edit source] <span>In the first pass, the forward–backward algorithm computes a set of forward probabilities which provide, for all k ∈ { 1 , … , t } {\displaystyle k\in \{1,\dots ,t\}} , the probability of ending up in any particular state given the first k {\displaystyle k} observations in the sequence, i.e. P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:k})} . In the second pass, the algorithm computes a set of backward probabilities which provide the probability of observing the remaining observations given any starting point k {\displaystyle k} , i.e. P ( o k + 1 : t | X k ) {\displaystyle P(o_{k+1:t}\ |\ X_{k})} . These two sets of probability distributions can then be combined to obtain the distribution over states at any specific point in time given the entire observation sequence: P ( X k | o 1 : t ) = P ( X k | o 1 : k , o k + 1 : t ) ∝ P ( o k + 1 : t | X k ) P ( X k | o 1 : k ) {\displaystyle P(X_{k}\ |\ o_{1:t})=P(X_{k}\ |\ o_{1:k},o_{k+1:t})\propto P(o_{k+1:t}\ |\ X_{k})P(X_{k}|o_{1:k})} The last step follows from an application of the Bayes' rule and the conditional independence of o k + 1 : t {\displaystyle o_{k+1:t}} and o 1 : k {\displaystyle o_{1:k}} given X k {\displaystyle X_{k}} . As outlined above, the algorithm involves three steps: computing forward probabilities computing backward probabilities computing smoothed values. The forward and backward steps m







#poisson-process #stochastics
For a collection of disjoint and bounded subregions of the underlying space, the number of points of a Poisson point process in each bounded subregion will be completely independent of all the others.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Poisson point process - Wikipedia
sed to define the Poisson distribution. If a Poisson point process is defined on some underlying space, then the number of points in a bounded region of this space will be a Poisson random variable. [45] Complete independence[edit source] <span>For a collection of disjoint and bounded subregions of the underlying space, the number of points of a Poisson point process in each bounded subregion will be completely independent of all the others. This property is known under several names such as complete randomness, complete independence, [21] or independent scattering [46] [47] and is common to all Poisson point processes.




Flashcard 1739942661388

Tags
#poisson-process #stochastics
Question
For a collection of [...] subregions of the underlying space, the number of points of a Poisson point process in each bounded subregion will be completely independent of all the others.
Answer
disjoint and bounded

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
For a collection of disjoint and bounded subregions of the underlying space, the number of points of a Poisson point process in each bounded subregion will be completely independent of all the others. </h

Original toplevel document

Poisson point process - Wikipedia
sed to define the Poisson distribution. If a Poisson point process is defined on some underlying space, then the number of points in a bounded region of this space will be a Poisson random variable. [45] Complete independence[edit source] <span>For a collection of disjoint and bounded subregions of the underlying space, the number of points of a Poisson point process in each bounded subregion will be completely independent of all the others. This property is known under several names such as complete randomness, complete independence, [21] or independent scattering [46] [47] and is common to all Poisson point processes.







#poisson-process #stochastics
If a Poisson point process has a parameter of the form , where is Lebegues measure, and is a constant, then the point process is called a homogeneous or stationary Poisson point process.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Poisson point process - Wikipedia
edit source] For all the different settings of the Poisson point process, the two key properties [b] of the Poisson distribution and complete independence play an important role. [25] [45] Homogeneous Poisson point process[edit source] <span>If a Poisson point process has a parameter of the form Λ = ν λ {\displaystyle \textstyle \Lambda =\nu \lambda } , where ν {\displaystyle \textstyle \nu } is Lebegues measure, which assigns length, area, or volume to sets, and λ {\displaystyle \textstyle \lambda } is a constant, then the point process is called a homogeneous or stationary Poisson point process. The parameter, called rate or intensity, is related to the expected (or average) number of Poisson points existing in some bounded region, [49] [50] where rate is usually used when the




Flashcard 1741238177036

Tags
#poisson-process #stochastics
Question
a homogeneous Poisson point process has a parameter of the form [...]
Answer
,

where is Lebegues measure, and is a constant

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
If a Poisson point process has a parameter of the form , where is Lebegues measure, and is a constant, then the point process is called a homogeneous or stationary Poisson point process.

Original toplevel document

Poisson point process - Wikipedia
edit source] For all the different settings of the Poisson point process, the two key properties [b] of the Poisson distribution and complete independence play an important role. [25] [45] Homogeneous Poisson point process[edit source] <span>If a Poisson point process has a parameter of the form Λ = ν λ {\displaystyle \textstyle \Lambda =\nu \lambda } , where ν {\displaystyle \textstyle \nu } is Lebegues measure, which assigns length, area, or volume to sets, and λ {\displaystyle \textstyle \lambda } is a constant, then the point process is called a homogeneous or stationary Poisson point process. The parameter, called rate or intensity, is related to the expected (or average) number of Poisson points existing in some bounded region, [49] [50] where rate is usually used when the







Flashcard 1741240536332

Tags
#poisson-process #stochastics
Question
In a homogeneous Poisson point process with , is [...] and is [...]
Answer
Lebegues measure, constant

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
If a Poisson point process has a parameter of the form , where is Lebegues measure, and is a constant, then the point process is called a homogeneous or stationary Poisson point process.

Original toplevel document

Poisson point process - Wikipedia
edit source] For all the different settings of the Poisson point process, the two key properties [b] of the Poisson distribution and complete independence play an important role. [25] [45] Homogeneous Poisson point process[edit source] <span>If a Poisson point process has a parameter of the form Λ = ν λ {\displaystyle \textstyle \Lambda =\nu \lambda } , where ν {\displaystyle \textstyle \nu } is Lebegues measure, which assigns length, area, or volume to sets, and λ {\displaystyle \textstyle \lambda } is a constant, then the point process is called a homogeneous or stationary Poisson point process. The parameter, called rate or intensity, is related to the expected (or average) number of Poisson points existing in some bounded region, [49] [50] where rate is usually used when the







Flashcard 1766930124044

Tags
#linear-algebra #matrix-decomposition
Question

The eigendecomposition of a real symmetric matrix can be represented as [...]

Answer

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill
Eigendecomposition of a matrix - Wikipedia
{\displaystyle A=A^{*}} ), which implies that it is also complex normal, the diagonal matrix Λ has only real values, and if A is unitary, Λ takes all its values on the complex unit circle. Real symmetric matrices[edit source] <span>As a special case, for every N×N real symmetric matrix, the eigenvalues are real and the eigenvectors can be chosen such that they are orthogonal to each other. Thus a real symmetric matrix A can be decomposed as A = Q Λ Q T {\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{T}} where Q is an orthogonal matrix, and Λ is a diagonal matrix whose entries are the eigenvalues of A. Useful facts[edit source] Useful facts regarding eigenvalues[edit source] The product of the eigenvalues is equal to the determinant of A det







Flashcard 1767406963980

Tags
#logic
Question
The Aristotelian conceptions of language and metaphysics fell out of favour in [...when...] with the rise of a new scientific paradigm.
Answer
the dawn of the modern era

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
ture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour <span>in the dawn of the modern era with the rise of a new scientific paradigm. <span><body><html>

Original toplevel document

The rise and fall and rise of logic | Aeon Essays
s Diafoirus resorts to disputational vocabulary to make a point about love: Distinguo, Mademoiselle; in all that does not concern the possession of the loved one, concedo, I grant it; but in what does regard that possession, nego, I deny it. <span>The fall of disputational culture wasn’t the only cause for the demise of scholastic logic, however. Scholastic logic was also viewed – rightly or wrongly – as being tied to broadly Aristotelian conceptions of language and metaphysics, which themselves fell out of favour in the dawn of the modern era with the rise of a new scientific paradigm. Despite all this, disputations continued to be practised in certain university contexts for some time – indeed, they live on in the ceremonial character of PhD defences. The point, thou







#has-images
HOW-TO Install CalculiX in Windows 10 with Windows Subsystem for Linux (WSL) from original source

With the recent updates in Microsoft Windows 10, it is possible to execute several of CLI programs for Linux directly at Windows, without the need of a virtualization tool such as VirtualBox. It installs a subsystem similar to what WINE does on Linux to run Windows software. The feature is called Windows Subsystem for Linux (WSL) and its available with the Creators Update of Windows 10. (More information about WSL here )

Windows Subsystem for Linux allows us to run CalculiX CrunchiX and GraphiX in Windows 10 using the official binaries available here. The main advantage of this is that it allows the user to have the latest update from the official source, decreasing room for instability in ported versions. As mentioned above, performance differences should be negligible.

If you still haven’t heard of CalculiX, I invite you to have a look to the official website. CCX is a FE solver totally open source. It uses a quite similar syntax as ABAQUS, so that different preprocessing tools can be used when selecting ABAQUS INP compatible input decks.

Is it fast?

The performance, while is not the same as native software, looks promising and constantly improving. For CPU intensive tasks it can be as good as native software! For intensive read/write operations is definitely slower, in case you plan to simulate with very big models.

In my personal tests with static contact problems I haven’t noticed differences.

Installation Procedure

Within a few steps, is possible to execute the binaries of CalculiX CrunchiX (CCX) and CalculiX GraphiX (CGX) that are directly available from the official CalculiX website.

Installation Steps for CalculiX CrunchiX

Install WSL

A good guide to do so is the official one from Microsoft here.

  1. Open PowerShell as Administrator and run:
 Enable-WindowsOptionalFeature -Online -FeatureName Microsoft-Windows-Subsystem-Linux
  1. Restart your computer when prompted.

Opening Bash for Windows

As explained in the aforementioned tutorial, you can access the Linux terminal (bash) after installing WSL by typing bash in the command line by accessing it through the Start Menu. There is another way that may become handy. By typing bash in the address bar in the Windows Explorer, it will be opened at the same working directory of the Explorer.

Microsoft does not recommend to edit files from the Linux subsystem from Windows software. Still, I have edited the files with Sublime Text without any problem.

After setting up your UNIX account, lets update the current installation by typing:

1 2 sudo apt-get update sudo apt-get upgrade

Install CCX executable

Once that the Bash is runni

...
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

HOW-TO Install CalculiX in Windows 10 with Windows Subsystem for Linux (WSL) from original source – Bits and Bolts
Contact [imagelink] Bits and Bolts A collection of ideas regarding computational mechanics Written by carlomontecJuly 10, 2017November 27, 2017 <span>HOW-TO Install CalculiX in Windows 10 with Windows Subsystem for Linux (WSL) from original source With the recent updates in Microsoft Windows 10, it is possible to execute several of CLI programs for Linux directly at Windows, without the need of a virtualization tool such as VirtualBox. It installs a subsystem similar to what WINE does on Linux to run Windows software. The feature is called Windows Subsystem for Linux (WSL) and its available with the Creators Update of Windows 10. (More information about WSL here ) Windows Subsystem for Linux allows us to run CalculiX CrunchiX and GraphiX in Windows 10 using the official binaries available here. The main advantage of this is that it allows the user to have the latest update from the official source, decreasing room for instability in ported versions. As mentioned above, performance differences should be negligible. If you still haven’t heard of CalculiX, I invite you to have a look to the official website. CCX is a FE solver totally open source. It uses a quite similar syntax as ABAQUS, so that different preprocessing tools can be used when selecting ABAQUS INP compatible input decks. Is it fast? The performance, while is not the same as native software, looks promising and constantly improving. For CPU intensive tasks it can be as good as native software! For intensive read/write operations is definitely slower, in case you plan to simulate with very big models. In my personal tests with static contact problems I haven’t noticed differences. Installation Procedure Within a few steps, is possible to execute the binaries of CalculiX CrunchiX (CCX) and CalculiX GraphiX (CGX) that are directly available from the official CalculiX website. Installation Steps for CalculiX CrunchiX Install WSL A good guide to do so is the official one from Microsoft here. Open PowerShell as Administrator and run: Enable-WindowsOptionalFeature -Online -FeatureName Microsoft-Windows-Subsystem-Linux Restart your computer when prompted. Opening Bash for Windows As explained in the aforementioned tutorial, you can access the Linux terminal (bash) after installing WSL by typing bash in the command line by accessing it through the Start Menu. There is another way that may become handy . By typing bash in the address bar in the Windows Explorer, it will be opened at the same working directory of the Explorer. Microsoft does not recommend to edit files from the Linux subsystem from Windows software. Still, I have edited the files with Sublime Text without any problem. After setting up your UNIX account, lets update the current installation by typing: 1 2 sudo apt-get update sudo apt-get upgrade Install CCX executable Once that the Bash is running we can proceed to install CCX. In order to get both programs running, we will need some additional libraries and compilers. Type in the bash the following command: 1 sudo apt-get install -y gfortran gcc Then we will download, extract and move the CalculiX executable for Linux, directly from the official website. 1 2 3 4 5 wget http: //www .dhondt.de /ccx_2 .13. tar .bz2 bunzip2 ccx_2.13. tar .bz2 tar -xvf ccx_2.13. tar sudo mv ccx_2.13 /usr/local/bin/ccx_2 .13 chmod ao+rx /usr/local/bin/ccx_2 .13 # change permissions of file As last step we will set the amount of logical processors to work with OpenMP in our .bashrc file, which has several default settings of the bash. In the case of my computer (i7-6700K) is 8 but should be set accordingly for each CPU. It can be checked following this instructions. 1 2 echo "export OMP_NUM_THREADS=8" >> ~/.bashrc source ~/.bashrc Now it should be ready to run by using the command ccx_2.13 . Install CalculiX GraphiX There no official support for software with graphic interface in Bash for Windows for now, but some workarounds have been found for a very good way to do this. The installation process is similar, but we will need couple extra steps for the graphic libraries to run. Install Xming The first step is to install a Xserver for Windows. A good alternative is Xming. Follow the default installation options. In order to use it, run Xming and you will notice in the tray bar ready to be used. Install Gedit Before installing CGX, we still need some libraries to run GUI software. An easy way to install the remaining libraries is to install Gedit from apt-get: 1 sudo apt-get install -y gedit I also installed Gvim and even Thunar and seemed to worked reasonably well. Install CGX executable Then, we can proceed to install CGX with the following commands: 1 2 3 4 wget http: //www .dhondt.de /cgx_2 .13.bz2 bunzip2 cgx_2.13.bz2 sudo mv cgx_2.13 /usr/local/bin/cgx_2 .13 chmod ao+rx /usr/local/bin/cgx_2 .13 # change permissions of file Ready! Have a look at the official documentation here to learn how to use CalculiX. Enjoy Advertisements Report this ad Report this ad Leave a Reply Cancel r