Edited, memorised or added to reading queue

on 05-Jun-2024 (Wed)

Do you want BuboFlash to help you learning these things? Click here to log in or create user.

#feature-engineering #lstm #recurrent-neural-networks #rnn
As the number of hyperparameters and their range grow, the search space becomes exponentially complex, and tuning the models manually or by grid-search becomes impractical . Bayesian optimization for hyperparameter tuning provides hyperparameters (step 1) iteratively based on previous performance (Shahriari, Swersky, Wang, Adams, & De Freitas, 2015). We use Bayesian optimization to search the hyperparameter space for our model extensively.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

pdf

cannot see any pdfs




#feature-engineering #lstm #recurrent-neural-networks #rnn
Interestingly, the LSTM model beats both benchmark model across the board and performs well at all lift levels.
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

pdf

cannot see any pdfs




#deep-learning #keras #lstm #python #sequence

Below are some common configurations for the batch size:

batch size=1 :
Weights are updated after each sample and the procedure is called stochas- tic gradient descent.

batch size=32 :
weights are updated after a specified number of samples and the procedure is called mini-batch gradient descent. Common values are 32, 64, and 128, tailored to the desired efficiency and rate of model updates. If the batch size is not a factor of the number of samples in one epoch, then an additional batch size of the left over samples is run at the end of the epoch.

batch size=n :
Where n is the number of samples in the training dataset. Weights are updated at the end of each epoch and the procedure is called batch gradient descent

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

pdf

cannot see any pdfs




Flashcard 7629495209228

Tags
#advanced #deep-learning #keras #python
Question

# Import the sigmoid function from scipy

from scipy.special import expit as [...]

# Weight from the model

weight = 0.14

# Print the approximate win probability predicted close game

print(sigmoid(1 * 0.14))

# Print the approximate win probability predicted blowout game

print(sigmoid(10 * 0.14))

Answer
sigmoid

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Open it
# Import the sigmoid function from scipy from scipy.special import expit as sigmoid # Weight from the model weight = 0.14 # Print the approximate win probability predicted close game print(sigmoid(1 * 0.14)) # Print the approximate win probability predicted blowout gam







#feature-engineering #lstm #recurrent-neural-networks #rnn
As the number of hyperparameters and their range grow, the search space becomes exponentially complex, and tuning the models manually or by grid-search becomes impractical . Bayesian optimization for hyperparameter tuning provides hyperparameters (step 1) iteratively based on previous performance
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
As the number of hyperparameters and their range grow, the search space becomes exponentially complex, and tuning the models manually or by grid-search becomes impractical . Bayesian optimization for hyperparameter tuning provides hyperparameters (step 1) iteratively based on previous performance (Shahriari, Swersky, Wang, Adams, & De Freitas, 2015). We use Bayesian optimization to search the hyperparameter space for our model extensively.

Original toplevel document (pdf)

cannot see any pdfs




Flashcard 7629499403532

Tags
#deep-learning #keras #lstm #python #sequence
Question
The [...], or LSTM, network is a type of Recurrent Neural Network.
Answer
Long Short-Term Memory

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
The Long Short-Term Memory, or LSTM, network is a type of Recurrent Neural Network.

Original toplevel document (pdf)

cannot see any pdfs







Flashcard 7629500976396

Tags
#DAG #causal #edx #has-images
[unknown IMAGE 7093205732620]
Question
For example, suppose that the causal DAG includes an [...] common cause of A and Y, U and also a measured variable L that is an effect of U.
Answer
unmeasured

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
For example, suppose that the causal DAG includes an unmeasured common cause of A and Y, U and also a measured variable L that is an effect of U.

Original toplevel document (pdf)

cannot see any pdfs







#causality #statistics
We havte to ensure that there is no non-causal association flowing between 𝑋 and π‘Œ . This is true if 𝑋 and π‘Œ are d-separated in the augmented graph where we remove outgoing edges from 𝑋 .
statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it

Original toplevel document (pdf)

cannot see any pdfs




#ML-engineering #ML_in_Action #learning #machine #software-engineering

ML engineers

They need just enough data engineering skills to build (and schedule the ETL for) feature datasets for their models, but not to construct a petabyte-scale streaming ingestion framework.

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on


Parent (intermediate) annotation

Open it
ust enough software development skills to be able to write modular code and implement unit tests. They don’t need to know about the intricacies of non-blocking asynchronous messaging brokering. <span>They need just enough data engineering skills to build (and schedule the ETL for) feature datasets for their models, but not to construct a petabyte-scale streaming ingestion framework. They need just enough visualization skills to create plots and charts that communicate clearly what their research and models are doing, but not to develop dynamic web apps that have co

Original toplevel document (pdf)

cannot see any pdfs




Flashcard 7629509889292

Tags
#causality #has-images #statistics


Question

There are two categories of things that could go wrong if we condition on descendants of 𝑇:

1. We block the flow of causation from 𝑇 to π‘Œ.

2. We [...] between 𝑇 and π‘Œ.

Answer
induce non-causal association

statusnot learnedmeasured difficulty37% [default]last interval [days]               
repetition number in this series0memorised on               scheduled repetition               
scheduled repetition interval               last repetition or drill

Parent (intermediate) annotation

Open it
There are two categories of things that could go wrong if we condition on descendants of 𝑇: 1. We block the flow of causation from 𝑇 to π‘Œ. 2. We induce non-causal association between 𝑇 and π‘Œ.

Original toplevel document (pdf)

cannot see any pdfs