Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[id] cs-229-unsupervised-learning #139

Open
wants to merge 35 commits into
base: master
Choose a base branch
from
Open
Changes from 1 commit
Commits
Show all changes
35 commits
Select commit Hold shift + click to select a range
54096cd
Create folder id for Indonesia translation
swicaksono Apr 5, 2019
f96d857
Delete id
swicaksono Apr 5, 2019
1c1c7ca
Create README.md
swicaksono Apr 5, 2019
7522015
Update README.md
swicaksono Apr 5, 2019
0bdc9d7
Update README.md
swicaksono Apr 5, 2019
1f08af9
Add cheatsheet-deep-learning
swicaksono Apr 5, 2019
e1e4953
Rename template/cheatsheet-machine-learning-tips-and-tricks.md to id/…
swicaksono Apr 5, 2019
b828f96
Rename id/cheatsheet-machine-learning-tips-and-tricks.md to template/…
swicaksono Apr 5, 2019
e8141a6
Translating into Bahasa Indonesia
swicaksono Apr 5, 2019
2d6b7ea
Update cheatsheet-deep-learning.md
swicaksono Apr 6, 2019
dac448f
Update cheatsheet-deep-learning.md
swicaksono Apr 6, 2019
1e5a359
Update cheatsheet-deep-learning.md
swicaksono Apr 8, 2019
68b90cf
Update cheatsheet-deep-learning.md
swicaksono Apr 8, 2019
1bf5e3f
Fixing typo
swicaksono Apr 9, 2019
597295d
Adding some files and fixing typo
swicaksono Apr 9, 2019
344a424
Translating machine learning tips and tricks into Bahasa
swicaksono Apr 9, 2019
9b4f85d
Add translation to Bahasa Indonesia
swicaksono Apr 9, 2019
f1b014d
Translating some numbers into Bahasa Indonesia
swicaksono Apr 9, 2019
c819837
Translating into Bahasa Indonesia
swicaksono Apr 10, 2019
f3f225f
Translating cheatsheet-supervised-learning to Indonesia language.
swicaksono Apr 11, 2019
658ca3f
Translating cheatsheet-supervised-learning
swicaksono Apr 12, 2019
b0fa838
Translating cheatsheet supervised learning
swicaksono Apr 15, 2019
2916693
Translating into Bahasa Indonesia
swicaksono Apr 15, 2019
941c931
Translating into Indonesia Language
swicaksono Apr 16, 2019
dbffb1a
Translating supervised learning into Indonesia language
swicaksono Apr 18, 2019
12a47ab
Translate cheatsheet unspervised learning into Indonesia language
swicaksono May 6, 2019
c07c746
Translating cheatsheet unspervised learning into Indonesia
swicaksono May 6, 2019
625b6b6
Delete deep-learning-tips-and-tricks.md
shervinea May 29, 2019
9b578a9
Delete convolutional-neural-networks.md
shervinea May 29, 2019
981c088
Delete recurrent-neural-networks.md
shervinea May 29, 2019
ebe5952
Delete refresher-linear-algebra.md
shervinea May 29, 2019
1666ef9
Delete refresher-probability.md
shervinea May 29, 2019
80e9009
review unspervised learning
swicaksono Oct 19, 2019
252530a
Merge branch 'master' of https://github.com/swicaksono/cheatsheet-tra…
swicaksono Oct 19, 2019
303057c
review cheatsheet-unsupervised-learning
swicaksono Oct 19, 2019
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
23 changes: 12 additions & 11 deletions id/cheatsheet-deep-learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@

**11. Learning rate ― The learning rate, often noted α or sometimes η, indicates at which pace the weights get updated. This can be fixed or adaptively changed. The current most popular method is called Adam, which is a method that adapts the learning rate.**

⟶**11. Learning rate - Learning rate (Tingkat pembelajaran), sering dinotasikan sebagai α atau η, merupakan fase pembaruan pembobotan. Tingkat pembelajaran dapat diperbaiki atau diubah secara adaptif. Metode yang paling populer saat ini disebut Adam, yang merupakan metode yang dapat menyesuaikan tingkat pembelajaran.
⟶**11. Learning rate - Learning rate (Tingkat pembelajaran), sering dinotasikan sebagai α atau η, merupakan fase pembaruan pembobotan. Tingkat pembelajaran dapat diperbaiki atau diubah secara adaptif. Metode yang paling populer saat ini disebut Adam, yang merupakan metode yang dapat menyesuaikan tingkat pembelajaran.**
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Learning rate- Learning rate, sering dinotasikan sebagai α atau η, mendefinisikan seberapa cepat nilai weight diperbaharui. Learning rate bisa diset dengan nilai fix atau dirubah secara adaptif. Metode yang paling terkenal saat ini adalah Adam, sebuah method yang merubah learning rate secara adaptif.


<br>

Expand All @@ -78,61 +78,62 @@

**14. Updating weights ― In a neural network, weights are updated as follows:**

&#10230;
&#10230;**14. Memperbaharui bobot w - Dalam neural network, bobot w diperbarui nilainya dengan cara berikut:**

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Memperbaharui nilai weights - Dalam neural network, nilai weights diperbaharui nilainya dengan cara berikut:

Menurut saya weights tidak usah diterjemahkan, karena merupakan kosakata teknis pada neural network

<br>

**15. Step 1: Take a batch of training data.**

&#10230;
&#10230;**15. Langkah 1: Mengambil jumlah batch dari data latih.**

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. Langkah 1: Mengambil batch (sample data) dari keseluruhan training data.

<br>

**16. Step 2: Perform forward propagation to obtain the corresponding loss.**

&#10230;
&#10230;**16. Langkah 2: Melakukan forward propagation untuk mendapatkan nilai loss yang sesuai. **

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. Langkah 2: Melakukan forward propagation untuk mendapatkan nilai loss berdasarkan nilai masukan (input).

<br>

**17. Step 3: Backpropagate the loss to get the gradients.**

&#10230;
&#10230; **17. Langkah 3: Melakukan backpropagate terhadap loss untuk mendapatkan gradient.**

<br>

**18. Step 4: Use the gradients to update the weights of the network.**

&#10230;
&#10230;**18. Langkah 4: Menggunakan gradient untuk untuk memperbarui nilai dari network.**

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. Langkah 4: Menggunakan gradient untuk untuk memperbarui nilai weight dari network.

<br>

**19. Dropout ― Dropout is a technique meant at preventing overfitting the training data by dropping out units in a neural network. In practice, neurons are either dropped with probability p or kept with probability 1−p**

&#10230;
&#10230;**19. Dropout - Dropout adalah teknik untuk mencegah overfitting data latih dengan menghilangkan satu atau lebih unit layer dalam neural network. Pada praktiknya, neurons melakukan drop dengan probabilitas p atau tidak melakukannya dengan probabilitas 1-p**
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. Dropout - Dropout adalah teknik untuk mencegah model overfit terhadap training data dengan menghilangkan satu atau lebih unit layer dalam neural network. Pada praktiknya, neurons di-drop dengan probabilitas p atau dipertahankan dengan probabilitas 1-p


<br>

**20. Convolutional Neural Networks**

&#10230;
&#10230; **20. Convolutional Neural Networks**

<br>

**21. Convolutional layer requirement ― By noting W the input volume size, F the size of the convolutional layer neurons, P the amount of zero padding, then the number of neurons N that fit in a given volume is such that:**

&#10230;
&#10230; **21. Kebutuhan layer convolutional - W adalah ukuran volume input, F adalah ukuran dari layer neuron convolutional, P adalah jumlah zero padding, maka jumlah neurons N yang dapat dibentuk dari volume yang diberikan adalah: **

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maka jumlah neuron N yang sesuai dengan ukuran dimensi masukan adalah:

<br>

**22. Batch normalization ― It is a step of hyperparameter γ,β that normalizes the batch {xi}. By noting μB,σ2B the mean and variance of that we want to correct to the batch, it is done as follows:**

&#10230;
&#10230; **22. Batch normalization - Adalah salah satu step hyperparameter γ,β yang menormalisasikan batch {xi}. Dengan notasi μB,σ2B adalah rata-rata dan variansi nilai yang digunakan untuk perbaikan dalam batch, dapat diselesaikan sebagai berikut:**

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Dengan mendefinisikan μB,σ2B sebagai nilai rata-rata dan variansi dari batch yang ingin kita normalisasi, hal tersebut dapat dilakukan dengan cara:

<br>

**23. It is usually done after a fully connected/convolutional layer and before a non-linearity layer and aims at allowing higher learning rates and reducing the strong dependence on initialization.**

&#10230;
&#10230; **23. Biasanya dilakukan setelah layer sepenuhnya terhubung / konvolusional dan sebelum layer non-linearitas, yang bertujuan untuk peningkatan tingkat pembelajaran yang lebih tinggi dan mengurangi ketergantungan yang kuat pada inisialisasi.**
swicaksono marked this conversation as resolved.
Show resolved Hide resolved


<br>

Expand Down