lnshi / ml-exercises

Have some fun with ML. πŸ•΅πŸ€–πŸ§ πŸ€”

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Project GitHub Pages

Topics

ml_basics

Questions

  1. In gradient descent, must there be a learning rate transition point(safety threshold) for all kinds of cost functions?

  2. How do we extend this to the cross product of a four dimensional vector or more higher, like the right part of the above graph?

  3. When a lower dimensional space NOT discriminable dataset is PROJECTED to a PROPER higher dimensional space it always will be discriminable, the boundary is a hyper plane or just a discrimination function, what are the differences of the 'a hyper plane' or 'a discrimination function' here?

  4. What are the best practices / skills / underlying theories for the features expansion/extraction?

  5. For the hand-written digits recognition NN why the hidden layer input size is 25? How about design to use a different size? What is the theory/best practices/skills/prior knowledges here?

Accumulations / References

  1. ε¦‚δ½•η†θ§£ζœ€ε°δΊŒδΉ˜ζ³•οΌŸ

  2. np.array([0, 0]) vs np.array([0., 0.])

    >>> import numpy as np
    >>> t = np.array([0, 0])
    >>> t[0] = 0.97
    >>> t
    array([0, 0])
    >>> t[0] = 1.97
    >>> t
    array([1, 0])
    >>> t = np.array([0., 0.])
    >>> t[0] = 0.97
    >>> t
    array([0.97, 0.  ])
    >>> t = np.array([0, 0.])
    >>> t[0] = 0.97
    >>> t
    array([0.97, 0.  ])
    

Memos

  1. Exponential family form of multinomial distribution <- my answer -> link to topic 'GLM and exponential family distributions'

  2. Batch Normalization Tensorflow Keras Example

About

Have some fun with ML. πŸ•΅πŸ€–πŸ§ πŸ€”

License:MIT License


Languages

Language:Jupyter Notebook 100.0%