这篇稿子应博主Ju的邀约而写，主要是总结一下北美博士找咨询工作的经验和教训。我的背景是国内Top2本科，哥大生物专业博士，下个月入职一家咨询公司的New York Office。在求职期间得到了很多的帮助，希望也能尽一份自己的绵薄之力，帮到更多的后来人。
Now what we have discussed unconstrained optimization problems in previous post, it is now time to come to the reality. In the real world, we often have limitations, such as the total budget, motion angles, and some arbitrary desirable range of values. Life would be so easy (and boring) without boundary and conditions. Adding constraints certainly makes optimization problems less easy, but more interesting.
Give me a descent direction and a step length to move and I will find the optimum.
Why do we care about eigenvalues, eigenvectors, and singular values? Intuitively, what do they tell us about a matrix? When I first studied eigenvalues in college, I regarded it as yet another theoretical math trick that is hardly applicable to my life. Once I passed the final exam, I shelved all my eigen-knowledge to a corner in my memory. Years have passed, and I gradually realize the importance and brilliance of eigenvalues, particularly in the realm of machine learning. In this post, I will discuss how and why we perform eigendecomposition and singular value decomposition in machine learning.
In machine learning, we are often dealing with high-dimension data. For convenience, we often use matrix to represent data. Numerical optimization in machine learning often involves matrix transformation and computation. To make matrix computation more efficiently, we always factorize a matrix into several special matrices such as triangular matrices and orthogonal matrices. In this post, I will review essential concepts of matrix used in machine learning.
During logistic regression, in order to compute the optimal parameters in the model, we have to use an iterative numerical optimization approach (Newton method or Gradient descent method, instead of a simple analytical approach). Numerical optimization is a crucial mathematical concept in machine learning and function fitting, and it is deeply integrated in model training, regularization, support vector machine, neural network, and so on. In the next few posts, I will summarize key concepts and approaches in numerical optimization, and its application in machine learning.
Last February, I defended my PhD thesis and graduated from more than 2 decades of school life. Now, it’s been a full year of post-school life. There are no more exams and curriculum to quantify my GPA. In this post-school life, I start to realize that I have to be the one setting my own goal, designing my own curriculum, and evaluating my progress introspectively. In this post, I am sharing some lessons that I find useful in DIY curriculum.
In the previous post, I have discussed loss function in regression. In this post, I will elaborate how we develop loss function in classification when the output is discrete, rather than continuous.