Transfer Learning with various architectures of Convolutional Networks.
ResNet, AlexNet, VGG
Code – https://towardsdatascience.com/transfer-learning-using-keras-d804b2e04ef8
To understand bootstrap, suppose it were possible to draw repeated samples (of the same size) from the population of interest, a large number of times. Then, one would get a fairly good idea about the sampling distribution of a particular statistic from the collection of its values arising from these repeated samples. The idea behind bootstrap is to use the data of a sample study at hand as a “surrogate population”, for the purpose of approximating the sampling distribution of a statistic; i.e. to resample (with replacement) from the sample data at hand and create a large number of “phantom samples” known as bootstrap samples.
In other words, We randomly sample with replacement from the n known observations. We then call this a bootstrap sample. Since we allow for replacement, this bootstrap sample most likely not identical to our initial sample. Some data points may be duplicated, and others data points from the initial may be omitted in a bootstrap sample.
The following numerical example will help to demonstrate how the process works. If we begin with the sample 2, 4, 5, 6, 6, then all of the following are possible bootstrap samples:
2 ,5, 5, 6, 6
4, 5, 6, 6, 6
2, 2, 4, 5, 5
2, 2, 2, 4, 6
2, 2, 2, 2, 2
4,6, 6, 6, 6
Bootstrap aggregating (bagging) is a machine learning ensemble meta-algorithm designed to improve the stability and accuracy of machine learning algorithms used in statistical classification and regression. It also reduces variance and helps to avoid overfitting. Although it is usually applied to decision tree methods, it can be used with any type of method
Some of the useful Ubuntu commands along with its descriptions. This might be used as a handy reference to quickly know about the syntax of a command.
|wget URL||Downloads the file specified by the URL.|
|pwd||Displays the current directory you are in.|
|sudo||Allows the user to act like a superuser|
|sudo -i||Allows the user to get root access.|
|sudo apt-get install package_name||Allows the user to act like a superuser and install packages|
|cd directory_name||Changes from current directory to the mentioned directory.|
|cd ..||Moves back one directory|
|ls||To view the contents in a directory including files and sub-directories.|
|ls -a||To view the contents in a directory including hidden files.|
|man command||Displays the information about the command specified.|
|whereis file/directory||Shows where the specified file/directory is.|
|mkdir directory_name||Creates a directory with the given name.|
|mv oldname newname||Renames the file.|
|rm filename||Removes the specified filename.|
|rmdir directoryname||Removes the specified empty directory.|
|rm -r directoryname||Removes files and sub-directories in the specified directory.|
|ifconfig & iwconfig||Allows the user to look at the network configuration.|
|ping URL||Allows the user to test connectivity issues.|
|vi filename||Opens the specified file in the vi editor to view/make changes.|
|telnet ip_address||Connects to the specified IP address.|
|chmod 777 file_name||Modifies the permissions of the specified file.
The three digits in 777 represents users, groups and others.
|chmod -R 777 directory||Modifies the permissions of the specified directory recursively. Meaning it applies the changes for all the files and sub-directories.|
Fermi questions are questions designed to test your logic, and you may have encountered them before. So, they usually are formed or phrased in a way where the person asking you this question designs a scenario, asks you about a real world scenario, and asks you to provide an estimate very quickly. Here are few examples. Try to solve these 🙂
What is an eGPU?
GPU stands for Graphics Processing Unit, which is more commonly referred to as a Video Card or Graphics Chip. The “e” prefix stands for “external”. In short, an eGPU is the act of hooking up a desktop video card to a laptop, or a SFF system lacking actual desktop-sized slots (such as an Intel NUC).
But why?! Wouldn’t a desktop make more sense? Wouldn’t a desktop erform better?
Why not? Maybe. Most of the time.
Not all people want a desktop. Desktops are (typically) large, (typically) bulky and (by definition) immobile. There is convenience in having your own system with you on the road, while still being able to game in the comfort of your own home, without having to sync any data, or switch systems. One system is convenient, two are less so.
That said, if there is no particular wish to use a single machine, or if a laptop is not needed/desired in the first place, a desktop machine is undeniably superior (and often cheaper, if we compare the price of a laptop+eGPU setup to the price of a desktop built from scratch). There are a few considerations that make eGPUs desirable: Already owning a laptop (often a high-end one, for example, due to the requirements of an occupation) and wishing to be able to game on it, having an older laptop that could use a little boost in the arm in the graphics department but is otherwise perfectly usable, and saving space (because an eGPU plus a laptop take up very little space and are easier to fit into a small apartment or a dorm).
A desktop system with a near top-of-the-line desktop CPU (like the Intel i7 6700K or i5 6600K) and a given video card will nearly always outperform a laptop (no matter how high end the laptop is) with the same video card connected as an eGPU. This is an undeniable fact. However, eGPU performance can range from ~70% to ~95% of the equivalent desktop performance (depending on how the eGPU is connected to the laptop, whether you are using the internal or an external display, the game in question, the resolution you are at, as well as the frame rate you are getting), so the performance is still definitely there and is definitely viable…………………………………………. More
Source – https://www.reddit.com/r/eGPU/comments/5jpf2x/diy_egpu_101_introduction_to_egpu/
LibriSpeech ASR corpus
Regularization is a technique used in an attempt to solve the overfitting problem in statistical models.*
First of all, I want to clarify how this problem of overfitting arises.
When someone wants to model a problem, let’s say trying to predict the wage of someone based on his age, he will first try a linear regression model with age as an independent variable and wage as a dependent one. This model will mostly fail, since it is too simple.
Then, you might think: well, I also have the age, the sex and the education of each individual in my data set. I could add these as explaining variables.
Your model becomes more interesting and more complex. You measure its accuracy regarding a loss metric where is your design matrix and is the observations (also denoted targets) vector (here the wages).
You find out that your result are quite good but not as perfect as you wish.
So you add more variables: location, profession of parents, social background, number of children, weight, number of books, preferred color, best meal, last holidays destination and so on and so forth.
Your model will do good but it is probably overfitting, i.e. it will probably have poor prediction and generalization power: it sticks too much to the data and the model has probably learned the background noise while being fit. This isn’t of course acceptable.
So how do you solve this?
It is here where the regularization technique comes in handy.
You penalize your loss function by adding a multiple of an (LASSO) or an (Ridge) norm of your weights vector (it is the vector of the learned parameters in your linear regression). You get the following equation:
( is either the , or any other norm)
This will help you avoid overfitting and will perform, at the same time, features selection for certain regularization norms (the in the LASSO does the job).
Finally you might ask: OK I have everything now. How can I tune in the regularization term ?
One possible answer is to use cross-validation: you divide your training data, you train your model for a fixed value of and test it on the remaining subsets and repeat this procedure while varying . Then you select the best that minimizes your loss function.
I hope this was helpful. Let me know if there is any mistakes. I will try to add some graphs and eventually some R or Python code to illustrate this concept.
Also, you can read more about these topics (regularization and cross validation) here:
* Actually this is only one of the many uses. According to Wikipedia, it can be used to solve ill-posed problems. Here is the article for reference:.
As always, make sure to follow me for more insights about machine learning and its pitfalls:
Source – https://www.quora.com/What-is-regularization-in-machine-learning