cherryman
Member level 1
- Joined
- Feb 28, 2013
- Messages
- 32
- Helped
- 9
- Reputation
- 18
- Reaction score
- 9
- Trophy points
- 1,288
- Location
- Poland
- Activity points
- 1,477
Hello all, I have several basic questions about Machine Learning to better solidify my knowledge.
I hope to find the answers here.
1. SVM
I have a one dimensional feature vector (x) of length N and the target vector that haves labels of two classes (e.g. 0 and 1). The case is not linearly separable, so the histograms of those features given classes are overlapped.
Can You explain me, why the SVM generates different hypothesis according to the different kernel functions?
Is it related with that, for the SVM it is actually not a linear case? And when I use a kernel function I actually make a transformation z = phi(x)?
2. Kernels
Can You explain me the basic concept of a kernel? What it actually do to the data set? I know that it is an inner product of the given data (x'x) but how it affect on the data set?
3 SVM and classes with uneven number of representants
I heard that SVM is sensitive to uneven classes (classes with uneven number of representants). Can You tell me is it true? And if it is, can You tell me why?
On the other hand to evaluate hyperplane the SVM needs only support vectors, and other data are not necessary except those, which are misclassified
in soft margin SVM.
4. Data normalization
Having a set of M real signals, I extracted a set of 2D - features from them. From one signal there are N 2D features, so as a result I have the M matrices with Nx2 dimension. The task is to recognise 2 classes (0 and 1) from them. But the one class occurs much frequent than the other.
The main problem that I have is as follows.
Within each signal (matrix) the features have quite good separation, but when I want to create a training vector from all of the signals, the features from different classes are highly overlap. So I suspect, the problem is in normalization. But how to make good normalization? I have found that there is a good practice to subtract mean value from the features and then divide them by max(abs(x)) or standard deviation.
But if I do that on the vector that have features form two classes, the result will be different comparing to the vector with only one class representation.
The second hypothesis is that I have extracted weak features.
Probably I'll have more questions, but right now that is all. If somebody can help me I'll be very thankful.
I hope to find the answers here.
1. SVM
I have a one dimensional feature vector (x) of length N and the target vector that haves labels of two classes (e.g. 0 and 1). The case is not linearly separable, so the histograms of those features given classes are overlapped.
Can You explain me, why the SVM generates different hypothesis according to the different kernel functions?
Is it related with that, for the SVM it is actually not a linear case? And when I use a kernel function I actually make a transformation z = phi(x)?
2. Kernels
Can You explain me the basic concept of a kernel? What it actually do to the data set? I know that it is an inner product of the given data (x'x) but how it affect on the data set?
3 SVM and classes with uneven number of representants
I heard that SVM is sensitive to uneven classes (classes with uneven number of representants). Can You tell me is it true? And if it is, can You tell me why?
On the other hand to evaluate hyperplane the SVM needs only support vectors, and other data are not necessary except those, which are misclassified
in soft margin SVM.
4. Data normalization
Having a set of M real signals, I extracted a set of 2D - features from them. From one signal there are N 2D features, so as a result I have the M matrices with Nx2 dimension. The task is to recognise 2 classes (0 and 1) from them. But the one class occurs much frequent than the other.
The main problem that I have is as follows.
Within each signal (matrix) the features have quite good separation, but when I want to create a training vector from all of the signals, the features from different classes are highly overlap. So I suspect, the problem is in normalization. But how to make good normalization? I have found that there is a good practice to subtract mean value from the features and then divide them by max(abs(x)) or standard deviation.
But if I do that on the vector that have features form two classes, the result will be different comparing to the vector with only one class representation.
The second hypothesis is that I have extracted weak features.
Probably I'll have more questions, but right now that is all. If somebody can help me I'll be very thankful.