MLpedia is supposed to be a ML playground. the web application can apply various ML algorithms to simple data and saved the learned models to database. users can later use this database. There are also data visualization tools available too.
I have to say even though I am quite happy with the project, it is very counter intutive and it does not have lot of guidance in how to use it. hence this section.
- The plotter section has some very simple plotting tools. which are farely self explanatory. you can understand very easily what to do by looking at the forms. The only important thing to here is how to enter the X values and Y values . Enter same number of data points in both the field and make sure they are comma seperated with a single space after each comma. as an example consider the following, also added a screenshot-
Line1 X values: 5, 10, 15, 20
Line1 Y values: 17, 23, 5, 10
-
To use the Analytica section first go ahed and make an account. this step is necessary if you want to use the analytica section of the application
-
Most of the algo here takes in two files. One of which is the features file. this is a simple csv file. each line describes the feature input of a single data point. the other one is labels file. This is also a csv file but containing a single value each line. the value of ith line of this file should be the lable of ith input feature of the features file . There is one more important thing to note. in case of neural network the first line of the labels file is not the conventional field name of a csv. it should be modified to the number of labels that are present in the file. I have provided some example files to try with. here is the link example inputs
-
after the learned model is saved you can do prediction on new datapoint.(alas! you can do it on anly one datapoint at a time, future scope of work.). again fill in a new feature vactor in the form and hit submit. features vector is values seperated by commas. again remember to add a single space after each value.
-
Remember this site is supposed to handle simple datasets. algorythms does not have normalization layers and probably will suffer from vanishing gradiant problem.(future scope of work)