Pages: [1] 2 3 ... 10
 on: Today at 12:06:09 AM 
Started by fvbnju765r - Last post by fvbnju765r
'set role' don't work. always told me that the attribute I set is  ' ' nothing or don't find.
I have to type the code in xml by myself to make it work.
look's you can't set something only by mouse click when you use 'set role'.

 on: October 24, 2014, 08:54:43 PM 
Started by qwertz - Last post by qwertz
Dear all,

having recently started with writing scripts, I have an idea what causes this error but unfortunately no idea how to solve it  Undecided

This piece of code is supposed to create a new attribute with a prefix in the name for each attribute in the example set.
(e.g. "att1" leads to "prefix_att1" and "att2" leads to "prefix_att2")

I suppose the error occurs because the for loop shall iterate over each attribute in the example set but while doing so the code creates new attributes and adds them to the example set.

Additionally, the examples of all newly created attributes - and only the newly created - shall be edited (e.g. set to "123").

PS: I forgot to say: When I run this routine a second time (in the same script right after the first loop) it would like to create new attributes again (let's say prefix2_att1) but only for the original attributes an not those that have been recently created during the first loop.

Does anyone have a hint how to proceed?

<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<process version="6.1.000">
  <operator activated="true" class="process" compatibility="6.1.000" expanded="true" name="Process">
    <process expanded="true">
      <operator activated="true" class="generate_data" compatibility="6.1.000" expanded="true" height="60" name="Generate Data" width="90" x="45" y="30"/>
      <operator activated="true" class="execute_script" compatibility="6.1.000" expanded="true" height="76" name="Execute Script" width="90" x="179" y="30">
        <parameter key="script" value="import;&#10;&#10;ExampleSet es = input [0];&#10;&#10;for ( Attribute att : es.getAttributes() )&#10;{&#10;&#9;String att_name = &quot;prefix&quot; + att.getName();&#10;&#9;Attribute newAttribute = AttributeFactory.createAttribute( att_name, Ontology.ATTRIBUTE_VALUE_TYPE.NUMERICAL );&#10;&#9;es.getExampleTable().addAttribute( newAttribute );&#10;&#9;es.getAttributes().addRegular( newAttribute );&#10;&#9;es.getExample( 0 ).setValue( att, 123 );&#10;}&#10;&#10;return input;"/>
      <connect from_op="Generate Data" from_port="output" to_op="Execute Script" to_port="input 1"/>
      <connect from_op="Execute Script" from_port="output 1" to_port="result 1"/>
      <portSpacing port="source_input 1" spacing="0"/>
      <portSpacing port="sink_result 1" spacing="0"/>
      <portSpacing port="sink_result 2" spacing="0"/>

Kind regards

 on: October 24, 2014, 06:59:53 PM 
Started by qwertz - Last post by qwertz
I found it Smiley See,8345.0.html

 on: October 24, 2014, 06:58:36 PM 
Started by al - Last post by Jono Gallagher
Hello and thank you for your questions about RapidMiner Cloud.

Let us answer your questions and shed some more light on the RapidMiner Cloud offering.

There are a few things you should keep in mind when making these kind of comparisons:

1) The Cloud machines may have to be spawned when needed. Depending on the current usage, your job may immediately get a free machine or it will have to wait for a new one to spawn and be prepared. This can take a couple of minutes before your job is actually executed. However this time is NOT included in the consumed time of your process, so you do not pay for that.

2) Jobs you submit to the Cloud run in a fully virtualized environment to protect the integrity of your own job and any subsequent jobs. This reduces performance a bit, but it provides a major boost in security.

3) Once the machine is prepared, your job will be started on the machine in question. However this does not yet equal the press of your local "Play" button, as a fresh execution environment needs to be set up first. While this does not take long, it takes a few seconds.

Taking the above factors into account, the execution time of your process actually beat your state-of-the-art laptop by a few seconds.

Regarding the next question, why was X-Large slower than Regular? Of course we also ran a multitude of performance tests on these machines, so here is what is going on:

1) Your process is very short. Couple this with full virtualization and you are bound to get some time variance. In our tests for equal jobs (which were not maxing out memory usage) X-Large was usually faster, however there were also times where Regular beat X-Large.

2) In terms of computing power, all machines use state-of-the-art Intel Server CPUs. The real difference (as shown in the description of each machine) is the available memory. If you submit a job that is not memory bound, your best bet is a Regular machine.

On to your last question, why should you push processes to RapidMiner Cloud? Let's go through some scenarios where the Cloud will come in quite handy!

- You can use operators otherwise limited to bigger licenses. For example you can access Twitter with a RapidMiner Starter license, as long as you have an active Cloud subscription. Or if you need one-time access to a data source you cannot access with your current license you can do so with RapidMiner Cloud. Executing a process with a RapidMiner Cloud subscription will allow you to access any data source supported by RapidMiner Studio!
- You can execute multiple processes at the same time
- You can run your processes on machines that have way more memory than your average desktop/laptop
- You don't need to worry about technical issues you might otherwise experience if you were to setup, use and maintain your own server
- You don't pay for any idle time as you would with your own Server. Taking setup, maintenance and configuration as well as hardware and license costs into account, your own server(s) can quickly add up to quite a high total cost of ownership.

We hope that this information sheds some light on using RapidMiner Cloud.

The RapidMiner Team

 on: October 24, 2014, 05:11:04 PM 
Started by peleitor - Last post by peleitor
Herllo. There is a quite popular retail dataset from belgian anonymized stores, which can be found here:

First file lines are:

0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29
30 31 32
33 34 35
36 37 38 39 40 41 42 43 44 45 46
38 39 47 48
38 39 48 49 50 51 52 53 54 55 56 57 58
32 41 59 60 61 62
3 39 48
63 64 65 66 67 68
32 69
48 70 71 72
39 73 74 75 76 77 78 79

But I don't know retail.dat file format.

How can it be imported into rapidminer?

I could not find any format descriptors of this.


 on: October 24, 2014, 06:18:11 AM 
Started by nachiket - Last post by nachiket
I am new Rapid Miner
I want to assign more weight to slope than other attributes in predicting the zone
I implemented Naive Bayes after referring to sample given for golf dataset on my data set now I want to assign weights
Sample Data(1100 rows)
FID   Geology                                           Geomorphology                                               Land use_land cover    Rainfall          SLOPE   Soil                        zone
0   Fissile hornblende biotite gneiss   HIGHLY DISSECTED DIFLECTION SLOPE        FOREST                    1200-1400   >60%   BROWN CLAY     High
1   Fissile hornblende biotite gneiss   HIGHLY DISSECTED DIFLECTION SLOPE        FOREST                    1200-1400   30-60%   BROWN CLAY    Moderate

 on: October 23, 2014, 05:10:17 PM 
Started by nachiket - Last post by nachiket
I am new rapidminer
I tried doing k-means clustering and bayes classification separately it works fine
but k-means takes numeric values and Bayes gives standard deviation for numeric values(I want prediction output)
Can anyone please help me in combining these two processes  

Sample Data(1100 rows)
FID   Geology                                           Geomorphology                                               Land use_land cover    Rainfall          SLOPE   Soil                        zone
0   Fissile hornblende biotite gneiss   HIGHLY DISSECTED DIFLECTION SLOPE        FOREST                    1200-1400   >60%   BROWN CLAY     High
1   Fissile hornblende biotite gneiss   HIGHLY DISSECTED DIFLECTION SLOPE        FOREST                    1200-1400   30-60%   BROWN CLAY    Moderate

 on: October 23, 2014, 10:20:12 AM 
Started by sgenzer - Last post by Marco Boeck

the same way as you delete a normal repository folder, either by selecting it and pressing "Del" or by right-clicking and selecting "Delete" in the context menu.
Are you getting some kind of error?


 on: October 22, 2014, 11:50:30 PM 
Started by mafern76 - Last post by mafern76
Hi, how are you?

I was recently presented with the following problem, a company has around 20.000 answers to one open question "What are the aspects of your work you like the most?", and they would like to analyze those answers.

I already worked manually analyzing around 300 of them, getting several flags, for example, HELPING_CUSTOMERS, SHORT_HOURS, etc.

My idea was to simply make a model for each flag and predict the remaining 20.000 answers, obtaining percentages regarding how many employees value each flag.

1. I was wondering if there is another approach to this and what would be the advantage over simply sampling the 20.000, getting percentages and extrapolating those, statistically, regardless of predictive models based on text.

2. Another valid question would be what is the difference between text mining and simply a tag cloud, but that is something that remains to be seen and I guess it depends on each individual problem. For example a more neutral question like "What do you think about your job?" may contain positive and negative sentiments using the same words, but right now I'm working on a question biased towards recieving positive sentiments.

Thanks a lot for your insight, I'll make sure to share mine!

 on: October 22, 2014, 04:15:12 PM 
Started by sgenzer - Last post by sgenzer
ok stupid question - how do you delete a cloud repository folder?

Pages: [1] 2 3 ... 10