Well, I've made an agent based model in Netlogo.
I got a plot where I count the mean of some variables. Here's the result.
As you can see, there are no values higher than 5 points.
But, when i export data in a csv file, I get this:
It seems like there are moved commas...How is that possible? Is there any way that I could avoid this error?
Thanks for your time and help.
Related
I have several datasets (approx. 10) that comes from user input and the labels (x-axis) will therefore almost never overlap between the datasets.
What i would like to do is connecting points from the same datasets (for instance bloodPressure) thru a "non-existing" datapoint when this is necessary - like in the graph below. I would not want to fake a datapoint to achieve this.
Any suggestions how to do this?
I found the answer in the Chart.js documentation. It is possible to skip the array of labels and instead create the datasets with a x and y value like in the screen dump.
Works fine!
This is my first post, so apologies if I do something wrong or am not clear.
I am trying to display multiple .fig files in one figure to compare them to each other. I tried using montage() but the education license I am on apparently doesn't have that included.
With montage I did..
comp10 = dir('c:\data_figs\location_1\fig10*.fig');
montage(comp10,'size',[1 NaN]);
Is there a way to do something similar without montage? I'm comparing ultrasound imaging analyses, and want them next to each other in order to see which ones have significant movement and which ones can be thrown out.
I am very inexperienced in matlab, so I probably am missing something super simple.
Thank you!
I need help regarding the time series in Tableau. So far Here is what I can do.
Connect to TabPY
Call / Run scripts on TabPy
My current issue is that tableau doesn't seem to allow more output than input elements. Say I want to use the last 100 data points to predict the coming 10 points. Input of the data to python isn't a problem. The problem comes when I want to return a list with 110 elements. I've also tried returning the 10 elements and it complaints that it expects 100 elements list.
Thanks for reading
I've found a work around. You can see the post here for more information. Basically you shift the original values by the prediction amount and then have the prediction return the same amount as the shifted original
I've asked this before, but I feel I wasn't clear enough so I'll try again.
I am running a network simulation, and I have several hundreds output files. Each file holds the simulation's test result for different parameters.
There are 5 different parameters and 16 different tests for each simulation. I need a method to store all this information (and again, there's a lot of it) in Matlab with the purpose of plotting graphs using a script. suppose the script input is parameter_1 and test_2, so I get a graph where parameter_1 is the X axis and test_2 is the Y axis.
My problem is that I'm not quite familier to Matlab, and I need to be directed so it doesn't take me forever (I'm short on time).
How do I store this information in Matlab? I was thinking of two options:
Each output file is imported separately to a different variable (matrix)
All output files are merged to one output file and imprted together. In the resulted matrix each line is a different output file, and each column is a different test. Problem is, I don't know how to store the simulation parameters
Edit: maybe I can use a dataset?
So, I would appreciate any suggestion of how to store the information, and what functions might help me fetch the only the data I need.
If you're still looking to give matlab a try with this problem, you can iterate through all the files and import them one by one. You can create a list of the contents of a folder with the function
ls(name)
and you can import data like this:
A = importdata(filename)
if your data is in txt files, you should consider this Prev Q
A good strategy to avoid cluttering your workspace is to import them all into a single matrix. SO if you have a matrix called VAR, then VAR{1,1}.{1,1} could be where you put your test results and VAR{1,1}.{2,1} could be where you put your simulation parameters of the first file. I think that is simpler than making a data structure. Just make sure you uniformly place the information in the same indexes of the arrays. You could also organize your VAR row v col by parameter vs test.
This is more along the lines of your first suggestion
Each output file is imported separately to a different variable
(matrix)
Your second suggestion seems unnecessary since you can just iterate through your files.
You can use the command save to store your data.
It is very convenient, and can store as much data as your hard disk can bear.
The documentation is there:
http://www.mathworks.fr/help/techdoc/ref/save.html
Describe the format of text files. Because if it has a systematic format then you can use dlmread or similar commands in matlab and read the text file in a matrix. From there, you can plot easily. If you try to do it in excel, it will be much slower than reading from a text file. If speed is an issue for you, I suggest that you don't go for Excel.
There is a huge file mask.txt containing floating point numbers arranged in column format(single column of approx 2 million numbers) I want to extract the data in blocks of 512*512. How do I fetch the next block of data. I have done the following but it is erroneous.
rawData=dlmread('mask.txt');
a1=reshape(rawData(1:262144),512,512);
a2=reshape(rawData(262145:524289),512,512);
What to do? Please resolve the problem. Thanking you
You method is correct, it's simply your numbers that's wrong. You did the classical mistake of not counting the first number. The vector should be from [n:n+512^2-1], not [n:n+512^2] as you did. So to fix it, just do
a2=reshape(rawData(262145:524288),512,512);