We have an installation of Moodle with several courses and users.
Recently we have reported that at least one user cannot pass an exam. He retried several times and always obtains 0 score with message:
Results Your Score: 0% (0 points)
What is almost impossible because the quiz has 25 questions and many of them are True/False type
I tried to change the Auto Save Delay option to 30 seconds to save the questions while are responded but it does not work.
Any help will be appreciated
Related
I am very much puzzled about using H5P in Moodle.
The idea is great, obviously, yet I cannot make it work as I expected.
My principles/idea:
There are bunch of activities in each course
Each activity can build up several Student's skills, say Creative Thinking or Problem Solving
After finishing each activity Student, based on the result, can go
to the next activity or re-do it if failed
For the testing purposes I set up 3 outcomes (0-30 > NO pass, 30-70 >
1 point, 71-100 > 2 points) in the H5P module - this one is working
fine.
The outcome should be passed to Moodle, so the course can then decide
what to do: pass with 1 or 2 points, or fail and request to do the
activity again
This outcome then will be added to Student's skillset
Say, I have this basic crossword. After finishing it the Student can achieve two Skills mentioned above yet still this depends on the outcome, eg. result 1 means 0 in Creative Thinking and +1 in Problem Solving, and/or 2 means +1 in Creative Thinking and +2 in Problem Solving.
The activity itself works as expected, as I mentioned above, see the images (note ONE point circled):
and , but then nothing happens.
The student is NOT taken to the next activity, all s/he can do is to retry same activity over and over again.
Is it possible to force Moodle/H5P to act as described above?
For the testing purposes I used two 'activities': one being 'h5p' itself and the other being 'lesson' with same h5p modules being added inside, see the image:
I run this all on WAMP
I tried to follow xAPI https://h5p.org/documentation/x-api
which resulted in js error:
Sorry for the long post - tried to cover everything.
If anybody knows the answers - this reply will be much appreciated.
Cheers,
Greg
I use rampUsers(20)over (120) in my gatling load test. But I got the following result
I expected the active user should be constant during the test.
rampUsers injects the defined numbers of users linearly over a given time. So your use of rampUsers(20)over (120) will result in gatling starting one user every 6 seconds. The graph you're getting shows this, but what might be confusing is that since your scenario completes in less than 6 seconds there's never more than one user active at a time.
if you're aiming for 20 concurrent users over 120 seconds, there's a different injection profile for that...
constantConcurrentUsers(20) during (120 seconds)
I am working with ADs api and using keywordstats.
It worked well, now I am getting some weird responses, or maybe it should be like that.
I created AD 7 days ago, 24 days after using keywordstats and fields like, cpc, actions, impressions, unique_actions returned me good results, matched at what I was seeing on ADS manager. AD was live one day.
Now 6 days after after, actions and unique actions has only one element "attention_event"? Impressions by each interest are also weird, for each I am getting from 1 to 5? In total I have more then 5K impressions, how is this possible? I have tried with specific dates, for one day it returns this weird data, for other days just returns empty response.
Actions fields, 6 days ago contained and events from FB Pixel, now none is returned in response.
Is this supposed to work like this, or I am missing something?
I've also dealt with some bugginess with the keywordstats endpoint.
If you look at the API documentation, you'll see a particular line: "To query from a particular day in the last week:". It seems like Facebook doesn't allow you to query for keywordstats that aren't from the past week. The discrepancy between metrics might arise because of this.
For some more context, I've found that even the last seven days isn't a reliable indicator of retrievable dates. I had to query each date individually (using <API_VERSION>/<AD_ID>/keywordstats?date=YYYY-MM-DD) and see which ones don't return an empty object. The dates that do return a metric are the ones that you should use in your calculations.
I have made a Customer search that identifying the customers that haven't purchased an item in six months or more. I've used group summaries and maximum summaries for the company name and the maximum transaction date, which refers to the latest sales order containing a certain item. The idea is to send them an email. However, the workflow is only executing on 20 records at a time. i even conducted a search that was not summarized and the workflow still only executed on 20 records. I did use the "Execute Now" button in testing mode to see how many entered the work flow from the summary search. But each execution only yields 20 workflow instances. the searches yield about 213 and 300 records respectively. I appreciate any insight!
Workflows using searches will only process 20 records when in testing mode.
From SuiteAnswers Article 36738 (NetSuite Login required)
When you execute a workflow on demand, NetSuite only processes the first 20 records returned by the saved search. For example, if the saved search for a scheduled workflow returns 1000 records, the workflow only initiates on the first 20 records returned by the saved search.
I have ScriptA with some functions in files that have triggers that all run under UserA and consume about 2 hours of runtime per day.
I have another project ScriptB with some other functions in other files that have triggers that all run under UserA (the same user as ScriptB users) and consume about 3 hours of runtime per day.
Is my Trigger Aggregate Execution Time quota (from quota page here) aggregated per user or per script? That is, is it:
Five hours (2 + 3) for UserA or is it
Two hours for ScriptA and 3hrs for ScriptB?
I have seen this answer but it doesn't explicitly address the scoping question I'm asking.
Obviously is per user not ler script. Otherwise quotas wouldnt make sense.
In the interests of getting some evidence together for this:
At 4m25 in this March 2013 episode of Google Apps Unscripted, Kalyan Reddy says that the quotas are "per account type" and as you can see in the dashboard, the Quota table is gridded and has columns labelled with those account types too.
I have also done some testing and made a script that uses quite a bit of time. It started to max out other scripts running under the same account and many of that account's triggered scripts started to get errors "Service using too much computer time for one day". But... interestingly, after a couple of days of those errors have subsided. I believe on a consumer account I am now getting way more execution time than 1 hr per day.
While not a direct answer to the question and still a leap of logic/assumption, these two things make me feel that "per account" is more likely to be correct than "per script". I'll keep the question open for a bit longer for any comments (esp Googlers).