1 Introduction
Modern university courses often use eassessment systems. Especially when courses have a high number of participants elearning tools are very useful to give students individual feedback. Courses with quantitative contents such as statistics and introductory mathematics are particularly suitable for eassessment since fillin exercises – which require students to submit a numeric answer – unambiguously allow to assess whether students can solve an exercise. The eassessment system JACK is a framework for delivering and grading complex exercises of various kinds. It was originally created to check programming exercises in Java [31], but has been extended to several other exercise types such as multiplechoice and fillin exercises [32, 30, 28]. JACK offers parameterizable content, meaning that exercises can contain different values each time an exercise is practiced. This means not only that different students get a different parameterization but, moreover, that the same student sees different numbers at each different time s/he selects the exercise. Hence, the exercise remains challenging until s/he understands the underlying concept to solve the exercise.
In addition to fillin exercises, JACK allows to design exercises with dynamic programming content. For instance, JACK offers electronic Java or R – the standard statistical programming language – exercises. Programming exercises not only help to prepare students for modern statistical work, but also have been shown to be highly beneficial to foster their understanding of statistics, see [24, 20].
This study analyzes JACK data to more deeply understand students’ learning behavior in an introductory mathematical statistics course. The high correlation between learning effort in the semester and the final grades is well documented, see [21] and Section 2 for more examples. Here, we aim to investigate additional aspects, namely the relevance of the daytime when students learn. In order to do so, we use several statistical learning methods to study which factors in students’ learning behavior are relevant to predict their success in the exam. It turns out that daytime activity has a higher effectiveness than nighttime activity.
Additionally, we find that good and very good students favor to learn in the afternoon, while some students who failed our course had insufficient learning behavior late at night. Moreover, we discuss the average time spent on exercises. Regarding this, students who participated at an exam spent more time in exercises than students who dropped the course before.
2 Related work
The overall engagement of students is indisputably one of the main covariates of academic success. For the case of mathematical statistics this has been shown on several occasions. [29] show in a metastudy that the simultaneous usage of traditional classroom lectures and eassessment has a positive effect on students’ success. [21] substantiate the previous result by analyzing the learning activity on the eassessment platform JACK. The study reveals that learning effort and success, measured by the total number of (correct) submissions on JACK over the course, positively affects the final grade in the exam. [24] add additional Rprogramming exercises to the JACK framework and show that the newly introduced exercise type helps to improve the general understanding of fundamental statistical concepts and thus ultimately yields better results in the final exam.
Due to the empirically observed positive effect of a multitude of variables on academic performance, prediction of the latter has become possible. In this socalled branch of educational data mining various statistical learning methods are applied to educational data in order to predict student outcomes. Often, but not necessarily, this outcome is measured with a binary response of pass/fail in order to be able to provide an earlywarning to students. [17, 14, 22] give a comprehensive overview of popular statistical learning methods used in the literature. For an overview of how to implement an earlywarningsystem see, e.g., [7]. The literature has identified a number of important predictors. [13] find evidence for the importance of socioeconomic and psychometric variables as well as preuniversity grades, although [23] show that, especially among the socioeconomic variables, the predictive capability can vary across countries. [4] additionally identify postadmission variables like obtained credits, degree of exam participation and exam success rate to have an influence on students’ success. [19, 33, 10] analyze the learning activity on learning management systems and are able to accurately predict students’ performance with appropriate variables. In a similar but more assessmentbased fashion [16, 9, 20] use activity in elearning frameworks as well as the results of midterm exams to predict students’ success in the final exam. [3] identify the performance in a small number of selected courses as a predictor for the academic achievement at the end of the study program. For a broad literature review concerning the usage of educational data mining see [27].
In contrast to previous studies which mostly rely on quantitative and qualitative learning activity measured by timeinvariant variables, there is also a temporal dimension of engagement which has been studied from different perspectives throughout the academic literature. [15] use timedependent information provided by a learning management system to predict academic performance. [34]
incorporate students’ response time as an additional feature into a random forest to investigate the predictive capability for students’ performance and find evidence that it can indeed improve prediction accuracy.
[26] and [25] further elaborate on the latter by using more sophisticated techniques and are able to support the preceding result.Only a few studies focus on the intraday engagement of students, that is, the actual daytime of learning, as a predictor for academic success. This topic is relevant as various studies show a significant influence of sleep quality and patterns on academic performance [2, 5, 6, 11]. Based on these insights [12] incorporate sleep variables into a prediction setting. With a stepwise regression approach they identify sleep frequency, night outings and sleep quality as among the most important predictors of academic success.
3 Course Structure
This section outlines the inital setup of the study. In particular, we sketch the structure of the analyzed course.
The eassessment system JACK was used for a lecture and exercises course in mathematical statistics at the German university of DuisburgEssen. 753 undergraduate firstyear students started the course. The course is compulsory for several business and economics programs as well as in teachers’ education. Out of these 753 students, only 379 took an exam at the end of the course, while the others dropped the course in this term (see Table 1). The course also introduces statistical programming skills using the statistics software R. In order to do so, the eassessment system JACK offers programming exercises where the correctness of students’ code is assessed, in addition to classical fillin and multiplechoice exercises.
Students, who…  counts  homework submissions 

took course  753  163,444 
registered to an exam  438  152,232 
participated at an exam  379  147,868 
passed an exam  186  87,382 
The course consisted of a weekly 2hours lecture, which introduced statistical concepts, and a 2hours exercise class, which presented explanatory exercises and problems. Both classes were held classically in front of the auditorium. Due to the large number of students, these classes are limited in addressing students’ different speeds of learning and individual questions. To overcome this issue and to encourage selfreliant learning, as well as to support students who had difficulties to attend classes, we offered all homework on JACK.
All in all, we offer 173 different exercises on JACK, of which 48 are designed as Rprogramming exercises and the remainder as fillin or multiplechoice exercises. The individual learning success is supported by offering specific automated feedback and, furthermore, by optional hints. In case of additional questions which were neither covered by hints nor feedback, the students were able to ask questions in our moodle helpforum.
In order to further encourage students to learn continuously during the semester, and not only in the weeks prior to the exams, we offered five online tests using JACK. These tests lasted 40 minutes at fixed times in the evening. Four of the online tests contained fillin or multiplechoice exercises only. The fifth online test contained R exercises exclusively. Participation only required a device with internet access, but no compulsory attendance at university. This summative assessment allowed students to assess their individual state of knowledge during the lecture period. It was not compulsory for students to participate at online tests in order to take the final exam at the end of the course. Instead, we offered bonus points for the final exams to encourage participating at the tests (a maximum of 10 bonus points in total for fillin online tests). The bonus points were only added to final exam points if at least 25 out of 60 exam points were reached, i.e., if students passed the exam without bonus. The R online test was worth at most 2 bonus points which were awarded even if students achieved less than 25 points. The reason for this was to motivate students to focus on programming skills since [24] and [20] show that this has a substantially (three times) higher impact on exam success than classical fillin exercises.
The final exams (3 in total) were also held electronically. While online tests during the semester could be solved at home with open books, the final exams were offered exclusively at university PC pools and supervised by academic staff. The exam consisted of R exercises (), short handwritten proofs () and the remainder of fillin exercises. Students can only retake an exam if they failed or did not take the previous ones (so that students can pass at most once), but can fail several times.^{1}^{1}1Students obtain 6 “malus points” for each failed exam of which they may collect at most 180 during their whole bachelor program. The last grade a student achieved in an exam will be denoted as the final grade. The corresponding exam will be denoted as the final exam.
3.1 Data and Models
In this section we present the available database and the models used. For each homework submission by a student on JACK we observe the exercise ID, the student ID, the number of points (on a scale from 0 to 100) and the time stamp with minutelong precision.
The response variables are given by the final exam success. We consider two possible responses. The first is a binary variable indicating whether a student passed (1) or did not pass (0) the course. Second, we consider the final grade as a response. We have the following grading scheme: very good (“100”), good (“200”), satisfactory (“300”), sufficient to pass (“400”) and failed (“500”). We assign “600” to students who took the course but did not participate at any of the exams. This is actually not a grade. However, this reflects the view that students who did not take any exam were even less prepared than students who failed the exams. Table
2 reports an overview of final grades. We do not report grades for one specific exam date but the grade given at the end of the course.Grade (points)  counts 

“100” (48 – 60)  9 
“200” (40 – 47)  38 
“300” (31 – 39)  100 
“400” (25 – 30)  39 
“500” (0 – 24)  193 
376  
failure rate  .508 
“average grade”  396.8 
JACK registered 163,444 submissions of homework exercises. See Table 1 for how these submissions are distributed among students. Figure 1 plots the number of daily submissions on JACK aggregated for all students. Characteristically, the number of submissions peaks shortly before a summative assessment such as an exam or an online test. This was also observed by [21].
We compile the following information for each student from the raw data:

the number of submissions (# submissions in short),

the number of fully correct submissions (100 points),

the number of submissions in the morning from 8am to 12pm,

the number of submissions in the afternoon from 12pm to 4pm,

the number of submissions in the evening from 4pm to 8pm,

the number of submissions in the late evening from 8pm to 12am,

the number of submissions at night from 12am to 8am,

the median submission time (see Subsection 4.2).

The score, which is defined as follows: let be a day during the semester. Then
where is the number of points of the latest submission up to time of student in exercise , . In other words, the score is the sum of points of the last submissions to every exercise. This helps tracking the learning progress for every student. In particular, we consider the final score, which is the score evaluated at the end of the term.

The frequency of submissions, i.e., the mean time between two following submissions at different days, measured in days.

The time until a student hands in the first submission from the beginning of the term, measured in days.

The time until a student hands in the last submission before his/her last exam, measured in days.

The number of days a student submitted solutions.

The average time spent per exercise measured in minutes (see Subsection 4.3).
Table 3 reports summary statistics. Figure 2 plots the average score of students with different grades and of students who dropped the course. Evidently, good and very good students had a strong learning progress from the beginning of the semester on. Students with the sufficient pass grade “400” and students who failed (“500”) start similarly weak but improve shortly before the exam. Students with “400”, however, improve slightly more, which may be the reason that they pass the exam. On the other hand, they may just have been lucky in the exam. The students who dropped the course show very little progress on average.
Variable name  Min  1st Quartile 
Median  Mean  3rd Quartile  Max  Sd 

# submissions  0  16  85  219.2  329  2726  296.8 
# correct  0  7  32  94.7  126  1576  144.3 
# morning  0  0  9  32.97  38  653  62.2 
# afternoon  0  3  28  73.75  103  1273  111.7 
# evening  0  5  32  71.56  105  819  97.9 
# late ev  0  0  5  31.52  36  465  58.3 
# night  0  0  0  9.432  5  396  32 
Median submission time  12:17am  2:16pm  3:56pm  3:37pm  7:19pm  11:31pm  2:52 
Final score  0  500  2372  4105  7333  15681  4100 
Frequency  1  5.25  9.556  50.433  26  234  84.8 
First submission  0  7  25.3  36.69  29.3  259.3  48.7 
Last submission  0  0.5  17.9  75.71  173.5  234  88.1 
# days  0  2  9  15.47  25  90  16.6 
Ave Time spent  0  5.8  10.3  10.5  13.9  46.3  6.8 
Overview of empirical quartiles mean and standard deviation for the considered covariates.
We choose the following modeling approaches for the classification problem:^{2}^{2}2
We also tried other modeling approaches than the ones stated here (e.g. neural networks, support vector machines, etc.). However their predictive performance proved not to be competitive.

Logistic regression models the probability of an event given
regressor variables via(1) The idea is to regress the logodds,
, on a linear combination of . So equation (1) can be rewritten as(2) The unknown coefficients
are estimated based on the available data. We use maximum likelihood, see
[17] for more details. We measure the variability of the estimatesvia the standard errors
, of the estimates. From this we obtain the statistics . [18] recommends to use the absolute value of the statistic of each nonconstant regressor as importance measure for logistic regression.The above approach can easily be extended to ordered logistic regression in which we want to predict a variable with
possible outcomes (multiclassclassification), see
[1]. We use binary logistic regression to predict the response “student passed”, and ordered logistic regression to predict the grade. 
Random Forests.
Treebased methods can be used for twoclassclassification as well as multiclassclassification. Single decision trees are very easy to interpret but have the drawback of having a high variance. To avoid this problem
[8] proposed an algorithm for averaging decision trees to obtain a socalled random forest. The idea is to take bootstrap samples from the single training data set. Then, a tree is trained on every bootstrapped training data sample. Finally, the prediction is the majority vote, which is the most common occurring class over all predictions, see [17]. Each of the single trees has a high variance but a low bias. Averaging over all trees reduces the variance. Another problem is that in each split of the trees, every variable in the predictor space is considered. If there is, for example, one very strong predictor it will be used in each tree for the first split. This leads to a high correlation between the trees. To avoid this problem, only a random sample of the predictors is used in each split to find an optimal split. The number of predictors in this random sample is usually set to . [8] also proposed to use the Mean Decrease Accuracy as importance measure for the input variables. We build trees to grow the forest and try variables at each split.
4 Empirical results
This section analyzes students’ learning behavior. We discuss which learning strategy turns out to predict students’ exam success.
4.1 Variable Importance
Our first analysis discusses which of the explanatory variables have a high predictive relevance. To model the target variable, i.e., passing the course or achieving a certain grade, we use the following set of variables as predictors in all of our models: {# correct, # morning, # afternoon, # evening, # late evening, # night, score at day of first online test, score at day of third online test, final score, frequency, first submission, last submission, # days, total time spent for exercises}. We dropped some variables like the total number of submissions to avoid high correlation between the predictor variables.
To compare the performance of the two models we use the accuracy, i.e., the rate of correctly classified observations. To avoid overfitting we use fold cross validation. Table 4 contains the cross validation results for the two different models. We see that in both, twoclass and multiclass classification, the random forest works best with an accuracy of or but logistic regression works well, too. In the full data set of the students do not pass the course, which leads to an accuracy of if we predict all students to not pass the course. Hence the random forest leads to an increase in accuracy of around percentage points which leads us to use the results of the random forest from now on.
Model  Accuracy twoclass  Accuracy multiclass 

Logistic Regression  0.821  0.72 
Random Forest  0.830  0.73 
We now investigate the variables which are chosen to build the single trees for the random forest. Figure 3 shows the importance of the variables used in the analysis.
We can see that the variable last_submission, i.e., the time until a student hands in the last submission before his/her last exam, measured in days, is by far the most important variable. Unfortunately, Figure 3 is silent on the direction of impact on the target variable. A solution to this problem is the partial dependence plot, which can help to understand how the logodds of realizing the respective class depend on the input variables.^{3}^{3}3The yaxis shows , where is the number of classes, is the class of interest, and is the proportion of votes for class . A high positive value of the partial dependence for a given value of the predictor means that it is more likely to belong to the class of interest than to the other class, see [14]. Here the class of interest is not passing the course. Figure 4 shows the partial dependence plot for last_submission. We see that not passing the course is more likely for high values of last_submission. This means that students who learn until the day of the final exam unsurprisingly have a higher probability to pass the course than students who quit learning far before the exam. This is because 374 out of 753 students did not participate in an exam. Most of these students did not learn until the exam but only made a few submissions at the beginning of the semester. Hence the variable last_submission has high values for these students. On the other hand most of the students who participated in the exams learned until shortly before the exam. This implies the high importance of the last submission. Other important variables are the final score, the frequency of submissions and the number of submissions in the morning. Figure 5 shows the partial dependence plot of the final score. We see that a high final score leads to a low probability not to pass the course.^{4}^{4}4Note that in logistic regression the sign of the estimated coefficients tells the direction of the impact of a variable. These are mostly in line with the exemplary partial dependence plots.
Furthermore, the importance of the variables in Figure 3 shows that the time of the first submission, the number of submissions at night and in the late evening and the first score in the term do not help for the predictive performance in the final exam. For the former and the latter this could be due to the fact that, at the beginning of the course, almost all students start to learn at the same level of knowledge, so there is no information that helps to decide between students passing or failing the final exam.
Since 374 out of 753 students did not participate in the exams we only focus on students who participated in an exam for the remainder of this subsection. This will obviously reduce the impact of the variable last_submission. We now estimate the corresponding binary classification random forest for pass vs. fail. Figure 6 shows its variable importance plot.^{5}^{5}5Note that a negative value for the mean decrease accuracy implies that randomly permuting the respective variable (ceteris paribus) yields to a lower MSE of the random forest. Now, final score and frequency, i.e., the mean time between two days of submissions measured in days, are the most important variables in the random forest model. For example, Figure 7 shows the partial dependence plot for the frequency. Small values of frequency make it more likely to pass the course. This means that students who learn regularly with only a few hours between their submissions have a higher probability to pass.
In case of multiclass classification, again, the time until a student hands in the last submission before his/her last exam is by far the most important variable in the model, for the same reasons as above. All other variables have low importance in this model. For reasons of brevity we shall now focus on the results of the binary model.
4.2 Learning Times
We now analyze more deeply at which time of the day good and less successful students prefer to learn. In order to investigate this we compute the median submission time for each student. We compare the median submission times for students who passed or did not pass in final exams.
Figure 8 shows kernel density plots for the median submission time for passing students in solid black and nonpassing students in dashed red. There is a higher variance of median submission times for students who did not pass; students who passed prefer to learn in the afternoon. Weaker students tend to learn later. Moreover, quite a few nonpassing students have median learning time in the morning. This is usually the time of the day when students should attend lecture and exercise classes.
Figure 9 further supports this claim. We compare very good and good students with students who failed all exams and students who dropped the course. Evidently, good students prefer to submit exercises during daytime. The earliest median submission time of a good student is about 11:30am and the latest is 7:20pm. Comparing these students with students who failed and, more visibly, who dropped shows that there are quite a few who study very late or very early. For example, the earliest is about 12:20am and the latest about 11:30pm.
This leads us to conclude that there are more nonpassing students who have difficulties to learn in the afternoon. As stated in Section 2 lack of sleep caused by studying at night has a negative impact on students’ performance.
Needless to say, the more important reason that poor students fail is mainly because they learn too little and not because of bad timing, cf. Subsection 4.1. It also needs to be emphasized that unfavorable time management can also be due to a high amount of responsibilities not connected to their studies. Unfortunately, our data set does not allow us to distinguish between these aspects. A data set including both submission data for eassessment and information on students’ other daily activities is hard to collect.
4.3 Submission duration
We now highlight another influential factor for success: we consider how long students work on a single submission, i.e., how much time they spent to solve an exercise. This analysis faces some challenges. First, we only observe the end and not the beginning of solving an exercise and hence do not have exact start and end times. We bypass this problem by measuring the time between two succeeding submissions. For example, if a student submits an exercise at 12pm and submits a second exercise (which might be the same as the first) at 12:15pm we consider 15 minutes as time spent for the second exercise. This means we do not observe duration of the first submission but of the following submissions. We omit duration times which are longer than two hours because students then likely took a break. This is also part of the second issue because we only monitor submission times in JACK and not whether students used this time to learn or whether they got distracted. We cannot rule out times of distractedness but still believe the following analysis offers interesting insights.
For each student we accumulate all duration times. These totals are of course higher for students who submitted more exercises than for students who submitted only a few. We thus divide total duration by the number of exercises submitted for each student.
Figure 10 shows a kernel density estimate for the average time spent per submission of students who passed versus students who did not pass. Evidently, there are many students who did not pass who invested little time for each exercise. Again, we next distinguish between students who failed an exam and students who dropped the course. Figure 11 compares students who achieved the best or second best grade with students who failed and students who dropped the course. Interestingly, time spent per submission is similar for both good students and students who failed (The plot for mediocre students looks very similar, too.). However, students who dropped the course invested perceptibly less time for each submission. Apparently, these students had too little motivation and/or time to participate in the course. They likely did not seriously attempt to solve the exercises.
5 Conclusions
This study analyzes when students should learn to be successful in a final exam. For this purpose, we analyzed data from the online learning platform JACK from an introductory mathematical statistics course in the summer term 2017. This data on students’ submissions on JACK offered information about the daytime when a student submits a solution to an exercise.
We used logistic regression and random forests to predict the success of a student in the final exam and, also try to predict the final grade. An advantage of these methods is that they offer information about the importance of the variables used in the model. We analyze the variable importance obtained by the random forest.
The two most important variables in this model are the day between the last submission of an exercise and the exam as well as the score the students achieve when they study with JACK. We further identify the frequency with which the students work on JACK and the number of submissions between 8am and 12pm as important variables. We identify good students to submit exercises during the daytime, while some students who quit the course or fail in the final exam learn very early in the morning or very late in the evening. Needless to say, the total amount of learning has a high impact on success. Additionally, we cannot rule out external factors (e.g. working during daytime) causing this effect rather than students who purposely did not study during daytime. Still we may conclude that students who did not pass the course study little during the afternoon. Moreover the time a student spends on a single exercise is very short for students who dropped the course.
All in all, our results stress the importance for students to decide when and how often to learn. With a good time management, students can possibly increase their probability to pass a course like the one investigated here.
6 Acknowledgments
We thank all colleagues who contributed to the course “Induktive
Statistik” in the summer term 2017, especially Paul Navas Alban.
We thank Anna Janssen, Kim J. Hermann, Timo Rammert and Alexander Zyweck for excellent research assistance.
Part of the work on this project was funded by the German Federal
Ministry of Education and Research under grant numbers 01PL11075 and 01 JA 1610.
References
 [1] Agresti, A. Categorical Data Analysis. John Wiley & Sons Inc, Hoboken, New Jersey, 2002.
 [2] Ahrberg, K., Dresler, M., Niedermaier, S., Steiger, A., and Genzel, L. The interaction between sleep quality and academic performance. Journal of Psychiatric Research, 46(12):1618–1622, 2012.
 [3] Asif, R., Merceron, A., Ali, S.A., and Haider, N.G. Analyzing undergraduate students’ performance using educational data mining. Computers & Education, 113:177–194, 2017.
 [4] Baars, G.J.A., Stijnen, T., and Splinter, T.A.W. A model to predict student failure in the first year of the undergraduate medical curriculum. Health Professions Education, 3(1):5–14, 2017.
 [5] Baert, S., Omey, E., Verhaest, D., and Vermeir, A. Mister sandman, bring me good marks! On the relationship between sleep quality and academic achievement. Social Science & Medicine, 130:91–98, 2015.
 [6] Beşoluk, Ş, Önder, İ, and Deveci, İ. Morningnesseveningness preferences and academic achievement of university students. Chronobiology International, 28(2):118–125, 2011.
 [7] bin Mat, U., Buniyamin, N., Arsad, P.M., and Kassim, R. An overview of using academic analytics to predict and improve students’ achievement: A proposed proactive intelligent intervention. In 5th Conference on Engineering Education (ICEED), pages 126–130. IEEE, 2013.
 [8] Breiman, L. Random forests. Machine Learning, 45(1):5–32, 2001.
 [9] Burgos, C., C. L, de la Peña, D., Lara, J.A., and Lizcano, D. and Martínez, M.A. Data mining for modeling students’ performance: A tutoring action plan to prevent academic dropout. Computers & Electrical Engineering, 66:541–556, 2018.
 [10] Elbadrawy, A., Studham, R.S., and Karypis, G. Collaborative MultiRegression Models for Predicting Students’ Performance in Course Activities. In Proceedings of the Fifth International Conference on Learning Analytics And Knowledge, LAK ’15, pages 103–107. ACM, 2015.
 [11] Eliasson, A., Eliasson, A., King, J., Gould, B., and Eliasson, A. Association of Sleep and Academic Performance. Sleep and Breathing, 6(1):45–48, 2002.
 [12] Gomes, A.A., Tavares, J., and de Azevedo, M.H.P. Sleep and Academic Performance in Undergraduates: A MultiMeasure, MultiPredictor Approach. Chronobiology International, 28(9):786–801, 2011.
 [13] Gray, G., McGuinness, C., and Owende, P. An application of classification models to predict learner progression in tertiary education. In International Advance Computing Conference (IACC), pages 549–554. IEEE, 2014.
 [14] Hastie, T., Tibshirani, R., and Friedman, J. The Elements of Statistical Learning. Springer, New York, 2009.
 [15] Hu, Y.H., Lo, C.L., and Shih, S.P. Developing early warning systems to predict students’ online learning performance. Computers in Human Behavior, 36:469–478, 2014.
 [16] Huang, S. and Fang, N. Predicting student academic performance in an engineering dynamics course: A comparison of four types of predictive mathematical models. Computers & Education, 61:133–145, 2013.
 [17] James, G., Witten, D., Hastie, T., and Tibshirani, R. An Introduction to Statistical Learning: with Applications in R. Springer, New York, 2013.
 [18] Kuhn, M. Variable selection using the caret package, 2012.
 [19] Macfadyen, L.P. and Dawson, S. Mining LMS data to develop an ”early warning system” for educators: A proof of concept. Computers & Education, 54(2):588–599, 2010.
 [20] Massing, T., Reckmann, N., Otto, B., Hermann, K.J., Hanck, C., and Goedicke, M. Klausurprognose mit Hilfe von EAssessmentNutzerdaten. In DeLFI 2018  Die 16. ELearning Fachtagung Informatik, pages 171–176, 2018.
 [21] Massing, T., Schwinning, N., Striewe, M., Hanck, C., and Goedicke, M. Eassessment using variablecontent exercises in mathematical statistics. Journal of Statistics Education, 26(3):174–189, 2018.
 [22] Meier, Y., Xu, J., Atan, O., and van der Schaar, M. Predicting grades. IEEE Transactions on Signal Processing, 64(4):959–972, 2016.
 [23] Oskouei, R.J. and Askari, M. Predicting Academic Performance with Applying Data Mining Techniques (Generalizing the results of Two Different Case Studies). Computer Engineering and Applications Journal, 3(2):79–88, 2014.
 [24] Otto, B., Massing, T., Schwinning, N., Reckmann, N., Blasberg, A., Schumann, S., Hanck, C., and Goedicke, M. Evaluation einer Statistiklehrveranstaltung mit dem JACK RModul. In DeLFI 2017  Die 15. eLearning Fachtagung Informatik, Lecture Notes in Informatic, Gesellschaft für Informatik, pages 75–86, 2017.
 [25] Papamitsiou, Z., Karapistoli, E., and Economides, A.A. Applying classification techniques on temporal trace data for shaping student behavior models. In Proceedings of the Sixth International Conference on Learning Analytics & Knowledge, LAK ’16, pages 299–303, New York, NY, USA, 2016. ACM.
 [26] Papamitsiou, Z.K., Terzis, V., and Economides, A.A. Temporal learning analytics for computer based testing. In Proceedings of the Fourth International Conference on Learning Analytics And Knowledge, LAK ’14, pages 31–35, New York, NY, USA, 2014. ACM.
 [27] Romero, C. and Ventura, S. Educational Data Mining: A Review of the State of the Art. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 40(6):601–618, 2010.
 [28] Schwinning, N., Striewe, M., Massing, T., Hanck, C., and Goedicke, M. Towards digitalisation of summative and formative assessments in academic teaching of statistics. In Proceedings of the Fifth International Conference on Learning and Teaching in Computing and Engineering, 2017.
 [29] Sosa, G.W., Berger, D.E., Saw, A.T., and Mary, J.C. Effectiveness of computerassisted instruction in statistics: A metaanalysis. Review of Educational Research, 81(1):97–128, 2011.
 [30] Striewe, M. An architecture for modular grading and feedback generation for complex exercises. Science of Computer Programming, 129:35–47, 2016.
 [31] Striewe, M., Balz, M., and Goedicke, M. A Flexible and Modular Software Architecture for Computer Aided Assessments and Automated Marking. In Proceedings of the First International Conference on Computer Supported Education (CSEDU), 23  26 March 2009, Lisboa, Portugal, volume 2, pages 54–61. INSTICC, 2009.
 [32] Striewe, M., Zurmaar, B., and Goedicke, M. Evolution of the EAssessment Framework JACK. In Gemeinsamer Tagungsband der Workshops der Tagung Software Engineering 2015., pages 118–120, 2015.
 [33] Wolff, A., Zdrahal, Z., Nikolov, A., and Pantucek, M. Improving retention: Predicting atrisk students by analysing clicking behaviour in a virtual learning environment. In Proceedings of the Third International Conference on Learning Analytics and Knowledge, LAK ’13, pages 145–149, New York, NY, USA, 2013. ACM.
 [34] Xiong, X., Pardos, Z.A., and Heffernan, N.T. An Analysis of Response Time Data for Improving Student Performance Prediction, 2011.
Comments
There are no comments yet.