1 Introduction
Curriculum learning is an approach to training neural networks inspired by human learning. Humans learn skills such as maths and languages by tackling increasingly difficult tasks. It has long been proposed
[4] that neural networks could benefit from a similar approach to learning. Curriculum learning involves presenting examples in some order during training such as to aid learning. On some tasks, curriculum learning has been shown to be necessary for learning [9, 23, 15] and to improve learning speed [5, 22] and generalization ability [2] on other tasks.For any particular application of curriculum learning the two key components of this approach to training neural networks are the division of the training data into tasks of varying difficulty and the definition of a syllabus. We follow [5] in defining a syllabus to be a “timevarying sequence of distributions over tasks”. Where a task is defined to be a subset of all training examples e.g. all examples with an input sequence of length 10. At any instance during training a batch called an example is drawn from some task and presented to the network.
In this work we focus on the effect of syllabus design on curriculum learning. Syllabuses fall in two main categories. Handcrafted syllabuses [22, 23, 9, 2] define an ordering of tasks and a level of success on each task to be attained before moving onto the next point in the syllabus. The level of success on a task is typically evaluated periodically on a validation set.
Automated syllabuses [5, 16] attempt to solve the problem of having to handdesign a syllabus and choosing the level success that determines progression from task to task required when using a handcrafted syllabus. Automated syllabuses only require the engineer to define a set of tasks (potentially ordered) and examples are chosen by some automatic mechanism e.g. in proportion with the error rate on that task.
The choice of syllabus has a significant effect on the efficacy of curriculum learning for a particular problem [5, 22]. Syllabuses have been primarily developed on an adhoc basis for individual problems and only limited empirical comparison of different syllabuses has been conducted [5, 22]. Additionally explanations for why curriculum learning works have mostly been grounded in learning theory [4, 13, 2] rather than empirical results.
Following [5], we consider curriculum learning in two settings. In the multitask setting, after training we are concerned with the performance of the network on all tasks. Whereas in the targettask setting, after training we are concerned only with the performance of the network on a single “target” task.
In this work we evaluate how the choice of syllabus effects the learning speed and generalization ability of neural networks in the multitask and targettask settings. We train LSTM [11] networks on three synthetic sequence learning tasks [8]. Sequence learning tasks are naturally suited to curriculum learning as it is often assumed that the difficulty of the task increases with the length of the input sequence [22, 15, 3, 18, 9]. LSTMs have achieved state of the art performance in many sequence learning tasks [6, 21, 7]. It is common to use LSTMs trained on synthetic tasks for experimental results in curriculum learning [22, 5, 3, 18, 13].
Our results provide the most extensive empirical comparison of the effect of curriculum learning syllabuses on speed of learning and generalization ability that we are aware of. Additionally, our results reveal insights on why curriculum learning outperforms standard training approaches and the key components of successful syllabuses. Thus our work provides evidence for the choice of syllabus in new applications of curriculum learning and provides the basis for future work in syllabus design.
2 Related Work
Bengio, et al. [2]
examined curriculum learning under handcrafted syllabuses. They found that a Perceptron trained on a toy task with a syllabus which steps through tasks in their order of difficulty learns faster than a Perceptron trained on randomly sampled examples. Their work demonstrated the potential utility of curriculum learning, but the division between curriculum learning and transfer learning was unclear in their experiments and only a basic handcrafted syllabus was considered.
Zaremba and Sutskever [22] showed that curriculum learning was necessary for a LSTM network to learn to “execute” simple Python programs. The authors provide an empirical comparison of two handcrafted syllabuses. Their problem had two dimensions of difficulty the length of the program and the level of nesting. They defined a syllabus which they call the Combined Strategy that combines the basic approach of stepping through the tasks one at a time with a uniform distribution over all tasks. The Combined Strategy syllabus lead to faster learning and better final performance on the target task than baseline strategies.
The authors hypothesize that by including examples from tasks more difficult than the current task, the LSTM network has an incentive to use only a portion of its memory to solve the current task in the syllabus. The authors argue that syllabuses which do not provide tasks from future points on the syllabus incur a significant time penalty at each point of progression as the network must undergo substantial retraining of the network weights to adapt to the new task. In a separate work, the same authors proposed a slightly more complex syllabus with the same goal of sampling tasks based on a syllabus with “nonnegligible mass over the hardest difficulty levels” which enabled reinforcement learning of a discrete Neural Turing Machine
[20], where a network trained directly on the target task failed to learn.Zaremba and Sutskever only considered syllabuses with a linear progression through the tasks in order of difficulty. Yet it has been argued that linear progression through the tasks in a syllabus can result in quadratic training time in the length of the sequences for a given problem [15]. Following a simple syllabus of drawing tasks from a uniform distribution over all tasks up to the current task and exponentially increasing the level of difficulty of the problem at each progression point it is possible to scale a variant of the Neural Turing Machine [8] to sequences of length 4,000 [15].
Graves, et al. [5] propose an automated syllabus by treating the selection of the task from which to draw the next example from as a stochastic adversarial multiarmed bandit problem [1]. Their paper examines six definitions of a reward function to promote learning. The reward functions split into two classes, those driven by the improvement in the loss of the network as a result of being trained on an example from a particular task and those driven by an increase in network complexity motivated by the Minimum Description Length principle [17, 10]. The authors show that the best reward function from each class results in nonuniform syllabuses that on some tasks converge to an optimal solution twice as fast as a baseline uniform sampling syllabus. Across all tasks however the authors find that the uniform sampling syllabus provides a strong baseline. Other more simplistic automated syllabuses have been used successfully, for example sampling from tasks in proportion with the average error on that task [16].
3 Methodology
3.1 Syllabuses
We considered six syllabuses that captured the main features of those identified in the literature. The six syllabuses include three handcrafted, one automated and two benchmark syllabuses. Below, we follow the notation that there are T tasks on the syllabus that are ordered in difficulty from 1 to T and that the learner’s current task is denoted C. A full distribution over tasks for each syllabus is given in table 1.
Syllabus  C  Uniform(1, max{1, C1})  Uniform(1, T)  T 

Naive  100%       
Look Back  90%  10%     
Look Back and Forward  80%    20%   
None        100% 
Uniform      100%   
The syllabus termed the Naive Strategy by Zaremba and Sutskever [22] was one of the first proposed syllabuses [2] for curriculum learning. Under the Naive Strategy examples at each timestep are sampled solely from the current point C on the syllabus. Once the learner reaches the defined level of success for progression on the current task as measured on a validation set, the learner moves onto the next task on the syllabus and all examples are now drawn from that task. We call this syllabus Naive.
It was observed that while the Naive syllabus increased the rate of learning on certain tasks, the learner rapidly forgot previously learned tasks [16, 14]. This is undesirable in the multitask setting and it was proposed [9] that drawing some examples from previous tasks on the syllabus may prevent this “catastrophic forgetting”. We define a syllabus which we call Look Back in which a fixed percentage of examples are drawn from a Uniform distribution over all previous tasks on the syllabus and the remainder are drawn from the current task on the syllabus as per the Naive syllabus. In practice for our experiments we chose to draw 10% of tasks from previous points on the syllabus, with the remaining 90% coming from the current task.
While the Look Back syllabus addresses the issue of catastrophic forgetting, it was further hypothesized that by only drawing examples from the current and past tasks in the syllabus considerable retraining of the network weights would be required as the learner moved forward through the syllabus [22]. A syllabus which we call Look Back and Forward addresses this issue. Look Back and Forward corresponds to the syllabus which Zaremba and Sutskever call the Combined Strategy [22]. With the Look Back and Forward syllabus a fixed percentage of examples are drawn from a Uniform distribution over all tasks on the syllabus with the remaining examples drawn from the current task. Thus, when on the early tasks in the syllabus almost all examples drawn from the uniform distribution will be drawn from future tasks on the syllabus. Once the learner approaches the target task almost all such examples will be drawn from previously learned tasks. In this way the Look Back and Forward syllabus seeks to address both the issue of catastrophic forgetting and the potential retraining required when the learner is not given examples from upcoming tasks on the syllabus. In our experiments, we drew 20% of examples from the Uniform distribution over all tasks, with the remaining 80% coming from the current task on the syllabus. We note that for the Look Back and Look Back and Forward syllabuses, we chose the percentage splits through experimentation.
We adopt the best performing automated syllabus consistent with maximum likelihood training from recent work on automated curriculum learning [5], which the authors call Predictive Gain. The authors follow their general approach of selecting the next task based on the Exp3.S algorithm [1] for stochastic adversarial multiarmed bandits. The reward function for Predictive Gain is defined to be the scaled reduction in loss on the same example after training on that example, i.e. the reward to the bandit is defined to be rescaled to the range by a simple procedure [5], where and are the weights parameterizing the network before and after training on respectively.
None is our benchmark syllabus in the targettask setting, for which we draw all examples from the target task in the syllabus.
As our benchmark syllabus in the multitask setting we draw examples from a Uniform distribution over all tasks. Unsurprisingly, we call this syllabus Uniform. This syllabus can also be seen as a simple syllabus in the targettask setting and as mentioned above has been found to be a strong benchmark in this setting [5].
In practice for the above handcrafted syllabuses we, as other authors have found [15], that learning was slow if the learner progressed through the syllabus one task at a time. To alleviate this, for each of these syllabuses after meeting the defined success metric on a task instead of moving onto the next task in the syllabus we followed an exponential strategy of doubling the current point on the syllabus along one dimension of difficulty on the problem. For the Repeat Copy problem (described below) which has two dimensions of difficulty, we alternated which dimension of difficulty to double.
3.2 Problems
As noted above curriculum learning has primarily been applied to sequence learning problems where the sequence length is typically used to divide the data into tasks ordered by difficulty. We follow this approach by adopting three synthetic sequence learning problems that have been shown to be difficult for LSTMs to solve [8].
Copy
 for the Copy problem, the network is fed a sequence of random bit vectors followed by an end of sequence marker. The network must then output the input sequence. This requires the network to store the input sequence and then read it back from memory. In our experiments we trained on input sequences up to 32 in length with 8dimensional random bit vectors.
Repeat Copy  similarly to the Copy problem, with Repeat Copy the network is fed an input sequence of random bit vectors. Unlike the Copy problem, this is followed by a scalar that indicates how many times the network should repeat the input sequence in its output sequence. In our experiments we trained on input sequences up to 13 in length with maximum number of repeats set to 13, again with 8dimensional random bit vectors. This means the target task for Repeat Copy required an output sequence of length 169.
Associative Recall  Associative Recall is also a sequence learning problem, with sequences consisting of random bit vectors. In this case the inputs are divided into items, with each item consisting of 3 x 6dimensional vectors. After being fed a sequence of items and an end of sequence marker, the network is then fed a query item which is an item from the input sequence. The correct output is the next item in the input sequence after the query item. We trained on sequences of up to 12 items, thus our target task contained input sequences of 36 6dimensional random bit vectors followed by the query item.
3.3 Experiments
For all of the above problems we ran experiments to measure the effect of syllabus choice on speed of learning in both the targettask and multitask setting and the generalization ability of the trained network. For each problem, syllabus pair we ran training from 10 different random initializations.
In order to measure the speed of learning we measured the performance during training every 200 steps on two heldout validation sets, one for the targettask and one for the multitask setting. The validation set for the targettask setting consisted solely of examples from the target task on that problem. Whereas, the validation set for the multitask setting consisted of examples uniformly distributed over all tasks in the syllabus. The number of examples in the targettask and multitask validation sets were 512 and 1024 respectively for all experiments.
To test the effect of syllabus choice on the generalization ability of our networks, for each problem we created test sets of 384 examples which gradually increased the difficulty of the problem. For the Copy problem which was trained with a target task of sequences of length 32, the test set comprised on sequences of length 40.
For Repeat Copy we wished to test the generalization ability of the trained networks along both dimensions of difficulty of the problem  sequence length and number of repeats. We created two test sets, one comprised of sequences of length 16 with the number of repeats fixed to 13. The other test set comprised of sequences of length 13 with the number of repeats set to 16. Our test set for Associative Recall consisted of sequences of 16 items, having been trained with a target task of 12 items. We measured the performance of each network on the test set when the network’s error on the validation set was lowest.
In all our experiments we used a stacked 3 x 256 units LSTM network, with a crossentropy loss function. For all networks and syllabuses we used the Adam optimizer
[12] with an initial learning rate of 0.01. We used a batch size of 32 for all experiments, with each batch containing only examples from a single task. We defined the maximum error achieved on a task before progression to the next point on the syllabus to be 2 bit errors per sequence for the Copy problem and 1 bit error per sequence on the Repeat Copy and Associative Recall problems.4 Results
4.1 Copy
Figure 1 shows the median learning curves over the 10 training runs per syllabus for the targettask and multitask setting. Training directly on the target task converges to near zero error 2.2 times faster than the next fastest syllabus  Predictive Gain. This demonstrates that in the targettask setting a syllabus is not required to solve the task. Predictive Gain adapts most rapidly to this situation and learns to focus on tasks close to the target task. Only the Look Back and Forward syllabus manages to reach near zero error in the targettask setting in the time provided. All the handcrafted syllabuses must reach the same level of success on a task to progress so the performance of Look Back and Forward relative to the other handcrafted syllabuses shows that sampling from tasks ahead of the current task in the syllabus allows much faster progression through the tasks.
As expected, in the multitask setting training directly on the target task performs very poorly when evaluated on all tasks. Despite making some progress towards the target task, the Naive syllabus shows almost no improvement on random performance in the multitask setting. This demonstrates that the network rapidly forgets previously learned tasks if not shown further examples of them. Predictive Gain, Uniform and Look Back and Forward all converge to near zero error in the multitask setting at similar rates, fig. 1.
Figure 2 shows the range of error achieved by the 10 trained networks for each syllabus when the networks are asked to generalize to sequences of length 40 on the Copy problem. Despite the networks of three syllabuses converging to near zero error on the target task with sequences of length 32 none of the networks succeed in generalizing to sequences of length 40. On sequences of length 40, the Prediction Gain and the Uniform syllabus demonstrate similar performance and have approximately 1.351.59 times lower median error than the Look Back and Forward, None and Look Back Syllabuses. There is substantial overlap in the range of generalization error for all syllabuses, so we cannot say that any one syllabus clearly outperforms the others in terms of improving generalization on the Copy problem.
4.2 Repeat Copy
Figure 3 shows the median learning curves for each syllabus on the Repeat Copy problem for the targettask and multitask setting. Unlike for the Copy problem In the targettask setting for the Repeat Copy problem, the network fails to learn a solution in the time provided by training directly on the target task, fig. 3. Interestingly the three handcrafted syllabuses converge to near zero error at approximately the same time and 3.5 times faster than the Uniform syllabus which is the next fastest. This is a clear win for the handcrafted syllabuses over the benchmark and automated syllabuses.
The Uniform syllabus converges to near zero error twice as fast as Predictive Gain. It is unclear why Predictive Gain has slower convergence although we posit a potential explanation  that scaling the rewards by the length of the input sequence as per the specification [5] may bias the bandit towards tasks with high repeats as such tasks incur no penalty for their added difficulty. This highlights that despite the automated nature of Predictive Gain’s syllabus generation it still relies on several tunable parameters.
Despite the Uniform syllabus’ slow convergence on the target task, in the multitask setting training on the same distribution as the test distribution is beneficial, as would be expected in non curriculum learning settings, fig. 3. The Uniform syllabus reaches near zero error 1.35 times faster than Look Back and Forward, the next fastest syllabus.
All the syllabuses but Predictive Gain and None consistently generalize with near zero error to sequences of length 16 on Repeat Copy, fig. 4. Predictive Gain’s wide generalization error range is explained by the fact that training with Predictive Gain on the Repeat Copy problem is unstable and 4 out of the 10 training runs failed to converge to near zero error. Whereas when we attempt to increase the number of repeats, all the syllabuses fail to generalize successfully, fig. 4, this is consistent with previous results [5].
4.3 Associative Recall
Figure 5 shows the median learning curves for each syllabus on Associative Recall for the targettask and multitask setting. Only Predictive Gain converges to a near zero error solution to the Associative Recall problem in either setting. The Uniform and Look Back and Forward syllabuses perform similarly in both settings and make some progress towards a low error solution. The Naive syllabus makes some progress towards learning the targettask and perhaps given enough time would learn a solution to the task. The Look Back and None syllabuses make very limited progress from random initialization in either setting.
Despite converging to a near zero error solution in both the targettask and multitask setting, figure 6 shows that Prediction Gain fails to generalize with similar error to sequences of 16 items. As expected, the other syllabuses which do not reach near zero error on the target task also exhibit similarly high generalization error.
The experiments on the above three problems demonstrate that none of the syllabuses tested has a clear advantage in all problems or settings. Over the three problems Look Back and Forward learned consistently faster than the other two handcrafted syllabuses; Look Back and Naive. Prediction Gain was fastest of the nonbenchmark syllabuses to converge to near zero error solutions on the Copy and Associative Recall problems but the opposite was true on the Repeat Copy problem (when training with Predictive Gain was unstable). Similar to other results [5] we found that the Uniform syllabus provides a strong baseline in the targettask setting. In the multitask setting uniformly sampling from all tasks leads to the fastest learning in two of the three problems. No syllabus consistently improved generalization on any of the above problems.
5 Conclusion
Curriculum learning has been demonstrated to be necessary for learning on several problems that neural networks cannot solve when presented examples solely from the target task [9, 23, 15] and to improve speed to convergence on other problems [5, 22]. These empirical results along with some recently observed theoretical properties of curriculum learning [19]
make it a potentially important area for progress in machine learning.
Despite this, curriculum learning is still a relatively unexplored area of machine learning. In this paper, we have conducted an empirical comparison of six syllabuses identified in the curriculum learning literature. We have demonstrated that the best choice of syllabus is task dependent and that the syllabuses we compared showed little effect on the generalization ability of our trained networks, contrary to some earlier findings [2].
We have shown that including examples from previously learned tasks is vital to prevent “catastrophic forgetting” of these tasks. Our results also demonstrated that including examples from future tasks on the syllabus provided substantial increases in speed of learning. This provides evidence for existing theories that naive progression through a syllabus requires substantial retraining of network weights at each progression point.
This is the first empirical comparison of a recently proposed automated syllabus  Predictive Gain, to stronger handcrafted syllabuses than the Uniform and None syllabuses. Despite being compared to stronger syllabuses Predictive Gain performed competitively. We conclude that automated approaches to syllabus design in curriculum learning may be a fruitful future area of development but that further work is required on how to set the hyperparameters governing the existing automated approaches.
Our empirical results provide the basis for the choice of syllabus in new applications of curriculum learning. When applying curriculum learning to problems in the multitask setting we recommend using the non curriculum learning syllabus of uniformly sampling from all tasks. Other syllabuses may provide marginal gains on some problems in the multitask setting, but this is not reliable across all problems and requires additional hyperparameter tuning. We recommend that when applying curriculum learning to problems in the targettask setting, practitioners use either Predictive Gain or the Look Back and Forward syllabus.
5.0.1 Acknowledgements
This publication emanated from research conducted with the financial support of Science Foundation Ireland (SFI) under Grant Number 13/RC/2106.
References
 [1] Auer, P., CesaBianchi, N., Freund, Y., Schapire, R.E.: The nonstochastic multiarmed bandit problem. SIAM journal on computing 32(1), 48–77 (2002)
 [2] Bengio, Y., Louradour, J., Collobert, R., Weston, J.: Curriculum learning. In: Proceedings of the 26th annual international conference on machine learning. pp. 41–48. ACM
 [3] Cirik, V., Hovy, E., Morency, L.P.: Visualizing and understanding curriculum learning for long shortterm memory networks. arXiv preprint arXiv:1611.06204 (2016)
 [4] Elman, J.L.: Learning and development in neural networks: The importance of starting small. Cognition 48(1), 71–99 (1993)
 [5] Graves, A., Bellemare, M.G., Menick, J., Munos, R., Kavukcuoglu, K.: Automated curriculum learning for neural networks. arXiv preprint arXiv:1704.03003 (2017)
 [6] Graves, A., Liwicki, M., Fernández, S., Bertolami, R., Bunke, H., Schmidhuber, J.: A novel connectionist system for unconstrained handwriting recognition. IEEE transactions on pattern analysis and machine intelligence 31(5), 855–868 (2009)

[7]
Graves, A., Mohamed, A.r., Hinton, G.: Speech recognition with deep recurrent neural networks. In: Acoustics, speech and signal processing (icassp), 2013 ieee international conference on. pp. 6645–6649. IEEE
 [8] Graves, A., Wayne, G., Danihelka, I.: Neural turing machines. arXiv preprint arXiv:1410.5401 (2014)
 [9] Graves, A., Wayne, G., Reynolds, M., Harley, T., Danihelka, I., GrabskaBarwińska, A., Colmenarejo, S.G., Grefenstette, E., Ramalho, T., Agapiou, J.: Hybrid computing using a neural network with dynamic external memory. Nature 538(7626), 471 (2016)
 [10] Grünwald, P.D.: The minimum description length principle. MIT press (2007)

[11]
Hochreiter, S., Schmidhuber, J.: Long shortterm memory. Neural computation 9(8), 1735–1780 (1997)
 [12] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
 [13] Krueger, K.A., Dayan, P.: Flexible shaping: How learning in small steps helps. Cognition 110(3), 380–394 (2009)
 [14] McCloskey, M., Cohen, N.J.: Catastrophic interference in connectionist networks: The sequential learning problem. In: Psychology of learning and motivation, vol. 24, pp. 109–165. Elsevier (1989)
 [15] Rae, J., Hunt, J.J., Danihelka, I., Harley, T., Senior, A.W., Wayne, G., Graves, A., Lillicrap, T.: Scaling memoryaugmented neural networks with sparse reads and writes. In: Advances in Neural Information Processing Systems. pp. 3621–3629
 [16] Reed, S., De Freitas, N.: Neural programmerinterpreters. arXiv preprint arXiv:1511.06279 (2015)
 [17] Rissanen, J.: Stochastic complexity and modeling. The annals of statistics pp. 1080–1100 (1986)
 [18] Spitkovsky, V.I., Alshawi, H., Jurafsky, D.: From baby steps to leapfrog: How less is more in unsupervised dependency parsing. In: Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics. pp. 751–759. Association for Computational Linguistics (2010)
 [19] Weinshall, D., Cohen, G.: Curriculum learning by transfer learning: Theory and experiments with deep networks. CoRR abs/1802.03796 (2018)
 [20] Williams, R.J.: Simple statistical gradientfollowing algorithms for connectionist reinforcement learning, pp. 5–32. Springer (1992)
 [21] Wu, Y., Schuster, M., Chen, Z., Le, Q.V., Norouzi, M., Macherey, W., Krikun, M., Cao, Y., Gao, Q., Macherey, K.: Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)
 [22] Zaremba, W., Sutskever, I.: Learning to execute. arXiv preprint arXiv:1410.4615 (2014)
 [23] Zaremba, W., Sutskever, I.: Reinforcement learning neural turing machinesrevised. arXiv preprint arXiv:1505.00521 (2015)
Comments
There are no comments yet.