Following the post on the future of work, was thinking about what implications this would have for education, and the most obvious connection between work and education is about credentials. These are the signposts that tell (current or future) employers that a person has a certain set of characteristics. The most obvious example of credentials is the degree which your college/university has given you, telling the world that you meet a certain set of criteria. Often, this criteria is somewhat obscure, and may mean all things to all people, as we can see from the fact that the same credential from different universities mean different things, as seen from the value that people assign to them.
Today, a college degree has immense value for an employer, because the college degree tells the employer that the student has gone through a certain set of courses, and therefore is the right person to meet the requirements of the employers. From the employer’s perspective, the degree tells them that the prospective employee has the skills to be able to build a career. What employers look for is the assurance that the prospective employee has what it takes to fit into the grand scheme of things, to become a part of the larger picture that their organisation represents.
However, as the nature of work changes, as I said before, would such a credential of an ability to learn all things be as important? I believe that in such a scenario, where an individual would be contributing their specific quantum of work in a larger value chain as a ‘freelancer’ the skills of the individual in that particular space would become much more important than their generic ability. This means that organisations would naturally be more interesting in evidence of achievement in that specific area.
Such a shift in focus from organisations would necessarily mean that the ability to demonstrate ability in a particular area would become more valuable than the ability to demonstrate overall/generic ability. Hence, I feel, artefacts generated by individuals in the course of their learning, whether in the form of project reports, or papers authored, or creative work, would probably have a far greater impact than the degree. So, for instance, a paper written by a student on a particular topic, related to the work sphere of the student would likely have far more interest for employers than the degree or the grade would.
In other words, the evidence of achievement, in the form of artefacts, or in the form of eminence would become a far more valuable resource by which to evaluate prospective employees than simply the degree.
In todays L&D landscape, the way businesses determine who should participate in what training isnt far away from some sort of conjuring act. More often than not, the result of this is a mixed bag, and many of the L&D professionals I speak to tell me that the L1 scores (based on the Kirkpatrick model) are more often than not tending towards the lower end of the spectrum.
There are typically two ways a business determines training participation. One is based on mandated training (usually related to promotion/growth), while the other is nomination by the business manager. Both of these are based on picking up from a ‘menu’ of available programs, and neither really takes into consideration the actual learning needs of the individual.
This is where the idea of predictive learning comes in. The idea here is simple … today, with the technology available to us, especially in the Big Data/Analytics domains, the data about what has worked in the past in what context is available to the organization in a large scale. This data is available based on training, HR, and operations/business data. This rich data can be leveraged to determine what is the best training solution which would likely work in a particular employee context. Like Big Data, this neednt look at the reason (or connection) between cause and effect, rather, look at the linkages as they have been seen in the past.
An important aspect of this picture is that this shifts the focus from training and learning, and from L&D to the individual learner, and makes the entire process people-centric.
One concern with this, though, could be that the outcome of the requirements could be way too granular, and too tailored to individual needs, so as to be unviable from the delivery perspective. More about this later …
There is, for obvious interest, quite a bit of interest in the subject of subjects, or, rather, in the subject of examinations on those subjects. The newspapers recently reported that Central Advisory Board of Education has recommended re-introduction of the class X exams. Another subject the article talks about is the policy of student detention based on exam results.
This brings us to a basic question … what is the purpose of exams? While there is definitely a need within the education system to assess achievement of learning objectives, the problem begins when exams are seen as a mechanism to weed out students who may not meet the criterion of meeting learning objectives. If the intent is to ensure that students learn the things they are supposed to, would studying the same thing again help a student understand better than the first time? This is akin to repeating something in the hope that just by repeating it, the other person will understand it. If the student didnt understand it the first time, isnt it more than likely that he wont understand the next time either?
Instead of having the student go through the entire year, it would be more helpful for the student if the focus was to be on topics the student was facing difficulty in understanding. A quick look at the answer sheet for the exam will give the answer. This, though, wont scale without the use of technology to support this, and today, we have the technology to move assessment in this direction.
Another aspect is to find out what we are testing. Are we testing memory of the subject, or are we testing understanding? If we are trying to assess achievement of learning objectives, we need to focus on understanding. This means the pattern of testing needs to change towards application of concepts from simple recitation of concepts, and we, as a nation, probably need an examination/assessment policy to complement the education and learning frameworks in the country.
We are told that marks (or grades) and qualifications are signals which serve to tell prospective employers about the worthiness of candidates for jobs … this as per classical economic theory. However, reading this article makes one think … what are marks measuring in the contemporary examination system in India?
There are a few possible things one could deduce from here:
- Children graduating schools are made up of different stuff, and are extremely bright.
- The University folks have lost it.
- The exam system is not exactly measuring earning.
Back when we were in school (this is another millenium, remember!) getting 80% in English meant you were really, really good at the subject. Mere mortals managed anything in the low to mid-70s, with some folks managing the 60s. Today, we are seeing a cut-off of 100% for Computer Science courses. If this is based on PCM (Physics, Chemistry, Mathematics), then one can assume that the kids are graduating school with exceptional understanding of the subjects. However, by the time these kids graduate, we find that corporates struggle to meet their hiring numbers. On the other hand, scoring 90s in English today should mean the kids should have an exceptional grasp of the language, but that isnt borne by observation.
Personally, I believe that the exam system is barking up the wrong tree (for biologists), or climbing up the wrong pole (for the rest of us). Marks dont seem to be measuring learning, though I dont know what they are measuring. To get a real understanding, exams need to test the kids, not on straight application of formulae, but to ask questions two or three steps removed from the data. And this isnt quite difficult to do.
About blogging, I find, the ideas are relatively simple to define and build upon, while the topic is the difficult part. The topic being done, let’s get to the idea of this post.
There are two basic problems with education delivery today, I feel. One is the shortage of skilled teachers, and the other is the piecemeal approach to concepts taught to children. Let’s look at these one by one.
When talking about shortage of skilled teachers, I don’t mean at the aggregate level. The problem facing education is that while in pockets skilled teachers are available, there are also pockets where they are not. In the age of technology, this gap should be one that should be easy to bridge. Having said that, I believe that on can’t rely on purely virtual education delivery when introducing a new subject, which is what is the a of the education system for the most part. So we might need to look at a blended approach towards doing that. For this blended approach, content taught could be divided into three parts: pure theory, guided problem-solving, and practical application. Concepts of a subject could be taught through recorded lectures, for which a panel of distinguished teachers could be identified from the vast pool available. This would help being the best teachers to students in areas where they are not available. At the same time, this would standardize the delivery of basic concepts. Building upon these concepts, to develop skills of applying these concepts, a set of real-life scenarios could be developed. These would need to be delivered in-person, and would help students to learn how these concepts could be applied to solve problems. These could be simulations or case studies, depending on the requirements of the course being taught. Building on these would be labs to experiment and to apply concepts, as required.
The other aspect is the piecemeal approach to teaching. Children are taught mathematics and physics and economics and history as separate subjects. The concepts they learn and their application are demarcated by subject, while in reality, these subjects are interrelated, and so also should their teaching be. To address this, and to develop holistic problem-solving skills, which enable students to see the big picture, appreciate systems, and building systemic thinking, classwork, homework, and exams should be based on a systemic approach, where students are able to see the system as a whole, understand it in it’s entirety, and be able to understand impact of one thing on another.
Let me describe a scenario that i came across, of gamification changing the way one tends to interact with something. I have been playing, as i suppose a lot of you have been, sudoku for some time. I have a version of sudoku which is a straightforward one. You select the level of the puzzle you want to play, and it generates the puzzle for you, and you go ahead. It tracks the best times for completing puzzles with different levels of difficulty.
Some time back i got another version of sudoku which has puzzles at different levels of difficulty. The difference is that in this one, the higher levels of difficulty need to be unlocked. Only once you have solved the puzzles at easy level, with a particular best time, and some other parameters, does the medium, difficult, and super levels get unlocked. You cant just jump to a higher level puzzle.
The same game, with simple functionality included creates a different level of motivation to solve the puzzle less time. Earlier, i wasnt too aware of the best timing for the different levels, but now, i was keeping track. And unlocking the level gave a sense of achievement. In short, i tended to look at two versions of the same game in different ways, and this was because of the small component of gamification introduced.
This is the impact of gamification, and this could be harnessed to create the motivation, engagement, and a sense of achievement, like in a game, in learning.
A recurring question which keeps coming up in discussions is how mathematics should be taught. There is a strong view that given the computing power which is available to us, we should relook the basic maths curriculum. So i went looking, and found this video which i feel looks at the problem in a logical way.
Conrad Wolfram is giving some compelling reasoning for why maths education should change, and gives a description of how it should change, too.
Hand-culcating the mundane way should not be the focus on the curriculum. Rather, the focus should be on building and enforcing the concepts, and leave the calculating to computers. In other words, as he says, students should be taught the procedures which define fundamental concepts, but the implementation of those should be left to computer programs. For example, students should know what a square root it, how it is calculated, but they shouldnt have to calculate, beyond illustrations. And here is the cool part he says … focus on teaching students how to write programs to calculate square roots, rather than having them mechanically execute the procedure for calculating. This will immensely help students clarify their concepts (how can one write a program without understanding the underlying principles to a very large extent?), while at the same time help them become more comfortable with the concept of application of these concepts. In other words, our mathematics curriculum should stress understanding and application (application to real world problems is a very good way of teaching these concepts) rather than stress on the mundane calculations which stress out students as well as parents alike. After all, why should a child lose marks in an exam (thats what happens) if he or she takes the square of 5 to be 10 when all the conceptual aspects of the solution are correct, and the only mistake is a calculation mistake?
Connect this with the post i had written earlier, and a rather innovative picture of mathematics teaching emerges.