Following the post on the future of work, was thinking about what implications this would have for education, and the most obvious connection between work and education is about credentials. These are the signposts that tell (current or future) employers that a person has a certain set of characteristics. The most obvious example of credentials is the degree which your college/university has given you, telling the world that you meet a certain set of criteria. Often, this criteria is somewhat obscure, and may mean all things to all people, as we can see from the fact that the same credential from different universities mean different things, as seen from the value that people assign to them.
Today, a college degree has immense value for an employer, because the college degree tells the employer that the student has gone through a certain set of courses, and therefore is the right person to meet the requirements of the employers. From the employer’s perspective, the degree tells them that the prospective employee has the skills to be able to build a career. What employers look for is the assurance that the prospective employee has what it takes to fit into the grand scheme of things, to become a part of the larger picture that their organisation represents.
However, as the nature of work changes, as I said before, would such a credential of an ability to learn all things be as important? I believe that in such a scenario, where an individual would be contributing their specific quantum of work in a larger value chain as a ‘freelancer’ the skills of the individual in that particular space would become much more important than their generic ability. This means that organisations would naturally be more interesting in evidence of achievement in that specific area.
Such a shift in focus from organisations would necessarily mean that the ability to demonstrate ability in a particular area would become more valuable than the ability to demonstrate overall/generic ability. Hence, I feel, artefacts generated by individuals in the course of their learning, whether in the form of project reports, or papers authored, or creative work, would probably have a far greater impact than the degree. So, for instance, a paper written by a student on a particular topic, related to the work sphere of the student would likely have far more interest for employers than the degree or the grade would.
In other words, the evidence of achievement, in the form of artefacts, or in the form of eminence would become a far more valuable resource by which to evaluate prospective employees than simply the degree.
The way we work has been undergoing massive changes over the last decade or more, but today, I believe, we are at the cusp of a fundamental shift in the relations of work, facilitated by the developments in technology. By relations of work, I mean the role each individual plays in a ‘value chain’ and how the part contributes to the whole.
Before the advent of the modern corporation, people worked not for a corporation (they weren’t around, remember?). Rather, artisans, for instance, manufactured their final product, say a bicycle (if they were around …) as a single entity, and sold their products in a marketplace.
With the advent of the corporation came the concept of people working in jobs where they did specific work, which contributed (often in indefinable ways) to the overall value chain. In this way, the individual would do their part of the work, and pass on their output to someone else, who would do their part of the work (value add) and so on …
This aspect is changing, and, I believe, set to change in bigger ways. As we are seeing there is a trend towards organizations outsourcing their work to freelance contractors. As this grows (and we are seeing this happening more so in the technology sector) we would likely come to a state where instead of many individuals being brought together under the ambit of the organizations, people would work more in their capacity as individuals, being brought together under the ambit of the value chain. This value chain, by definition, would span organizations, which means that we can expect to see, more and more, the value chain being formed as a loose federation of individual freelance contributors, their output orchestrated by a set of organizations partnering together to create a certain set of products or services.
So in terms of work structures this could likely be a move towards towards ways of working the modern corporation replaced, though in ways which are very much the new millennium. This has massive implications on the aspirations of youngsters (I don’t quite rely on the generation nomenclature, partly because I don’t understand it …), in that they can probably no longer aspire to long term jobs and designations may lose their meaning, the content of work, and the satisfaction that generates being the main defining factors there.
In a way, going back in time, but in a 21st century way.
In todays L&D landscape, the way businesses determine who should participate in what training isnt far away from some sort of conjuring act. More often than not, the result of this is a mixed bag, and many of the L&D professionals I speak to tell me that the L1 scores (based on the Kirkpatrick model) are more often than not tending towards the lower end of the spectrum.
There are typically two ways a business determines training participation. One is based on mandated training (usually related to promotion/growth), while the other is nomination by the business manager. Both of these are based on picking up from a ‘menu’ of available programs, and neither really takes into consideration the actual learning needs of the individual.
This is where the idea of predictive learning comes in. The idea here is simple … today, with the technology available to us, especially in the Big Data/Analytics domains, the data about what has worked in the past in what context is available to the organization in a large scale. This data is available based on training, HR, and operations/business data. This rich data can be leveraged to determine what is the best training solution which would likely work in a particular employee context. Like Big Data, this neednt look at the reason (or connection) between cause and effect, rather, look at the linkages as they have been seen in the past.
An important aspect of this picture is that this shifts the focus from training and learning, and from L&D to the individual learner, and makes the entire process people-centric.
One concern with this, though, could be that the outcome of the requirements could be way too granular, and too tailored to individual needs, so as to be unviable from the delivery perspective. More about this later …
About blogging, I find, the ideas are relatively simple to define and build upon, while the topic is the difficult part. The topic being done, let’s get to the idea of this post.
There are two basic problems with education delivery today, I feel. One is the shortage of skilled teachers, and the other is the piecemeal approach to concepts taught to children. Let’s look at these one by one.
When talking about shortage of skilled teachers, I don’t mean at the aggregate level. The problem facing education is that while in pockets skilled teachers are available, there are also pockets where they are not. In the age of technology, this gap should be one that should be easy to bridge. Having said that, I believe that on can’t rely on purely virtual education delivery when introducing a new subject, which is what is the a of the education system for the most part. So we might need to look at a blended approach towards doing that. For this blended approach, content taught could be divided into three parts: pure theory, guided problem-solving, and practical application. Concepts of a subject could be taught through recorded lectures, for which a panel of distinguished teachers could be identified from the vast pool available. This would help being the best teachers to students in areas where they are not available. At the same time, this would standardize the delivery of basic concepts. Building upon these concepts, to develop skills of applying these concepts, a set of real-life scenarios could be developed. These would need to be delivered in-person, and would help students to learn how these concepts could be applied to solve problems. These could be simulations or case studies, depending on the requirements of the course being taught. Building on these would be labs to experiment and to apply concepts, as required.
The other aspect is the piecemeal approach to teaching. Children are taught mathematics and physics and economics and history as separate subjects. The concepts they learn and their application are demarcated by subject, while in reality, these subjects are interrelated, and so also should their teaching be. To address this, and to develop holistic problem-solving skills, which enable students to see the big picture, appreciate systems, and building systemic thinking, classwork, homework, and exams should be based on a systemic approach, where students are able to see the system as a whole, understand it in it’s entirety, and be able to understand impact of one thing on another.
A recurring question which keeps coming up in discussions is how mathematics should be taught. There is a strong view that given the computing power which is available to us, we should relook the basic maths curriculum. So i went looking, and found this video which i feel looks at the problem in a logical way.
Conrad Wolfram is giving some compelling reasoning for why maths education should change, and gives a description of how it should change, too.
Hand-culcating the mundane way should not be the focus on the curriculum. Rather, the focus should be on building and enforcing the concepts, and leave the calculating to computers. In other words, as he says, students should be taught the procedures which define fundamental concepts, but the implementation of those should be left to computer programs. For example, students should know what a square root it, how it is calculated, but they shouldnt have to calculate, beyond illustrations. And here is the cool part he says … focus on teaching students how to write programs to calculate square roots, rather than having them mechanically execute the procedure for calculating. This will immensely help students clarify their concepts (how can one write a program without understanding the underlying principles to a very large extent?), while at the same time help them become more comfortable with the concept of application of these concepts. In other words, our mathematics curriculum should stress understanding and application (application to real world problems is a very good way of teaching these concepts) rather than stress on the mundane calculations which stress out students as well as parents alike. After all, why should a child lose marks in an exam (thats what happens) if he or she takes the square of 5 to be 10 when all the conceptual aspects of the solution are correct, and the only mistake is a calculation mistake?
Connect this with the post i had written earlier, and a rather innovative picture of mathematics teaching emerges.