On 2019/04/19 10:04, Simone Piccardi wrote:
Il 16/04/19 14:45, Mark Clarke ha scritto:
I would suggest that its not an either or approach. We could have a part that is multiple choice and a practical part. The practical part doesn't have to be under exam conditions. It could be a task like write a bash script that does x or some other assignemtn. The student is given 2 days to do the task and submit the script/assignement and the testing can be automated.

And how do you avoid having the student getting "help" from a friend?


That's an excellent point.

Another is how will an automated tester account for every variation that the candidate might have or do? Perhaps a candidate might validate an IP Address (sensible) and naturally uses Python with netaddr. Automated testing is likely to fail and the assignment, whilst correct, is marked wrong. Now manual intervention is needed and that means salaries. The cost of an exam just multiplies many times.

I've stayed out of this current discussion as it rears it's head every few years and never goes anywhere. Such discussions are tiring.

Someone earlier mentioned the perception that hands-on testing is better. I very much agree that it is a perception. It might not be true.

So what is hands-on testing good for? It's great for testing if a candidate can perform a series of predetermined steps in response to a given situation to produce a determined result. Hence why we test student pilots with it. And electricians, scuba divers and almost every action a sailor will do on the job (when sailors can't pass these tests, other sailors die).

It's why RedHat, Cisco and SuSE use practical tests - those distros provide specific tools to do specific functions and the candidate can rely on the tools to be present and work correctly. To do task X on RHEL regarding selinux, RHEL provides a tool, and it will be present on the test machine. The candidate is required to show they can drive the tool to produce the result RedHat demonstrated in the course.

In truth, this has very little to do with results, it has everything to do with the tool and how it is used, and the result is a side-effect. RedHat never puts anything in their low and mid level exams that is not covered in sufficient detail in their course materials, to do so would be very unfair. You can't expect someone to perform a task they were not taught how to do.

If we look at LPI's mission, we see that it is to a large degree exactly opposite to the above. LPI is not about RHEL tools, it is about the candidate proving they understand Linux systems within the scope of the level tested. Because the scope is not bound to a specific distro or release, testing has to be done on a somewhat abstract, conceptual level. There is nothing wrong with measuring the extent of conceptual knowledge and this is what LPI does.

Testing conceptual knowledge is not inherently better or worse than practical testing, they are simply different. Both have their place and they are answers to different questions about candidates and should not be conflated.


--
Alan McKinnon
[email protected]

_______________________________________________
lpi-examdev mailing list
[email protected]
https://list.lpi.org/mailman/listinfo/lpi-examdev

Reply via email to