This image/report itself doesn’t make much sense – probably it was generated by chatGPT itself.
“What makes your job exposed to GPT?” – OK I expect a list of possible answers:
“Low wages”: OK, having a low wage makes my job exposed to GPT.
“Manufacturing”: OK, manufacturing makes my job exposed to GPT. …No wait, what does that mean?? You mean if my job is about manufacturing, then it’s exposed to GPT? OK but then shouldn’t this be listed under the next question, “What jobs are exposed to GPT?”?
…
“Jobs requiring low formal education”: what?! The question was “what makes your job exposed to GPT?”. From this answer I get that “jobs requiring low formal education make my job exposed to GPT”. Or I get that who/whatever wrote this knows no syntax or semantics. OK, sorry, you meant “If your job requires low formal education, then it’s exposed to GPT”. But then shouldn’t this answer also be listed under the next question??
“What jobs are exposed to GPT?”
“Athletes”. Well, “athletes” semantically speaking is not a job; maybe “athletics” is a job. But who gives a shirt about semantics? there’s chatGPT today after all.
The same with the rest. “Stonemasonry” is a job, “stonemasons” are the people who do that job. At least the question could have been “Which job categories are exposed to GPT?”.
“Pile drivers operators”: this very specific job is thankfully Low Exposure.
“High exposure: Mathematicians”. Mmm… wait, wait. Didn’t you say that “Science skills” and “Critical thinking skills” were “Low Exposure”, in the previous question?
I’ve seen GPT struggling with pretty basic maths and “abstract” tasks such as making the letters add up in an anagram. Math requires insight that a language model cannot posess. I don’t really get why people like infographics so much. The format usually just distracts from the data presented, which is convenient given that the data is usually garbage too.
Math requires insight that a language model cannot posess
Amen to that! Good maths & science teachers have struggled for decades (if not centuries) so that students understand what they’re doing and don’t simply give answers based on some words or symbols they see in questions [there are also bad teachers who promote this instead]. Because on closer inspection such answers always collapse. And now comes chatGPT that does exactly that instead – and collapses in the same way – and gets glorified.
Amen to what you say on infographic content as well 😂
This image/report itself doesn’t make much sense – probably it was generated by chatGPT itself.
“What makes your job exposed to GPT?” – OK I expect a list of possible answers:
“What jobs are exposed to GPT?”
Icanhazcheezeburger? 🤣
I’ve seen GPT struggling with pretty basic maths and “abstract” tasks such as making the letters add up in an anagram. Math requires insight that a language model cannot posess. I don’t really get why people like infographics so much. The format usually just distracts from the data presented, which is convenient given that the data is usually garbage too.
Amen to that! Good maths & science teachers have struggled for decades (if not centuries) so that students understand what they’re doing and don’t simply give answers based on some words or symbols they see in questions [there are also bad teachers who promote this instead]. Because on closer inspection such answers always collapse. And now comes chatGPT that does exactly that instead – and collapses in the same way – and gets glorified.
Amen to what you say on infographic content as well 😂