{"id":19647,"date":"2026-04-28T07:31:09","date_gmt":"2026-04-28T07:31:09","guid":{"rendered":"https:\/\/www.europesays.com\/ai\/19647\/"},"modified":"2026-04-28T07:31:09","modified_gmt":"2026-04-28T07:31:09","slug":"5-ai-myths-and-why-we-must-move-past-them-opinion","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/ai\/19647\/","title":{"rendered":"5 AI Myths and Why We Must Move Past Them (opinion)"},"content":{"rendered":"<p>We are ready for an exciting set of conversations about AI and higher education. We want to talk with faculty about what <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.penguinrandomhouse.com\/books\/741805\/co-intelligence-by-ethan-mollick\/\" target=\"_blank\">Ethan Mollick\u2019s idea of \u201cco-intelligence\u201d<\/a> means in practice in the classroom. We need to be engaging the employers of our graduates to redesign the curriculum. We are even ready for difficult conversations like what happens when a professor wants a class to use AI but some students have ethical objections.<\/p>\n<p>These conversations are happening among the influencer class of AI writers and thinkers, but not as much in the day-to-day work with faculty. This past year, we\u2019ve realized there are a set of conversations and interventions we have to move beyond if we want to get to the good stuff.<\/p>\n<p>The artificial intelligence era in higher education seems too new for it to be overrun with myths and outdated advice, but here we are. Popular blogs highlight archaic tricks for detecting AI writing, and teaching and learning websites recommend tactics that have not worked for more than a year. Some of this is to be forgiven\u2014we have never seen a technology move this fast, so adaptation is ongoing rather than a one-off event.<\/p>\n<p>As faculty members who are deeply invested in understanding the good and bad of AI in modern classrooms, our goal here is to discuss some of the more prominent and stubborn myths that persist in conversations about AI and teaching and learning.<\/p>\n<p>I Know It When I See It<\/p>\n<p>Guides to spotting AI-produced work continue to <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.captechu.edu\/blog\/how-spot-ai-generated-content-it-fact-or-fiction\" target=\"_blank\">exist<\/a> and be <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.eastcentral.edu\/free\/ai-faculty-resources\/detecting-ai-generated-text\/\" target=\"_blank\">posted<\/a> on college and university websites. <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.tomsguide.com\/ai\/how-to-spot-ai-writing-5-telltale-signs-to-look-for\" target=\"_blank\">Tech publications<\/a> and <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/elearningindustry.com\/how-to-tell-if-something-is-written-by-ai-telltale-signs-to-look-for\" target=\"_blank\">journalists<\/a> continue to publish advice on this as well. Casey Newton from Platformer and Hard Fork recently said of <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.nytimes.com\/2025\/09\/19\/podcasts\/hardfork-charlie-kirk.html?showTranscript=1\" target=\"_blank\">AI writing<\/a>, \u201cI can always tell.\u201d<\/p>\n<p>This is empirically untrue. <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.rivista.ai\/wp-content\/uploads\/2024\/05\/1-s2.0-S2666920X24000109-main.pdf\" target=\"_blank\">Study<\/a> after <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.nature.com\/articles\/s41746-023-00819-6.pdf\" target=\"_blank\">study<\/a> confirms that humans are incredibly bad at detecting AI writing: It\u2019s more the case that we can spot people who are bad at using AI. Consider some of the advice common in these guides, such as that an abundance of em dashes or bullet points indicates likely AI involvement. You might well find students using lots of em dashes and in conversation discover they have used AI. However, this is just selection bias since you never knew about the other students who also used it but whom you didn\u2019t identify. Spotting unauthorized AI use ultimately becomes a process based on vibes that necessarily introduces instructor bias into another aspect of teaching.<\/p>\n<p>AI Can\u2019t Do Personal Reflection<\/p>\n<p>This was a key component of a popular article in The Atlantic from the fall in which the author, the linguist John McWhorter, posited, \u201cI have also found ways of posing questions that get past what AI can answer, such as asking for a personal take\u00a0\u2026 that draws from material discussed in class.\u201d This belief that asking for personal reflection is a form of AI-proofing is <a href=\"https:\/\/nmu.edu\/ctl\/creating-ai-resistant-assignments-activities-and-assessments-designing-out\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">remarkably<\/a> <a href=\"https:\/\/www.teachingentrepreneurship.org\/ai-proof-assignments\/?utm_source=chatgpt.com\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">widespread<\/a>, even on centers for <a href=\"https:\/\/www.uis.edu\/news\/center-online-learning-research-service\/faculty\/colrs-teaching-tip-ai-proofing-assignments?utm_source=chatgpt.com\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">teaching and learning websites<\/a>.<\/p>\n<p>But this is just basic prompting. Say you assign a personal reflection that asks students to critically think about how they are shaped by social inequalities (race, gender, class) in their everyday lives. Student A might copy the assignment instructions into an LLM and write to the LLM, \u201cDo this assignment for me.\u201d Student B copies the instructions but in this case the student adds personal details about themself such as race and ethnicity, income, age, place of residence, and such. We strongly suggest everyone try this out for themselves with their own assignments. There are even <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/humanizeai.com\/?ref=kyle&amp;utm_source=newsletter.theaireport.ai&amp;utm_medium=newsletter&amp;utm_campaign=chinese-hackers-use-anthropic-s-claude&amp;_bhlid=9e5852422f9e760fdc9829961231bbd2a54c99de\" target=\"_blank\">websites<\/a> that allow users to \u201chumanize\u201d their AI-generated text, meaning you may not even need to be good at prompting to do this.<\/p>\n<p>The Calendar Trick<\/p>\n<p>The 2023 versions of ChatGPT were trained on data only up to 2021. So a <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/larryferlazzo.edublogs.org\/2023\/06\/08\/i-asked-artificial-intelligence-to-recommend-student-writing-prompts-that-cant-be-easily-helped-by-ai-here-are-14-of-them\/?utm_source=chatgpt.com\" target=\"_blank\">popular tactic<\/a> early on was to ask questions about current events that the systems did not have access to. This has somehow persisted as an approach. It is something we have both encountered working with faculty, and it even shows up in modified form in some guides, like this one <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.montclair.edu\/faculty-excellence\/ofe-teaching-principles\/clear-course-design\/practical-responses-to-chat-gpt\/teaching-with-chatgpt-assignment-design-tips-ideas\/\" target=\"_blank\">from Montclair State University<\/a> that recommends connecting assignments to \u201cvery recent events\u201d as that is context \u201cAI will not have.\u201d<\/p>\n<p>Obviously, this advice is no longer viable, as some models like Grok and Llama are fully integrated into real-time social media networks and other providers are in negotiations with news outlets to have breaking news\u2013informed outputs.<\/p>\n<p>Trojan Horse<\/p>\n<p>In 2024, a teacher went viral with the strategy of embedding <a rel=\"noreferrer noopener nofollow\" href=\"https:\/\/www.newsweek.com\/teacher-clever-hack-catching-students-using-chatgpt-essay-1893623?utm_source=chatgpt.com\" target=\"_blank\">hidden instructions<\/a> in an essay prompt using small, white font. The idea is that the student will unwittingly copy and paste the whole assignment guide into the AI platform, which will read the hidden prompt. The resulting output will incorporate something the instructor called for, like incorporating references to a movie or a theorist not covered in class. Of course, this would also work with STEM prompts and result in errant pieces of code or components of a proof that are unnecessary.<\/p>\n<p>At best this is \u201cgotcha\u201d teaching, where instructors move from teachers to pranksters. Is this the role we want to be playing? If your goal is to catch students in the act, you\u2019re not starting out from a place of good pedagogy. This is also an unreliable approach with a limited shelf life. The Trojan horse is basically a variation of a prompt-injection attack, which is a known security issue for LLMs that engineers are actively trying to resolve. The greater risk is that if an instructor relies on this as a tactic, it immediately absolves everyone who passes their \u201ctest.\u201d Even if the trick works, you aren\u2019t spotting all the students who are using the tools: You are only spotting students who are bad at using them.<\/p>\n<p>Claude\u2019s Not in Our Classroom<\/p>\n<p>Some universities <a href=\"https:\/\/provost.uiowa.edu\/ai-classroom?utm_source=chatgpt.com\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">recommend<\/a> tying assignments closely to classroom discussions. Unlike the other approaches, this one is founded in good pedagogy. This sort of woven approach to instruction is absolutely a best practice\u2014but it\u2019s not an effective way to deter the use of large language models.<\/p>\n<p>A student can simply add the context from the classroom discussion to the prompt and the resulting output will incorporate it. Yet we do think there is merit here, insofar as we view this as a best practice in teaching students how to prompt well.<\/p>\n<p>Trying to get students to not use these tools seems to be a losing battle, and we are not even convinced it is one worth fighting. We did not even touch on the necessity of working with these tools to prepare students for a competitive job market where employers increasingly expect AI literacy and tool usage. The point of this piece was simply to clarify a series of misconceptions and the persistence of outdated advice.<\/p>\n<p>We recognize that people are in different places with AI depending on the ethical questions they have and how much they have used the tools. There are plenty of unsettled questions about sustainability, learning loss, human creativity and labor. We are ready to have more of those conversations. We would prefer to have fewer instances where we have to break the news to one of our colleagues that the \u201cone neat trick\u201d they learned from TikTok two years ago doesn\u2019t work anymore.<\/p>\n<p>Zach Justus is director of faculty development and a professor of communication arts and sciences at California State University, Chico.<\/p>\n<p>Nik Janos is a professor of sociology at California State University, Chico.<\/p>\n","protected":false},"excerpt":{"rendered":"We are ready for an exciting set of conversations about AI and higher education. We want to talk&hellip;\n","protected":false},"author":2,"featured_media":19648,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[24,25,501,76,293,500,382,66],"class_list":{"0":"post-19647","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-ai","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-career","11":"tag-education","12":"tag-events","13":"tag-higher","14":"tag-jobs","15":"tag-news"},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/19647","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/comments?post=19647"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/posts\/19647\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media\/19648"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/media?parent=19647"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/categories?post=19647"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/ai\/wp-json\/wp\/v2\/tags?post=19647"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}