GCFGlobal Logo

  • Get started with computers
  • Learn Microsoft Office
  • Apply for a job
  • Improve my work skills
  • Design nice-looking docs
  • Getting Started
  • Smartphones & Tablets
  • Typing Tutorial
  • Online Learning
  • Basic Internet Skills
  • Online Safety
  • Social Media
  • Zoom Basics
  • Google Docs
  • Google Sheets
  • Career Planning
  • Resume Writing
  • Cover Letters
  • Job Search and Networking
  • Business Communication
  • Entrepreneurship 101
  • Careers without College
  • Job Hunt for Today
  • 3D Printing
  • Freelancing 101
  • Personal Finance
  • Sharing Economy
  • Decision-Making
  • Graphic Design
  • Photography
  • Image Editing
  • Learning WordPress
  • Language Learning
  • Critical Thinking
  • For Educators
  • Translations
  • Staff Picks
  • English expand_more expand_less

Critical Thinking and Decision-Making  - What is Critical Thinking?

Critical thinking and decision-making  -, what is critical thinking, critical thinking and decision-making what is critical thinking.

GCFLearnFree Logo

Critical Thinking and Decision-Making: What is Critical Thinking?

Lesson 1: what is critical thinking, what is critical thinking.

Critical thinking is a term that gets thrown around a lot. You've probably heard it used often throughout the years whether it was in school, at work, or in everyday conversation. But when you stop to think about it, what exactly is critical thinking and how do you do it ?

Watch the video below to learn more about critical thinking.

Simply put, critical thinking is the act of deliberately analyzing information so that you can make better judgements and decisions . It involves using things like logic, reasoning, and creativity, to draw conclusions and generally understand things better.

illustration of the terms logic, reasoning, and creativity

This may sound like a pretty broad definition, and that's because critical thinking is a broad skill that can be applied to so many different situations. You can use it to prepare for a job interview, manage your time better, make decisions about purchasing things, and so much more.

The process

illustration of "thoughts" inside a human brain, with several being connected and "analyzed"

As humans, we are constantly thinking . It's something we can't turn off. But not all of it is critical thinking. No one thinks critically 100% of the time... that would be pretty exhausting! Instead, it's an intentional process , something that we consciously use when we're presented with difficult problems or important decisions.

Improving your critical thinking

illustration of the questions "What do I currently know?" and "How do I know this?"

In order to become a better critical thinker, it's important to ask questions when you're presented with a problem or decision, before jumping to any conclusions. You can start with simple ones like What do I currently know? and How do I know this? These can help to give you a better idea of what you're working with and, in some cases, simplify more complex issues.  

Real-world applications

illustration of a hand holding a smartphone displaying an article that reads, "Study: Cats are better than dogs"

Let's take a look at how we can use critical thinking to evaluate online information . Say a friend of yours posts a news article on social media and you're drawn to its headline. If you were to use your everyday automatic thinking, you might accept it as fact and move on. But if you were thinking critically, you would first analyze the available information and ask some questions :

  • What's the source of this article?
  • Is the headline potentially misleading?
  • What are my friend's general beliefs?
  • Do their beliefs inform why they might have shared this?

illustration of "Super Cat Blog" and "According to survery of cat owners" being highlighted from an article on a smartphone

After analyzing all of this information, you can draw a conclusion about whether or not you think the article is trustworthy.

Critical thinking has a wide range of real-world applications . It can help you to make better decisions, become more hireable, and generally better understand the world around you.

illustration of a lightbulb, a briefcase, and the world

/en/problem-solving-and-decision-making/why-is-it-so-hard-to-make-decisions/content/

Library homepage

  • school Campus Bookshelves
  • menu_book Bookshelves
  • perm_media Learning Objects
  • login Login
  • how_to_reg Request Instructor Account
  • hub Instructor Commons

Margin Size

  • Download Page (PDF)
  • Download Full Book (PDF)
  • Periodic Table
  • Physics Constants
  • Scientific Calculator
  • Reference & Cite
  • Tools expand_more
  • Readability

selected template will load here

This action is not available.

Social Sci LibreTexts

5.3: Using Critical Thinking Skills- Decision Making and Problem Solving

  • Last updated
  • Save as PDF
  • Page ID 24260

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

Introduction

In previous lessons, you learned about characteristics of critical thinkers and information literacy. In this module, you will learn how to put those skills into action through the important processes of decision making and problem solving.

As with the process of developing information literacy, asking questions is an important part of decision making and problem solving. Thinking is born of questions. Questions wake us up. Questions alert us to hidden assumptions. Questions promote curiosity and create new distinctions. Questions open up options that otherwise go unexplored. Besides, teachers love questions.

We make decisions all the time, whether we realize it or not. Even avoiding decisions is a form of decision making. The student who puts off studying for a test until the last minute, for example, might really be saying, “I’ve decided this course is not important” or “I’ve decided not to give this course much time.”

Decisions are specific and lead to focused action. When we decide, we narrow down. We give up actions that are inconsistent with our decision.

In addition to decision making, critical thinking skills are important to solving problems. We encounter problems every single day, and having a solid process in place is important to solving them.

At the end of the lesson, you will learn how to put your critical thinking skills to use by reviewing an example of how critical thinking skills can help with making those everyday decisions.

Using Critical Thinking Skills: Asking Questions

Questions have practical power. Asking for directions can shave hours off a trip. Asking a librarian for help can save hours of research time. Asking how to address an instructor—by first name or formal title—can change your relationship with that person. Asking your academic advisor a question can alter your entire education. Asking people about their career plans can alter your career plans.

You can use the following strategies to develop questions for problem solving and decision making:

Ask questions that create possibilities. At any moment, you can ask a question that opens up a new possibility for someone.

  • Suppose a friend walks up to you and says, “People just never listen to me.” You listen carefully. Then you say, “Let me make sure I understand. Who, specifically, doesn’t listen to you? And how do you know they’re not listening?”
  • Another friend tells you, “I just lost my job to someone who has less experience. That should never happen.” You respond, “Wow, that’s hard. I’m sorry you lost your job. Who can help you find another job?”
  • A relative seeks your advice. “My mother-in-law makes me mad,” she says. “You’re having a hard time with this person,” you say. “What does she say and do when you feel mad at her? And are there times when you don’t get mad at her?”

These kinds of questions—asked with compassion and a sense of timing—can help people move from complaining about problems to solving them.

Discover new questions. Students sometimes say, “I don’t know what questions to ask.” Consider the following ways to create questions about any subject you want to study or about any

area of your life that you want to change:

  • Let your pen start moving. Sometimes you can access a deeper level of knowledge by taking out your pen, putting it on a piece of paper, and writing down questions—even before you know what to write. Don’t think. Just watch the pen move across the paper. Notice what appears. The results might be surprising.
  • Ask about what’s missing . Another way to invent useful questions is to notice what’s missing from your life and then ask how to supply it. For example, if you want to take better notes, you can write, “What’s missing is skill in note taking. How can I gain more skill in taking notes?” If you always feel rushed, you can write, “What’s missing is time. How do I create enough time in my day to actually do the things that I say I want to do?”
  • Pretend to be someone else. Another way to invent questions is first to think of someone you greatly respect. Then pretend you’re that person. Ask the questions you think she would ask.
  • What can I do when ... an instructor calls on me in class and I have no idea what to say? When a teacher doesn’t show up for class on time? When I feel overwhelmed with assignments?
  • How can I ... take the kind of courses that I want? Expand my career options? Become much more effective as a student, starting today?
  • When do I ... decide on a major? Transfer to another school? Meet with an instructor to discuss an upcoming term paper?
  • What else do I want to know about ... my academic plan? My career plan? My options for job hunting? My friends? My relatives? My spouse?
  • Who can I ask about ... my career options? My major? My love life? My values and purpose in life?

Many times you can quickly generate questions by simply asking yourself, “What else do I want to know?” Ask this question immediately after you read a paragraph in a book or listen to someone speak.

Start from the assumption that you are brilliant. Then ask questions to unlock your brilliance.

Using Critical Thinking Skills in Decision Making

As you develop your critical thinking skills, you can apply them as you make decisions. The following suggestions can help in your decision-making process:

Recognize decisions. Decisions are more than wishes or desires. There’s a world of difference between “I wish I could be a better student” and “I will take more powerful notes, read with greater retention, and review my class notes daily.” Deciding to eat fruit for dessert instead of ice cream rules out the next trip to the ice cream store.

Establish priorities. Some decisions are trivial. No matter what the outcome, your life is not affected much. Other decisions can shape your circumstances for years. Devote more time and energy to the decisions with big outcomes.

Base decisions on a life plan. The benefit of having long-term goals for our lives is that they provide a basis for many of our daily decisions. Being certain about what we want to accomplish this year and this month makes today’s choices more clear.

Balance learning styles in decision making. To make decisions more effectively, use all four modes of learning explained in a previous lesson. The key is to balance reflection with action, and thinking with experience. First, take the time to think creatively, and generate many options. Then think critically about the possible consequences of each option before choosing one. Remember, however, that thinking is no substitute for experience. Act on your chosen option, and notice what happens. If you’re not getting the results you want, then quickly return to creative thinking to invent new options.

Choose an overall strategy. Every time you make a decision, you choose a strategy—even when you’re not aware of it. Effective decision makers can articulate and choose from among several strategies. For example:

  • Find all of the available options, and choose one deliberately. Save this strategy for times when you have a relatively small number of options, each of which leads to noticeably different results.
  • Find all of the available options, and choose one randomly. This strategy can be risky. Save it for times when your options are basically similar and fairness is the main issue.
  • Limit the options, and then choose. When deciding which search engine to use, visit many search sites and then narrow the list down to two or three from which to choose.

Use time as an ally. Sometimes we face dilemmas—situations in which any course of action leads to undesirable consequences. In such cases, consider putting a decision on hold. Wait it out. Do nothing until the circumstances change, making one alternative clearly preferable to another.

Use intuition. Some decisions seem to make themselves. A solution pops into your mind, and you gain newfound clarity. Using intuition is not the same as forgetting about the decision or refusing to make it. Intuitive decisions usually arrive after we’ve gathered the relevant facts and faced a problem for some time.

Evaluate your decision. Hindsight is a source of insight. After you act on a decision, observe the consequences over time. Reflect on how well your decision worked and what you might have done differently.

Think of choices. This final suggestion involves some creative thinking. Consider that the word decide derives from the same roots as suicide and homicide . In the spirit of those words, a decision forever “kills” all other options. That’s kind of heavy. Instead, use the word choice , and see whether it frees up your thinking. When you choose , you express a preference for one option over others. However, those options remain live possibilities for the future. Choose for today, knowing that as you gain more wisdom and experience, you can choose again.

Using Critical Thinking Skills in Problem Solving

Think of problem solving as a process with four Ps : Define the problem , generate possibilities ,

create a plan , and perform your plan.

Step 1: Define the problem. To define a problem effectively, understand what a problem is—a mismatch between what you want and what you have. Problem solving is all about reducing the gap between these two factors.

Tell the truth about what’s present in your life right now, without shame or blame. For example: “I often get sleepy while reading my physics assignments, and after closing the book I cannot remember what I just read.”

Next, describe in detail what you want. Go for specifics: “I want to remain alert as I read about physics. I also want to accurately summarize each chapter I read.”

Remember that when we define a problem in limiting ways, our solutions merely generate new problems. As Albert Einstein said, “The world we have made is a result of the level of thinking we have done thus far. We cannot solve problems at the same level at which we created them” (Calaprice 2000).

This idea has many applications for success in school. An example is the student who struggles with note taking. The problem, she thinks, is that her notes are too sketchy. The logical solution, she decides, is to take more notes; her new goal is to write down almost everything her instructors say. No matter how fast and furiously she writes, she cannot capture all of the instructors’ comments.

Consider what happens when this student defines the problem in a new way. After more thought, she decides that her dilemma is not the quantity of her notes but their quality . She adopts a new format for taking notes, dividing her notepaper into two columns. In the right-hand column, she writes down only the main points of each lecture. In the left-hand column, she notes two or three supporting details for each point.

Over time, this student makes the joyous discovery that there are usually just three or four core ideas to remember from each lecture. She originally thought the solution was to take more notes. What really worked was taking notes in a new way.

Step 2: Generate possibilities. Now put on your creative thinking hat. Open up. Brainstorm as many possible solutions to the problem as you can. At this stage, quantity counts. As you generate possibilities, gather relevant facts. For example, when you’re faced with a dilemma about what courses to take next semester, get information on class times, locations, and instructors. If you haven’t decided which summer job offer to accept, gather information on salary, benefits, and working conditions.

Step 3: Create a plan. After rereading your problem definition and list of possible solutions, choose the solution that seems most workable. Think about specific actions that will reduce the gap between what you have and what you want. Visualize the steps you will take to make this solution a reality, and arrange them in chronological order. To make your plan even more powerful, put it in writing.

Step 4: Perform your plan. This step gets you off your chair and out into the world. Now you actually do what you have planned.

Ultimately, your skill in solving problems lies in how well you perform your plan. Through the quality of your actions, you become the architect of your own success.

When facing problems, experiment with these four Ps, and remember that the order of steps is not absolute. Also remember that any solution has the potential to create new problems. If that happens, cycle through the four Ps of problem solving again.

Critical Thinking Skills in Action: Thinking About Your Major, Part 1

One decision that troubles many students in higher education is the choice of a major. Weighing the benefits, costs, and outcomes of a possible major is an intellectual challenge. This choice is an opportunity to apply your critical thinking, decision-making, and problem-solving skills. The following suggestions will guide you through this seemingly overwhelming process.

The first step is to discover options. You can use the following suggestions to discover options for choosing your major:

Follow the fun. Perhaps you look forward to attending one of your classes and even like completing the assignments. This is a clue to your choice of major.

See whether you can find lasting patterns in the subjects and extracurricular activities that you’ve enjoyed over the years. Look for a major that allows you to continue and expand on these experiences.

Also, sit down with a stack of 3 × 5 cards and brainstorm answers to the following questions:

  • What do you enjoy doing most with your unscheduled time?
  • Imagine that you’re at a party and having a fascinating conversation. What is this conversation about?
  • What kind of problems do you enjoy solving—those that involve people? Products? Ideas?
  • What interests are revealed by your choices of reading material, television shows, and other entertainment?
  • What would an ideal day look like for you? Describe where you would live, who would be with you, and what you would do throughout the day. Do any of these visions suggest a possible major?

Questions like these can uncover a “fun factor” that energizes you to finish the work of completing a major.

Consider your abilities. In choosing a major, ability counts as much as interest. In addition to considering what you enjoy, think about times and places when you excelled. List the courses that you aced, the work assignments that you mastered, and the hobbies that led to rewards or recognition. Let your choice of a major reflect a discovery of your passions and potentials.

Use formal techniques for self-discovery. Explore questionnaires and inventories that are designed to correlate your interests with specific majors. Examples include the Strong Interest Inventory and the Self-Directed Search. Your academic advisor or someone in your school’s career planning office can give you more details about these and related assessments. For some fun, take several of them and meet with an advisor to interpret the results. Remember inventories can help you gain self-knowledge, and other people can offer valuable perspectives. However, what you do with all this input is entirely up to you.

Critical Thinking Skills in Action: Thinking About Your Major, Part 2

As you review the following additional suggestions of discovering options, think about what strategies you already use in your own decision-making process. Also think about what new strategies you might try in the future.

Link to long-term goals. Your choice of a major can fall into place once you determine what you want in life. Before you choose a major, back up to a bigger picture. List your core values, such as contributing to society, achieving financial security and professional recognition, enjoying good health, or making time for fun. Also write down specific goals that you want to accomplish 5 years, 10 years, or even 50 years from today.

Many students find that the prospect of getting what they want in life justifies all of the time, money, and day-to-day effort invested in going to school. Having a major gives you a powerful incentive for attending classes, taking part in discussions, reading textbooks, writing papers, and completing other assignments. When you see a clear connection between finishing school and creating the life of your dreams, the daily tasks of higher education become charged with meaning.

Ask other people. Key people in your life might have valuable suggestions about your choice of major. Ask for their ideas, and listen with an open mind. At the same time, distance yourself from any pressure to choose a major or career that fails to interest you. If you make a choice solely on the basis of the expectations of other people, you could end up with a major or even a career you don’t enjoy.

Gather information. Check your school’s catalog or website for a list of available majors. Here is a gold mine of information. Take a quick glance, and highlight all the majors that interest you. Then talk to students who have declared these majors. Also read the descriptions of courses required for these majors. Do you get excited about the chance to enroll in them? Pay attention to your gut feelings.

Also chat with instructors who teach courses in a specific major. Ask for copies of their class syllabi. Go to the bookstore and browse the required texts. Based on all of this information, write a list of prospective majors. Discuss them with an academic advisor and someone at your school’s career-planning center.

Invent a major. When choosing a major, you might not need to limit yourself to those listed in your school catalog. Many schools now have flexible programs that allow for independent study. Through such programs, you might be able to combine two existing majors or invent an entirely new one of your own.

Consider a complementary minor. You can add flexibility to your academic program by choosing a minor to complement or contrast with your major. The student who wants to be a minister could opt for a minor in English; all of those courses in composition can help in writing sermons. Or the student with a major in psychology might choose a minor in business administration, with the idea of managing a counseling service some day. An effective choice of a minor can expand your skills and career options.

Think critically about the link between your major and your career. Your career goals might have a significant impact on your choice of major.

You could pursue a rewarding career by choosing among several different majors. Even students planning to apply for law school or medical school have flexibility in their choice of majors. In addition, after graduation, many people tend to be employed in jobs that have little relationship to their major. And you might choose a career in the future that is unrelated to any currently available major.

Critical Thinking Skills in Action: Thinking About Your Major, Part 3

Once you have discovered all of your options, you can move on to the next step in the process— making a trial choice.

Make a Trial Choice

Pretend that you have to choose a major today. Based on the options for a major that you’ve already discovered, write down the first three ideas that come to mind. Review the list for a few minutes, and then choose one.

Evaluate Your Trial Choice

When you’ve made a trial choice of major, take on the role of a scientist. Treat your choice as a hypothesis, and then design a series of experiments to evaluate and test it. For example:

  • Schedule office meetings with instructors who teach courses in the major. Ask about required course work and career options in the field.
  • Discuss your trial choice with an academic advisor or career counselor.
  • Enroll in a course related to your possible major. Remember that introductory courses might not give you a realistic picture of the workload involved in advanced courses. Also, you might not be able to register for certain courses until you’ve actually declared a related major.
  • Find a volunteer experience, internship, part-time job, or service-learning experience related to the major.
  • Interview students who have declared the same major. Ask them in detail about their experiences and suggestions for success.
  • Interview people who work in a field related to the major and “shadow” them—that is, spend time with those people during their workday.
  • Think about whether you can complete your major given the amount of time and money that you plan to invest in higher education.
  • Consider whether declaring this major would require a transfer to another program or even another school.

If your “experiments” confirm your choice of major, celebrate that fact. If they result in choosing a new major, celebrate that outcome as well.

Also remember that higher education represents a safe place to test your choice of major—and to change your mind. As you sort through your options, help is always available from administrators, instructors, advisors, and peers.

Choose Again

Keep your choice of a major in perspective. There is probably no single “correct” choice. Your unique collection of skills is likely to provide the basis for majoring in several fields.

Odds are that you’ll change your major at least once—and that you’ll change careers several times during your life. One benefit of higher education is mobility. You gain the general skills and knowledge that can help you move into a new major or career field at any time.

Viewing a major as a one-time choice that determines your entire future can raise your stress levels. Instead, look at choosing a major as the start of a continuing path that involves discovery, choice, and passionate action.

As you review this example of how you can use critical thinking to make a decision about choosing your major, think about how you will use your critical thinking to make decisions and solve problems in the future.

SEP home page

  • Table of Contents
  • Random Entry
  • Chronological
  • Editorial Information
  • About the SEP
  • Editorial Board
  • How to Cite the SEP
  • Special Characters
  • Advanced Tools
  • Support the SEP
  • PDFs for SEP Friends
  • Make a Donation
  • SEPIA for Libraries
  • Entry Contents

Bibliography

Academic tools.

  • Friends PDF Preview
  • Author and Citation Info
  • Back to Top

Critical Thinking

Critical thinking is a widely accepted educational goal. Its definition is contested, but the competing definitions can be understood as differing conceptions of the same basic concept: careful thinking directed to a goal. Conceptions differ with respect to the scope of such thinking, the type of goal, the criteria and norms for thinking carefully, and the thinking components on which they focus. Its adoption as an educational goal has been recommended on the basis of respect for students’ autonomy and preparing students for success in life and for democratic citizenship. “Critical thinkers” have the dispositions and abilities that lead them to think critically when appropriate. The abilities can be identified directly; the dispositions indirectly, by considering what factors contribute to or impede exercise of the abilities. Standardized tests have been developed to assess the degree to which a person possesses such dispositions and abilities. Educational intervention has been shown experimentally to improve them, particularly when it includes dialogue, anchored instruction, and mentoring. Controversies have arisen over the generalizability of critical thinking across domains, over alleged bias in critical thinking theories and instruction, and over the relationship of critical thinking to other types of thinking.

2.1 Dewey’s Three Main Examples

2.2 dewey’s other examples, 2.3 further examples, 2.4 non-examples, 3. the definition of critical thinking, 4. its value, 5. the process of thinking critically, 6. components of the process, 7. contributory dispositions and abilities, 8.1 initiating dispositions, 8.2 internal dispositions, 9. critical thinking abilities, 10. required knowledge, 11. educational methods, 12.1 the generalizability of critical thinking, 12.2 bias in critical thinking theory and pedagogy, 12.3 relationship of critical thinking to other types of thinking, other internet resources, related entries.

Use of the term ‘critical thinking’ to describe an educational goal goes back to the American philosopher John Dewey (1910), who more commonly called it ‘reflective thinking’. He defined it as

active, persistent and careful consideration of any belief or supposed form of knowledge in the light of the grounds that support it, and the further conclusions to which it tends. (Dewey 1910: 6; 1933: 9)

and identified a habit of such consideration with a scientific attitude of mind. His lengthy quotations of Francis Bacon, John Locke, and John Stuart Mill indicate that he was not the first person to propose development of a scientific attitude of mind as an educational goal.

In the 1930s, many of the schools that participated in the Eight-Year Study of the Progressive Education Association (Aikin 1942) adopted critical thinking as an educational goal, for whose achievement the study’s Evaluation Staff developed tests (Smith, Tyler, & Evaluation Staff 1942). Glaser (1941) showed experimentally that it was possible to improve the critical thinking of high school students. Bloom’s influential taxonomy of cognitive educational objectives (Bloom et al. 1956) incorporated critical thinking abilities. Ennis (1962) proposed 12 aspects of critical thinking as a basis for research on the teaching and evaluation of critical thinking ability.

Since 1980, an annual international conference in California on critical thinking and educational reform has attracted tens of thousands of educators from all levels of education and from many parts of the world. Also since 1980, the state university system in California has required all undergraduate students to take a critical thinking course. Since 1983, the Association for Informal Logic and Critical Thinking has sponsored sessions in conjunction with the divisional meetings of the American Philosophical Association (APA). In 1987, the APA’s Committee on Pre-College Philosophy commissioned a consensus statement on critical thinking for purposes of educational assessment and instruction (Facione 1990a). Researchers have developed standardized tests of critical thinking abilities and dispositions; for details, see the Supplement on Assessment . Educational jurisdictions around the world now include critical thinking in guidelines for curriculum and assessment.

For details on this history, see the Supplement on History .

2. Examples and Non-Examples

Before considering the definition of critical thinking, it will be helpful to have in mind some examples of critical thinking, as well as some examples of kinds of thinking that would apparently not count as critical thinking.

Dewey (1910: 68–71; 1933: 91–94) takes as paradigms of reflective thinking three class papers of students in which they describe their thinking. The examples range from the everyday to the scientific.

Transit : “The other day, when I was down town on 16th Street, a clock caught my eye. I saw that the hands pointed to 12:20. This suggested that I had an engagement at 124th Street, at one o’clock. I reasoned that as it had taken me an hour to come down on a surface car, I should probably be twenty minutes late if I returned the same way. I might save twenty minutes by a subway express. But was there a station near? If not, I might lose more than twenty minutes in looking for one. Then I thought of the elevated, and I saw there was such a line within two blocks. But where was the station? If it were several blocks above or below the street I was on, I should lose time instead of gaining it. My mind went back to the subway express as quicker than the elevated; furthermore, I remembered that it went nearer than the elevated to the part of 124th Street I wished to reach, so that time would be saved at the end of the journey. I concluded in favor of the subway, and reached my destination by one o’clock.” (Dewey 1910: 68–69; 1933: 91–92)

Ferryboat : “Projecting nearly horizontally from the upper deck of the ferryboat on which I daily cross the river is a long white pole, having a gilded ball at its tip. It suggested a flagpole when I first saw it; its color, shape, and gilded ball agreed with this idea, and these reasons seemed to justify me in this belief. But soon difficulties presented themselves. The pole was nearly horizontal, an unusual position for a flagpole; in the next place, there was no pulley, ring, or cord by which to attach a flag; finally, there were elsewhere on the boat two vertical staffs from which flags were occasionally flown. It seemed probable that the pole was not there for flag-flying.

“I then tried to imagine all possible purposes of the pole, and to consider for which of these it was best suited: (a) Possibly it was an ornament. But as all the ferryboats and even the tugboats carried poles, this hypothesis was rejected. (b) Possibly it was the terminal of a wireless telegraph. But the same considerations made this improbable. Besides, the more natural place for such a terminal would be the highest part of the boat, on top of the pilot house. (c) Its purpose might be to point out the direction in which the boat is moving.

“In support of this conclusion, I discovered that the pole was lower than the pilot house, so that the steersman could easily see it. Moreover, the tip was enough higher than the base, so that, from the pilot’s position, it must appear to project far out in front of the boat. Moreover, the pilot being near the front of the boat, he would need some such guide as to its direction. Tugboats would also need poles for such a purpose. This hypothesis was so much more probable than the others that I accepted it. I formed the conclusion that the pole was set up for the purpose of showing the pilot the direction in which the boat pointed, to enable him to steer correctly.” (Dewey 1910: 69–70; 1933: 92–93)

Bubbles : “In washing tumblers in hot soapsuds and placing them mouth downward on a plate, bubbles appeared on the outside of the mouth of the tumblers and then went inside. Why? The presence of bubbles suggests air, which I note must come from inside the tumbler. I see that the soapy water on the plate prevents escape of the air save as it may be caught in bubbles. But why should air leave the tumbler? There was no substance entering to force it out. It must have expanded. It expands by increase of heat, or by decrease of pressure, or both. Could the air have become heated after the tumbler was taken from the hot suds? Clearly not the air that was already entangled in the water. If heated air was the cause, cold air must have entered in transferring the tumblers from the suds to the plate. I test to see if this supposition is true by taking several more tumblers out. Some I shake so as to make sure of entrapping cold air in them. Some I take out holding mouth downward in order to prevent cold air from entering. Bubbles appear on the outside of every one of the former and on none of the latter. I must be right in my inference. Air from the outside must have been expanded by the heat of the tumbler, which explains the appearance of the bubbles on the outside. But why do they then go inside? Cold contracts. The tumbler cooled and also the air inside it. Tension was removed, and hence bubbles appeared inside. To be sure of this, I test by placing a cup of ice on the tumbler while the bubbles are still forming outside. They soon reverse” (Dewey 1910: 70–71; 1933: 93–94).

Dewey (1910, 1933) sprinkles his book with other examples of critical thinking. We will refer to the following.

Weather : A man on a walk notices that it has suddenly become cool, thinks that it is probably going to rain, looks up and sees a dark cloud obscuring the sun, and quickens his steps (1910: 6–10; 1933: 9–13).

Disorder : A man finds his rooms on his return to them in disorder with his belongings thrown about, thinks at first of burglary as an explanation, then thinks of mischievous children as being an alternative explanation, then looks to see whether valuables are missing, and discovers that they are (1910: 82–83; 1933: 166–168).

Typhoid : A physician diagnosing a patient whose conspicuous symptoms suggest typhoid avoids drawing a conclusion until more data are gathered by questioning the patient and by making tests (1910: 85–86; 1933: 170).

Blur : A moving blur catches our eye in the distance, we ask ourselves whether it is a cloud of whirling dust or a tree moving its branches or a man signaling to us, we think of other traits that should be found on each of those possibilities, and we look and see if those traits are found (1910: 102, 108; 1933: 121, 133).

Suction pump : In thinking about the suction pump, the scientist first notes that it will draw water only to a maximum height of 33 feet at sea level and to a lesser maximum height at higher elevations, selects for attention the differing atmospheric pressure at these elevations, sets up experiments in which the air is removed from a vessel containing water (when suction no longer works) and in which the weight of air at various levels is calculated, compares the results of reasoning about the height to which a given weight of air will allow a suction pump to raise water with the observed maximum height at different elevations, and finally assimilates the suction pump to such apparently different phenomena as the siphon and the rising of a balloon (1910: 150–153; 1933: 195–198).

Diamond : A passenger in a car driving in a diamond lane reserved for vehicles with at least one passenger notices that the diamond marks on the pavement are far apart in some places and close together in others. Why? The driver suggests that the reason may be that the diamond marks are not needed where there is a solid double line separating the diamond lane from the adjoining lane, but are needed when there is a dotted single line permitting crossing into the diamond lane. Further observation confirms that the diamonds are close together when a dotted line separates the diamond lane from its neighbour, but otherwise far apart.

Rash : A woman suddenly develops a very itchy red rash on her throat and upper chest. She recently noticed a mark on the back of her right hand, but was not sure whether the mark was a rash or a scrape. She lies down in bed and thinks about what might be causing the rash and what to do about it. About two weeks before, she began taking blood pressure medication that contained a sulfa drug, and the pharmacist had warned her, in view of a previous allergic reaction to a medication containing a sulfa drug, to be on the alert for an allergic reaction; however, she had been taking the medication for two weeks with no such effect. The day before, she began using a new cream on her neck and upper chest; against the new cream as the cause was mark on the back of her hand, which had not been exposed to the cream. She began taking probiotics about a month before. She also recently started new eye drops, but she supposed that manufacturers of eye drops would be careful not to include allergy-causing components in the medication. The rash might be a heat rash, since she recently was sweating profusely from her upper body. Since she is about to go away on a short vacation, where she would not have access to her usual physician, she decides to keep taking the probiotics and using the new eye drops but to discontinue the blood pressure medication and to switch back to the old cream for her neck and upper chest. She forms a plan to consult her regular physician on her return about the blood pressure medication.

Candidate : Although Dewey included no examples of thinking directed at appraising the arguments of others, such thinking has come to be considered a kind of critical thinking. We find an example of such thinking in the performance task on the Collegiate Learning Assessment (CLA+), which its sponsoring organization describes as

a performance-based assessment that provides a measure of an institution’s contribution to the development of critical-thinking and written communication skills of its students. (Council for Aid to Education 2017)

A sample task posted on its website requires the test-taker to write a report for public distribution evaluating a fictional candidate’s policy proposals and their supporting arguments, using supplied background documents, with a recommendation on whether to endorse the candidate.

Immediate acceptance of an idea that suggests itself as a solution to a problem (e.g., a possible explanation of an event or phenomenon, an action that seems likely to produce a desired result) is “uncritical thinking, the minimum of reflection” (Dewey 1910: 13). On-going suspension of judgment in the light of doubt about a possible solution is not critical thinking (Dewey 1910: 108). Critique driven by a dogmatically held political or religious ideology is not critical thinking; thus Paulo Freire (1968 [1970]) is using the term (e.g., at 1970: 71, 81, 100, 146) in a more politically freighted sense that includes not only reflection but also revolutionary action against oppression. Derivation of a conclusion from given data using an algorithm is not critical thinking.

What is critical thinking? There are many definitions. Ennis (2016) lists 14 philosophically oriented scholarly definitions and three dictionary definitions. Following Rawls (1971), who distinguished his conception of justice from a utilitarian conception but regarded them as rival conceptions of the same concept, Ennis maintains that the 17 definitions are different conceptions of the same concept. Rawls articulated the shared concept of justice as

a characteristic set of principles for assigning basic rights and duties and for determining… the proper distribution of the benefits and burdens of social cooperation. (Rawls 1971: 5)

Bailin et al. (1999b) claim that, if one considers what sorts of thinking an educator would take not to be critical thinking and what sorts to be critical thinking, one can conclude that educators typically understand critical thinking to have at least three features.

  • It is done for the purpose of making up one’s mind about what to believe or do.
  • The person engaging in the thinking is trying to fulfill standards of adequacy and accuracy appropriate to the thinking.
  • The thinking fulfills the relevant standards to some threshold level.

One could sum up the core concept that involves these three features by saying that critical thinking is careful goal-directed thinking. This core concept seems to apply to all the examples of critical thinking described in the previous section. As for the non-examples, their exclusion depends on construing careful thinking as excluding jumping immediately to conclusions, suspending judgment no matter how strong the evidence, reasoning from an unquestioned ideological or religious perspective, and routinely using an algorithm to answer a question.

If the core of critical thinking is careful goal-directed thinking, conceptions of it can vary according to its presumed scope, its presumed goal, one’s criteria and threshold for being careful, and the thinking component on which one focuses. As to its scope, some conceptions (e.g., Dewey 1910, 1933) restrict it to constructive thinking on the basis of one’s own observations and experiments, others (e.g., Ennis 1962; Fisher & Scriven 1997; Johnson 1992) to appraisal of the products of such thinking. Ennis (1991) and Bailin et al. (1999b) take it to cover both construction and appraisal. As to its goal, some conceptions restrict it to forming a judgment (Dewey 1910, 1933; Lipman 1987; Facione 1990a). Others allow for actions as well as beliefs as the end point of a process of critical thinking (Ennis 1991; Bailin et al. 1999b). As to the criteria and threshold for being careful, definitions vary in the term used to indicate that critical thinking satisfies certain norms: “intellectually disciplined” (Scriven & Paul 1987), “reasonable” (Ennis 1991), “skillful” (Lipman 1987), “skilled” (Fisher & Scriven 1997), “careful” (Bailin & Battersby 2009). Some definitions specify these norms, referring variously to “consideration of any belief or supposed form of knowledge in the light of the grounds that support it and the further conclusions to which it tends” (Dewey 1910, 1933); “the methods of logical inquiry and reasoning” (Glaser 1941); “conceptualizing, applying, analyzing, synthesizing, and/or evaluating information gathered from, or generated by, observation, experience, reflection, reasoning, or communication” (Scriven & Paul 1987); the requirement that “it is sensitive to context, relies on criteria, and is self-correcting” (Lipman 1987); “evidential, conceptual, methodological, criteriological, or contextual considerations” (Facione 1990a); and “plus-minus considerations of the product in terms of appropriate standards (or criteria)” (Johnson 1992). Stanovich and Stanovich (2010) propose to ground the concept of critical thinking in the concept of rationality, which they understand as combining epistemic rationality (fitting one’s beliefs to the world) and instrumental rationality (optimizing goal fulfillment); a critical thinker, in their view, is someone with “a propensity to override suboptimal responses from the autonomous mind” (2010: 227). These variant specifications of norms for critical thinking are not necessarily incompatible with one another, and in any case presuppose the core notion of thinking carefully. As to the thinking component singled out, some definitions focus on suspension of judgment during the thinking (Dewey 1910; McPeck 1981), others on inquiry while judgment is suspended (Bailin & Battersby 2009, 2021), others on the resulting judgment (Facione 1990a), and still others on responsiveness to reasons (Siegel 1988). Kuhn (2019) takes critical thinking to be more a dialogic practice of advancing and responding to arguments than an individual ability.

In educational contexts, a definition of critical thinking is a “programmatic definition” (Scheffler 1960: 19). It expresses a practical program for achieving an educational goal. For this purpose, a one-sentence formulaic definition is much less useful than articulation of a critical thinking process, with criteria and standards for the kinds of thinking that the process may involve. The real educational goal is recognition, adoption and implementation by students of those criteria and standards. That adoption and implementation in turn consists in acquiring the knowledge, abilities and dispositions of a critical thinker.

Conceptions of critical thinking generally do not include moral integrity as part of the concept. Dewey, for example, took critical thinking to be the ultimate intellectual goal of education, but distinguished it from the development of social cooperation among school children, which he took to be the central moral goal. Ennis (1996, 2011) added to his previous list of critical thinking dispositions a group of dispositions to care about the dignity and worth of every person, which he described as a “correlative” (1996) disposition without which critical thinking would be less valuable and perhaps harmful. An educational program that aimed at developing critical thinking but not the correlative disposition to care about the dignity and worth of every person, he asserted, “would be deficient and perhaps dangerous” (Ennis 1996: 172).

Dewey thought that education for reflective thinking would be of value to both the individual and society; recognition in educational practice of the kinship to the scientific attitude of children’s native curiosity, fertile imagination and love of experimental inquiry “would make for individual happiness and the reduction of social waste” (Dewey 1910: iii). Schools participating in the Eight-Year Study took development of the habit of reflective thinking and skill in solving problems as a means to leading young people to understand, appreciate and live the democratic way of life characteristic of the United States (Aikin 1942: 17–18, 81). Harvey Siegel (1988: 55–61) has offered four considerations in support of adopting critical thinking as an educational ideal. (1) Respect for persons requires that schools and teachers honour students’ demands for reasons and explanations, deal with students honestly, and recognize the need to confront students’ independent judgment; these requirements concern the manner in which teachers treat students. (2) Education has the task of preparing children to be successful adults, a task that requires development of their self-sufficiency. (3) Education should initiate children into the rational traditions in such fields as history, science and mathematics. (4) Education should prepare children to become democratic citizens, which requires reasoned procedures and critical talents and attitudes. To supplement these considerations, Siegel (1988: 62–90) responds to two objections: the ideology objection that adoption of any educational ideal requires a prior ideological commitment and the indoctrination objection that cultivation of critical thinking cannot escape being a form of indoctrination.

Despite the diversity of our 11 examples, one can recognize a common pattern. Dewey analyzed it as consisting of five phases:

  • suggestions , in which the mind leaps forward to a possible solution;
  • an intellectualization of the difficulty or perplexity into a problem to be solved, a question for which the answer must be sought;
  • the use of one suggestion after another as a leading idea, or hypothesis , to initiate and guide observation and other operations in collection of factual material;
  • the mental elaboration of the idea or supposition as an idea or supposition ( reasoning , in the sense on which reasoning is a part, not the whole, of inference); and
  • testing the hypothesis by overt or imaginative action. (Dewey 1933: 106–107; italics in original)

The process of reflective thinking consisting of these phases would be preceded by a perplexed, troubled or confused situation and followed by a cleared-up, unified, resolved situation (Dewey 1933: 106). The term ‘phases’ replaced the term ‘steps’ (Dewey 1910: 72), thus removing the earlier suggestion of an invariant sequence. Variants of the above analysis appeared in (Dewey 1916: 177) and (Dewey 1938: 101–119).

The variant formulations indicate the difficulty of giving a single logical analysis of such a varied process. The process of critical thinking may have a spiral pattern, with the problem being redefined in the light of obstacles to solving it as originally formulated. For example, the person in Transit might have concluded that getting to the appointment at the scheduled time was impossible and have reformulated the problem as that of rescheduling the appointment for a mutually convenient time. Further, defining a problem does not always follow after or lead immediately to an idea of a suggested solution. Nor should it do so, as Dewey himself recognized in describing the physician in Typhoid as avoiding any strong preference for this or that conclusion before getting further information (Dewey 1910: 85; 1933: 170). People with a hypothesis in mind, even one to which they have a very weak commitment, have a so-called “confirmation bias” (Nickerson 1998): they are likely to pay attention to evidence that confirms the hypothesis and to ignore evidence that counts against it or for some competing hypothesis. Detectives, intelligence agencies, and investigators of airplane accidents are well advised to gather relevant evidence systematically and to postpone even tentative adoption of an explanatory hypothesis until the collected evidence rules out with the appropriate degree of certainty all but one explanation. Dewey’s analysis of the critical thinking process can be faulted as well for requiring acceptance or rejection of a possible solution to a defined problem, with no allowance for deciding in the light of the available evidence to suspend judgment. Further, given the great variety of kinds of problems for which reflection is appropriate, there is likely to be variation in its component events. Perhaps the best way to conceptualize the critical thinking process is as a checklist whose component events can occur in a variety of orders, selectively, and more than once. These component events might include (1) noticing a difficulty, (2) defining the problem, (3) dividing the problem into manageable sub-problems, (4) formulating a variety of possible solutions to the problem or sub-problem, (5) determining what evidence is relevant to deciding among possible solutions to the problem or sub-problem, (6) devising a plan of systematic observation or experiment that will uncover the relevant evidence, (7) carrying out the plan of systematic observation or experimentation, (8) noting the results of the systematic observation or experiment, (9) gathering relevant testimony and information from others, (10) judging the credibility of testimony and information gathered from others, (11) drawing conclusions from gathered evidence and accepted testimony, and (12) accepting a solution that the evidence adequately supports (cf. Hitchcock 2017: 485).

Checklist conceptions of the process of critical thinking are open to the objection that they are too mechanical and procedural to fit the multi-dimensional and emotionally charged issues for which critical thinking is urgently needed (Paul 1984). For such issues, a more dialectical process is advocated, in which competing relevant world views are identified, their implications explored, and some sort of creative synthesis attempted.

If one considers the critical thinking process illustrated by the 11 examples, one can identify distinct kinds of mental acts and mental states that form part of it. To distinguish, label and briefly characterize these components is a useful preliminary to identifying abilities, skills, dispositions, attitudes, habits and the like that contribute causally to thinking critically. Identifying such abilities and habits is in turn a useful preliminary to setting educational goals. Setting the goals is in its turn a useful preliminary to designing strategies for helping learners to achieve the goals and to designing ways of measuring the extent to which learners have done so. Such measures provide both feedback to learners on their achievement and a basis for experimental research on the effectiveness of various strategies for educating people to think critically. Let us begin, then, by distinguishing the kinds of mental acts and mental events that can occur in a critical thinking process.

  • Observing : One notices something in one’s immediate environment (sudden cooling of temperature in Weather , bubbles forming outside a glass and then going inside in Bubbles , a moving blur in the distance in Blur , a rash in Rash ). Or one notes the results of an experiment or systematic observation (valuables missing in Disorder , no suction without air pressure in Suction pump )
  • Feeling : One feels puzzled or uncertain about something (how to get to an appointment on time in Transit , why the diamonds vary in spacing in Diamond ). One wants to resolve this perplexity. One feels satisfaction once one has worked out an answer (to take the subway express in Transit , diamonds closer when needed as a warning in Diamond ).
  • Wondering : One formulates a question to be addressed (why bubbles form outside a tumbler taken from hot water in Bubbles , how suction pumps work in Suction pump , what caused the rash in Rash ).
  • Imagining : One thinks of possible answers (bus or subway or elevated in Transit , flagpole or ornament or wireless communication aid or direction indicator in Ferryboat , allergic reaction or heat rash in Rash ).
  • Inferring : One works out what would be the case if a possible answer were assumed (valuables missing if there has been a burglary in Disorder , earlier start to the rash if it is an allergic reaction to a sulfa drug in Rash ). Or one draws a conclusion once sufficient relevant evidence is gathered (take the subway in Transit , burglary in Disorder , discontinue blood pressure medication and new cream in Rash ).
  • Knowledge : One uses stored knowledge of the subject-matter to generate possible answers or to infer what would be expected on the assumption of a particular answer (knowledge of a city’s public transit system in Transit , of the requirements for a flagpole in Ferryboat , of Boyle’s law in Bubbles , of allergic reactions in Rash ).
  • Experimenting : One designs and carries out an experiment or a systematic observation to find out whether the results deduced from a possible answer will occur (looking at the location of the flagpole in relation to the pilot’s position in Ferryboat , putting an ice cube on top of a tumbler taken from hot water in Bubbles , measuring the height to which a suction pump will draw water at different elevations in Suction pump , noticing the spacing of diamonds when movement to or from a diamond lane is allowed in Diamond ).
  • Consulting : One finds a source of information, gets the information from the source, and makes a judgment on whether to accept it. None of our 11 examples include searching for sources of information. In this respect they are unrepresentative, since most people nowadays have almost instant access to information relevant to answering any question, including many of those illustrated by the examples. However, Candidate includes the activities of extracting information from sources and evaluating its credibility.
  • Identifying and analyzing arguments : One notices an argument and works out its structure and content as a preliminary to evaluating its strength. This activity is central to Candidate . It is an important part of a critical thinking process in which one surveys arguments for various positions on an issue.
  • Judging : One makes a judgment on the basis of accumulated evidence and reasoning, such as the judgment in Ferryboat that the purpose of the pole is to provide direction to the pilot.
  • Deciding : One makes a decision on what to do or on what policy to adopt, as in the decision in Transit to take the subway.

By definition, a person who does something voluntarily is both willing and able to do that thing at that time. Both the willingness and the ability contribute causally to the person’s action, in the sense that the voluntary action would not occur if either (or both) of these were lacking. For example, suppose that one is standing with one’s arms at one’s sides and one voluntarily lifts one’s right arm to an extended horizontal position. One would not do so if one were unable to lift one’s arm, if for example one’s right side was paralyzed as the result of a stroke. Nor would one do so if one were unwilling to lift one’s arm, if for example one were participating in a street demonstration at which a white supremacist was urging the crowd to lift their right arm in a Nazi salute and one were unwilling to express support in this way for the racist Nazi ideology. The same analysis applies to a voluntary mental process of thinking critically. It requires both willingness and ability to think critically, including willingness and ability to perform each of the mental acts that compose the process and to coordinate those acts in a sequence that is directed at resolving the initiating perplexity.

Consider willingness first. We can identify causal contributors to willingness to think critically by considering factors that would cause a person who was able to think critically about an issue nevertheless not to do so (Hamby 2014). For each factor, the opposite condition thus contributes causally to willingness to think critically on a particular occasion. For example, people who habitually jump to conclusions without considering alternatives will not think critically about issues that arise, even if they have the required abilities. The contrary condition of willingness to suspend judgment is thus a causal contributor to thinking critically.

Now consider ability. In contrast to the ability to move one’s arm, which can be completely absent because a stroke has left the arm paralyzed, the ability to think critically is a developed ability, whose absence is not a complete absence of ability to think but absence of ability to think well. We can identify the ability to think well directly, in terms of the norms and standards for good thinking. In general, to be able do well the thinking activities that can be components of a critical thinking process, one needs to know the concepts and principles that characterize their good performance, to recognize in particular cases that the concepts and principles apply, and to apply them. The knowledge, recognition and application may be procedural rather than declarative. It may be domain-specific rather than widely applicable, and in either case may need subject-matter knowledge, sometimes of a deep kind.

Reflections of the sort illustrated by the previous two paragraphs have led scholars to identify the knowledge, abilities and dispositions of a “critical thinker”, i.e., someone who thinks critically whenever it is appropriate to do so. We turn now to these three types of causal contributors to thinking critically. We start with dispositions, since arguably these are the most powerful contributors to being a critical thinker, can be fostered at an early stage of a child’s development, and are susceptible to general improvement (Glaser 1941: 175)

8. Critical Thinking Dispositions

Educational researchers use the term ‘dispositions’ broadly for the habits of mind and attitudes that contribute causally to being a critical thinker. Some writers (e.g., Paul & Elder 2006; Hamby 2014; Bailin & Battersby 2016a) propose to use the term ‘virtues’ for this dimension of a critical thinker. The virtues in question, although they are virtues of character, concern the person’s ways of thinking rather than the person’s ways of behaving towards others. They are not moral virtues but intellectual virtues, of the sort articulated by Zagzebski (1996) and discussed by Turri, Alfano, and Greco (2017).

On a realistic conception, thinking dispositions or intellectual virtues are real properties of thinkers. They are general tendencies, propensities, or inclinations to think in particular ways in particular circumstances, and can be genuinely explanatory (Siegel 1999). Sceptics argue that there is no evidence for a specific mental basis for the habits of mind that contribute to thinking critically, and that it is pedagogically misleading to posit such a basis (Bailin et al. 1999a). Whatever their status, critical thinking dispositions need motivation for their initial formation in a child—motivation that may be external or internal. As children develop, the force of habit will gradually become important in sustaining the disposition (Nieto & Valenzuela 2012). Mere force of habit, however, is unlikely to sustain critical thinking dispositions. Critical thinkers must value and enjoy using their knowledge and abilities to think things through for themselves. They must be committed to, and lovers of, inquiry.

A person may have a critical thinking disposition with respect to only some kinds of issues. For example, one could be open-minded about scientific issues but not about religious issues. Similarly, one could be confident in one’s ability to reason about the theological implications of the existence of evil in the world but not in one’s ability to reason about the best design for a guided ballistic missile.

Facione (1990a: 25) divides “affective dispositions” of critical thinking into approaches to life and living in general and approaches to specific issues, questions or problems. Adapting this distinction, one can usefully divide critical thinking dispositions into initiating dispositions (those that contribute causally to starting to think critically about an issue) and internal dispositions (those that contribute causally to doing a good job of thinking critically once one has started). The two categories are not mutually exclusive. For example, open-mindedness, in the sense of willingness to consider alternative points of view to one’s own, is both an initiating and an internal disposition.

Using the strategy of considering factors that would block people with the ability to think critically from doing so, we can identify as initiating dispositions for thinking critically attentiveness, a habit of inquiry, self-confidence, courage, open-mindedness, willingness to suspend judgment, trust in reason, wanting evidence for one’s beliefs, and seeking the truth. We consider briefly what each of these dispositions amounts to, in each case citing sources that acknowledge them.

  • Attentiveness : One will not think critically if one fails to recognize an issue that needs to be thought through. For example, the pedestrian in Weather would not have looked up if he had not noticed that the air was suddenly cooler. To be a critical thinker, then, one needs to be habitually attentive to one’s surroundings, noticing not only what one senses but also sources of perplexity in messages received and in one’s own beliefs and attitudes (Facione 1990a: 25; Facione, Facione, & Giancarlo 2001).
  • Habit of inquiry : Inquiry is effortful, and one needs an internal push to engage in it. For example, the student in Bubbles could easily have stopped at idle wondering about the cause of the bubbles rather than reasoning to a hypothesis, then designing and executing an experiment to test it. Thus willingness to think critically needs mental energy and initiative. What can supply that energy? Love of inquiry, or perhaps just a habit of inquiry. Hamby (2015) has argued that willingness to inquire is the central critical thinking virtue, one that encompasses all the others. It is recognized as a critical thinking disposition by Dewey (1910: 29; 1933: 35), Glaser (1941: 5), Ennis (1987: 12; 1991: 8), Facione (1990a: 25), Bailin et al. (1999b: 294), Halpern (1998: 452), and Facione, Facione, & Giancarlo (2001).
  • Self-confidence : Lack of confidence in one’s abilities can block critical thinking. For example, if the woman in Rash lacked confidence in her ability to figure things out for herself, she might just have assumed that the rash on her chest was the allergic reaction to her medication against which the pharmacist had warned her. Thus willingness to think critically requires confidence in one’s ability to inquire (Facione 1990a: 25; Facione, Facione, & Giancarlo 2001).
  • Courage : Fear of thinking for oneself can stop one from doing it. Thus willingness to think critically requires intellectual courage (Paul & Elder 2006: 16).
  • Open-mindedness : A dogmatic attitude will impede thinking critically. For example, a person who adheres rigidly to a “pro-choice” position on the issue of the legal status of induced abortion is likely to be unwilling to consider seriously the issue of when in its development an unborn child acquires a moral right to life. Thus willingness to think critically requires open-mindedness, in the sense of a willingness to examine questions to which one already accepts an answer but which further evidence or reasoning might cause one to answer differently (Dewey 1933; Facione 1990a; Ennis 1991; Bailin et al. 1999b; Halpern 1998, Facione, Facione, & Giancarlo 2001). Paul (1981) emphasizes open-mindedness about alternative world-views, and recommends a dialectical approach to integrating such views as central to what he calls “strong sense” critical thinking. In three studies, Haran, Ritov, & Mellers (2013) found that actively open-minded thinking, including “the tendency to weigh new evidence against a favored belief, to spend sufficient time on a problem before giving up, and to consider carefully the opinions of others in forming one’s own”, led study participants to acquire information and thus to make accurate estimations.
  • Willingness to suspend judgment : Premature closure on an initial solution will block critical thinking. Thus willingness to think critically requires a willingness to suspend judgment while alternatives are explored (Facione 1990a; Ennis 1991; Halpern 1998).
  • Trust in reason : Since distrust in the processes of reasoned inquiry will dissuade one from engaging in it, trust in them is an initiating critical thinking disposition (Facione 1990a, 25; Bailin et al. 1999b: 294; Facione, Facione, & Giancarlo 2001; Paul & Elder 2006). In reaction to an allegedly exclusive emphasis on reason in critical thinking theory and pedagogy, Thayer-Bacon (2000) argues that intuition, imagination, and emotion have important roles to play in an adequate conception of critical thinking that she calls “constructive thinking”. From her point of view, critical thinking requires trust not only in reason but also in intuition, imagination, and emotion.
  • Seeking the truth : If one does not care about the truth but is content to stick with one’s initial bias on an issue, then one will not think critically about it. Seeking the truth is thus an initiating critical thinking disposition (Bailin et al. 1999b: 294; Facione, Facione, & Giancarlo 2001). A disposition to seek the truth is implicit in more specific critical thinking dispositions, such as trying to be well-informed, considering seriously points of view other than one’s own, looking for alternatives, suspending judgment when the evidence is insufficient, and adopting a position when the evidence supporting it is sufficient.

Some of the initiating dispositions, such as open-mindedness and willingness to suspend judgment, are also internal critical thinking dispositions, in the sense of mental habits or attitudes that contribute causally to doing a good job of critical thinking once one starts the process. But there are many other internal critical thinking dispositions. Some of them are parasitic on one’s conception of good thinking. For example, it is constitutive of good thinking about an issue to formulate the issue clearly and to maintain focus on it. For this purpose, one needs not only the corresponding ability but also the corresponding disposition. Ennis (1991: 8) describes it as the disposition “to determine and maintain focus on the conclusion or question”, Facione (1990a: 25) as “clarity in stating the question or concern”. Other internal dispositions are motivators to continue or adjust the critical thinking process, such as willingness to persist in a complex task and willingness to abandon nonproductive strategies in an attempt to self-correct (Halpern 1998: 452). For a list of identified internal critical thinking dispositions, see the Supplement on Internal Critical Thinking Dispositions .

Some theorists postulate skills, i.e., acquired abilities, as operative in critical thinking. It is not obvious, however, that a good mental act is the exercise of a generic acquired skill. Inferring an expected time of arrival, as in Transit , has some generic components but also uses non-generic subject-matter knowledge. Bailin et al. (1999a) argue against viewing critical thinking skills as generic and discrete, on the ground that skilled performance at a critical thinking task cannot be separated from knowledge of concepts and from domain-specific principles of good thinking. Talk of skills, they concede, is unproblematic if it means merely that a person with critical thinking skills is capable of intelligent performance.

Despite such scepticism, theorists of critical thinking have listed as general contributors to critical thinking what they variously call abilities (Glaser 1941; Ennis 1962, 1991), skills (Facione 1990a; Halpern 1998) or competencies (Fisher & Scriven 1997). Amalgamating these lists would produce a confusing and chaotic cornucopia of more than 50 possible educational objectives, with only partial overlap among them. It makes sense instead to try to understand the reasons for the multiplicity and diversity, and to make a selection according to one’s own reasons for singling out abilities to be developed in a critical thinking curriculum. Two reasons for diversity among lists of critical thinking abilities are the underlying conception of critical thinking and the envisaged educational level. Appraisal-only conceptions, for example, involve a different suite of abilities than constructive-only conceptions. Some lists, such as those in (Glaser 1941), are put forward as educational objectives for secondary school students, whereas others are proposed as objectives for college students (e.g., Facione 1990a).

The abilities described in the remaining paragraphs of this section emerge from reflection on the general abilities needed to do well the thinking activities identified in section 6 as components of the critical thinking process described in section 5 . The derivation of each collection of abilities is accompanied by citation of sources that list such abilities and of standardized tests that claim to test them.

Observational abilities : Careful and accurate observation sometimes requires specialist expertise and practice, as in the case of observing birds and observing accident scenes. However, there are general abilities of noticing what one’s senses are picking up from one’s environment and of being able to articulate clearly and accurately to oneself and others what one has observed. It helps in exercising them to be able to recognize and take into account factors that make one’s observation less trustworthy, such as prior framing of the situation, inadequate time, deficient senses, poor observation conditions, and the like. It helps as well to be skilled at taking steps to make one’s observation more trustworthy, such as moving closer to get a better look, measuring something three times and taking the average, and checking what one thinks one is observing with someone else who is in a good position to observe it. It also helps to be skilled at recognizing respects in which one’s report of one’s observation involves inference rather than direct observation, so that one can then consider whether the inference is justified. These abilities come into play as well when one thinks about whether and with what degree of confidence to accept an observation report, for example in the study of history or in a criminal investigation or in assessing news reports. Observational abilities show up in some lists of critical thinking abilities (Ennis 1962: 90; Facione 1990a: 16; Ennis 1991: 9). There are items testing a person’s ability to judge the credibility of observation reports in the Cornell Critical Thinking Tests, Levels X and Z (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005). Norris and King (1983, 1985, 1990a, 1990b) is a test of ability to appraise observation reports.

Emotional abilities : The emotions that drive a critical thinking process are perplexity or puzzlement, a wish to resolve it, and satisfaction at achieving the desired resolution. Children experience these emotions at an early age, without being trained to do so. Education that takes critical thinking as a goal needs only to channel these emotions and to make sure not to stifle them. Collaborative critical thinking benefits from ability to recognize one’s own and others’ emotional commitments and reactions.

Questioning abilities : A critical thinking process needs transformation of an inchoate sense of perplexity into a clear question. Formulating a question well requires not building in questionable assumptions, not prejudging the issue, and using language that in context is unambiguous and precise enough (Ennis 1962: 97; 1991: 9).

Imaginative abilities : Thinking directed at finding the correct causal explanation of a general phenomenon or particular event requires an ability to imagine possible explanations. Thinking about what policy or plan of action to adopt requires generation of options and consideration of possible consequences of each option. Domain knowledge is required for such creative activity, but a general ability to imagine alternatives is helpful and can be nurtured so as to become easier, quicker, more extensive, and deeper (Dewey 1910: 34–39; 1933: 40–47). Facione (1990a) and Halpern (1998) include the ability to imagine alternatives as a critical thinking ability.

Inferential abilities : The ability to draw conclusions from given information, and to recognize with what degree of certainty one’s own or others’ conclusions follow, is universally recognized as a general critical thinking ability. All 11 examples in section 2 of this article include inferences, some from hypotheses or options (as in Transit , Ferryboat and Disorder ), others from something observed (as in Weather and Rash ). None of these inferences is formally valid. Rather, they are licensed by general, sometimes qualified substantive rules of inference (Toulmin 1958) that rest on domain knowledge—that a bus trip takes about the same time in each direction, that the terminal of a wireless telegraph would be located on the highest possible place, that sudden cooling is often followed by rain, that an allergic reaction to a sulfa drug generally shows up soon after one starts taking it. It is a matter of controversy to what extent the specialized ability to deduce conclusions from premisses using formal rules of inference is needed for critical thinking. Dewey (1933) locates logical forms in setting out the products of reflection rather than in the process of reflection. Ennis (1981a), on the other hand, maintains that a liberally-educated person should have the following abilities: to translate natural-language statements into statements using the standard logical operators, to use appropriately the language of necessary and sufficient conditions, to deal with argument forms and arguments containing symbols, to determine whether in virtue of an argument’s form its conclusion follows necessarily from its premisses, to reason with logically complex propositions, and to apply the rules and procedures of deductive logic. Inferential abilities are recognized as critical thinking abilities by Glaser (1941: 6), Facione (1990a: 9), Ennis (1991: 9), Fisher & Scriven (1997: 99, 111), and Halpern (1998: 452). Items testing inferential abilities constitute two of the five subtests of the Watson Glaser Critical Thinking Appraisal (Watson & Glaser 1980a, 1980b, 1994), two of the four sections in the Cornell Critical Thinking Test Level X (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005), three of the seven sections in the Cornell Critical Thinking Test Level Z (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005), 11 of the 34 items on Forms A and B of the California Critical Thinking Skills Test (Facione 1990b, 1992), and a high but variable proportion of the 25 selected-response questions in the Collegiate Learning Assessment (Council for Aid to Education 2017).

Experimenting abilities : Knowing how to design and execute an experiment is important not just in scientific research but also in everyday life, as in Rash . Dewey devoted a whole chapter of his How We Think (1910: 145–156; 1933: 190–202) to the superiority of experimentation over observation in advancing knowledge. Experimenting abilities come into play at one remove in appraising reports of scientific studies. Skill in designing and executing experiments includes the acknowledged abilities to appraise evidence (Glaser 1941: 6), to carry out experiments and to apply appropriate statistical inference techniques (Facione 1990a: 9), to judge inductions to an explanatory hypothesis (Ennis 1991: 9), and to recognize the need for an adequately large sample size (Halpern 1998). The Cornell Critical Thinking Test Level Z (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005) includes four items (out of 52) on experimental design. The Collegiate Learning Assessment (Council for Aid to Education 2017) makes room for appraisal of study design in both its performance task and its selected-response questions.

Consulting abilities : Skill at consulting sources of information comes into play when one seeks information to help resolve a problem, as in Candidate . Ability to find and appraise information includes ability to gather and marshal pertinent information (Glaser 1941: 6), to judge whether a statement made by an alleged authority is acceptable (Ennis 1962: 84), to plan a search for desired information (Facione 1990a: 9), and to judge the credibility of a source (Ennis 1991: 9). Ability to judge the credibility of statements is tested by 24 items (out of 76) in the Cornell Critical Thinking Test Level X (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005) and by four items (out of 52) in the Cornell Critical Thinking Test Level Z (Ennis & Millman 1971; Ennis, Millman, & Tomko 1985, 2005). The College Learning Assessment’s performance task requires evaluation of whether information in documents is credible or unreliable (Council for Aid to Education 2017).

Argument analysis abilities : The ability to identify and analyze arguments contributes to the process of surveying arguments on an issue in order to form one’s own reasoned judgment, as in Candidate . The ability to detect and analyze arguments is recognized as a critical thinking skill by Facione (1990a: 7–8), Ennis (1991: 9) and Halpern (1998). Five items (out of 34) on the California Critical Thinking Skills Test (Facione 1990b, 1992) test skill at argument analysis. The College Learning Assessment (Council for Aid to Education 2017) incorporates argument analysis in its selected-response tests of critical reading and evaluation and of critiquing an argument.

Judging skills and deciding skills : Skill at judging and deciding is skill at recognizing what judgment or decision the available evidence and argument supports, and with what degree of confidence. It is thus a component of the inferential skills already discussed.

Lists and tests of critical thinking abilities often include two more abilities: identifying assumptions and constructing and evaluating definitions.

In addition to dispositions and abilities, critical thinking needs knowledge: of critical thinking concepts, of critical thinking principles, and of the subject-matter of the thinking.

We can derive a short list of concepts whose understanding contributes to critical thinking from the critical thinking abilities described in the preceding section. Observational abilities require an understanding of the difference between observation and inference. Questioning abilities require an understanding of the concepts of ambiguity and vagueness. Inferential abilities require an understanding of the difference between conclusive and defeasible inference (traditionally, between deduction and induction), as well as of the difference between necessary and sufficient conditions. Experimenting abilities require an understanding of the concepts of hypothesis, null hypothesis, assumption and prediction, as well as of the concept of statistical significance and of its difference from importance. They also require an understanding of the difference between an experiment and an observational study, and in particular of the difference between a randomized controlled trial, a prospective correlational study and a retrospective (case-control) study. Argument analysis abilities require an understanding of the concepts of argument, premiss, assumption, conclusion and counter-consideration. Additional critical thinking concepts are proposed by Bailin et al. (1999b: 293), Fisher & Scriven (1997: 105–106), Black (2012), and Blair (2021).

According to Glaser (1941: 25), ability to think critically requires knowledge of the methods of logical inquiry and reasoning. If we review the list of abilities in the preceding section, however, we can see that some of them can be acquired and exercised merely through practice, possibly guided in an educational setting, followed by feedback. Searching intelligently for a causal explanation of some phenomenon or event requires that one consider a full range of possible causal contributors, but it seems more important that one implements this principle in one’s practice than that one is able to articulate it. What is important is “operational knowledge” of the standards and principles of good thinking (Bailin et al. 1999b: 291–293). But the development of such critical thinking abilities as designing an experiment or constructing an operational definition can benefit from learning their underlying theory. Further, explicit knowledge of quirks of human thinking seems useful as a cautionary guide. Human memory is not just fallible about details, as people learn from their own experiences of misremembering, but is so malleable that a detailed, clear and vivid recollection of an event can be a total fabrication (Loftus 2017). People seek or interpret evidence in ways that are partial to their existing beliefs and expectations, often unconscious of their “confirmation bias” (Nickerson 1998). Not only are people subject to this and other cognitive biases (Kahneman 2011), of which they are typically unaware, but it may be counter-productive for one to make oneself aware of them and try consciously to counteract them or to counteract social biases such as racial or sexual stereotypes (Kenyon & Beaulac 2014). It is helpful to be aware of these facts and of the superior effectiveness of blocking the operation of biases—for example, by making an immediate record of one’s observations, refraining from forming a preliminary explanatory hypothesis, blind refereeing, double-blind randomized trials, and blind grading of students’ work. It is also helpful to be aware of the prevalence of “noise” (unwanted unsystematic variability of judgments), of how to detect noise (through a noise audit), and of how to reduce noise: make accuracy the goal, think statistically, break a process of arriving at a judgment into independent tasks, resist premature intuitions, in a group get independent judgments first, favour comparative judgments and scales (Kahneman, Sibony, & Sunstein 2021). It is helpful as well to be aware of the concept of “bounded rationality” in decision-making and of the related distinction between “satisficing” and optimizing (Simon 1956; Gigerenzer 2001).

Critical thinking about an issue requires substantive knowledge of the domain to which the issue belongs. Critical thinking abilities are not a magic elixir that can be applied to any issue whatever by somebody who has no knowledge of the facts relevant to exploring that issue. For example, the student in Bubbles needed to know that gases do not penetrate solid objects like a glass, that air expands when heated, that the volume of an enclosed gas varies directly with its temperature and inversely with its pressure, and that hot objects will spontaneously cool down to the ambient temperature of their surroundings unless kept hot by insulation or a source of heat. Critical thinkers thus need a rich fund of subject-matter knowledge relevant to the variety of situations they encounter. This fact is recognized in the inclusion among critical thinking dispositions of a concern to become and remain generally well informed.

Experimental educational interventions, with control groups, have shown that education can improve critical thinking skills and dispositions, as measured by standardized tests. For information about these tests, see the Supplement on Assessment .

What educational methods are most effective at developing the dispositions, abilities and knowledge of a critical thinker? In a comprehensive meta-analysis of experimental and quasi-experimental studies of strategies for teaching students to think critically, Abrami et al. (2015) found that dialogue, anchored instruction, and mentoring each increased the effectiveness of the educational intervention, and that they were most effective when combined. They also found that in these studies a combination of separate instruction in critical thinking with subject-matter instruction in which students are encouraged to think critically was more effective than either by itself. However, the difference was not statistically significant; that is, it might have arisen by chance.

Most of these studies lack the longitudinal follow-up required to determine whether the observed differential improvements in critical thinking abilities or dispositions continue over time, for example until high school or college graduation. For details on studies of methods of developing critical thinking skills and dispositions, see the Supplement on Educational Methods .

12. Controversies

Scholars have denied the generalizability of critical thinking abilities across subject domains, have alleged bias in critical thinking theory and pedagogy, and have investigated the relationship of critical thinking to other kinds of thinking.

McPeck (1981) attacked the thinking skills movement of the 1970s, including the critical thinking movement. He argued that there are no general thinking skills, since thinking is always thinking about some subject-matter. It is futile, he claimed, for schools and colleges to teach thinking as if it were a separate subject. Rather, teachers should lead their pupils to become autonomous thinkers by teaching school subjects in a way that brings out their cognitive structure and that encourages and rewards discussion and argument. As some of his critics (e.g., Paul 1985; Siegel 1985) pointed out, McPeck’s central argument needs elaboration, since it has obvious counter-examples in writing and speaking, for which (up to a certain level of complexity) there are teachable general abilities even though they are always about some subject-matter. To make his argument convincing, McPeck needs to explain how thinking differs from writing and speaking in a way that does not permit useful abstraction of its components from the subject-matters with which it deals. He has not done so. Nevertheless, his position that the dispositions and abilities of a critical thinker are best developed in the context of subject-matter instruction is shared by many theorists of critical thinking, including Dewey (1910, 1933), Glaser (1941), Passmore (1980), Weinstein (1990), Bailin et al. (1999b), and Willingham (2019).

McPeck’s challenge prompted reflection on the extent to which critical thinking is subject-specific. McPeck argued for a strong subject-specificity thesis, according to which it is a conceptual truth that all critical thinking abilities are specific to a subject. (He did not however extend his subject-specificity thesis to critical thinking dispositions. In particular, he took the disposition to suspend judgment in situations of cognitive dissonance to be a general disposition.) Conceptual subject-specificity is subject to obvious counter-examples, such as the general ability to recognize confusion of necessary and sufficient conditions. A more modest thesis, also endorsed by McPeck, is epistemological subject-specificity, according to which the norms of good thinking vary from one field to another. Epistemological subject-specificity clearly holds to a certain extent; for example, the principles in accordance with which one solves a differential equation are quite different from the principles in accordance with which one determines whether a painting is a genuine Picasso. But the thesis suffers, as Ennis (1989) points out, from vagueness of the concept of a field or subject and from the obvious existence of inter-field principles, however broadly the concept of a field is construed. For example, the principles of hypothetico-deductive reasoning hold for all the varied fields in which such reasoning occurs. A third kind of subject-specificity is empirical subject-specificity, according to which as a matter of empirically observable fact a person with the abilities and dispositions of a critical thinker in one area of investigation will not necessarily have them in another area of investigation.

The thesis of empirical subject-specificity raises the general problem of transfer. If critical thinking abilities and dispositions have to be developed independently in each school subject, how are they of any use in dealing with the problems of everyday life and the political and social issues of contemporary society, most of which do not fit into the framework of a traditional school subject? Proponents of empirical subject-specificity tend to argue that transfer is more likely to occur if there is critical thinking instruction in a variety of domains, with explicit attention to dispositions and abilities that cut across domains. But evidence for this claim is scanty. There is a need for well-designed empirical studies that investigate the conditions that make transfer more likely.

It is common ground in debates about the generality or subject-specificity of critical thinking dispositions and abilities that critical thinking about any topic requires background knowledge about the topic. For example, the most sophisticated understanding of the principles of hypothetico-deductive reasoning is of no help unless accompanied by some knowledge of what might be plausible explanations of some phenomenon under investigation.

Critics have objected to bias in the theory, pedagogy and practice of critical thinking. Commentators (e.g., Alston 1995; Ennis 1998) have noted that anyone who takes a position has a bias in the neutral sense of being inclined in one direction rather than others. The critics, however, are objecting to bias in the pejorative sense of an unjustified favoring of certain ways of knowing over others, frequently alleging that the unjustly favoured ways are those of a dominant sex or culture (Bailin 1995). These ways favour:

  • reinforcement of egocentric and sociocentric biases over dialectical engagement with opposing world-views (Paul 1981, 1984; Warren 1998)
  • distancing from the object of inquiry over closeness to it (Martin 1992; Thayer-Bacon 1992)
  • indifference to the situation of others over care for them (Martin 1992)
  • orientation to thought over orientation to action (Martin 1992)
  • being reasonable over caring to understand people’s ideas (Thayer-Bacon 1993)
  • being neutral and objective over being embodied and situated (Thayer-Bacon 1995a)
  • doubting over believing (Thayer-Bacon 1995b)
  • reason over emotion, imagination and intuition (Thayer-Bacon 2000)
  • solitary thinking over collaborative thinking (Thayer-Bacon 2000)
  • written and spoken assignments over other forms of expression (Alston 2001)
  • attention to written and spoken communications over attention to human problems (Alston 2001)
  • winning debates in the public sphere over making and understanding meaning (Alston 2001)

A common thread in this smorgasbord of accusations is dissatisfaction with focusing on the logical analysis and evaluation of reasoning and arguments. While these authors acknowledge that such analysis and evaluation is part of critical thinking and should be part of its conceptualization and pedagogy, they insist that it is only a part. Paul (1981), for example, bemoans the tendency of atomistic teaching of methods of analyzing and evaluating arguments to turn students into more able sophists, adept at finding fault with positions and arguments with which they disagree but even more entrenched in the egocentric and sociocentric biases with which they began. Martin (1992) and Thayer-Bacon (1992) cite with approval the self-reported intimacy with their subject-matter of leading researchers in biology and medicine, an intimacy that conflicts with the distancing allegedly recommended in standard conceptions and pedagogy of critical thinking. Thayer-Bacon (2000) contrasts the embodied and socially embedded learning of her elementary school students in a Montessori school, who used their imagination, intuition and emotions as well as their reason, with conceptions of critical thinking as

thinking that is used to critique arguments, offer justifications, and make judgments about what are the good reasons, or the right answers. (Thayer-Bacon 2000: 127–128)

Alston (2001) reports that her students in a women’s studies class were able to see the flaws in the Cinderella myth that pervades much romantic fiction but in their own romantic relationships still acted as if all failures were the woman’s fault and still accepted the notions of love at first sight and living happily ever after. Students, she writes, should

be able to connect their intellectual critique to a more affective, somatic, and ethical account of making risky choices that have sexist, racist, classist, familial, sexual, or other consequences for themselves and those both near and far… critical thinking that reads arguments, texts, or practices merely on the surface without connections to feeling/desiring/doing or action lacks an ethical depth that should infuse the difference between mere cognitive activity and something we want to call critical thinking. (Alston 2001: 34)

Some critics portray such biases as unfair to women. Thayer-Bacon (1992), for example, has charged modern critical thinking theory with being sexist, on the ground that it separates the self from the object and causes one to lose touch with one’s inner voice, and thus stigmatizes women, who (she asserts) link self to object and listen to their inner voice. Her charge does not imply that women as a group are on average less able than men to analyze and evaluate arguments. Facione (1990c) found no difference by sex in performance on his California Critical Thinking Skills Test. Kuhn (1991: 280–281) found no difference by sex in either the disposition or the competence to engage in argumentative thinking.

The critics propose a variety of remedies for the biases that they allege. In general, they do not propose to eliminate or downplay critical thinking as an educational goal. Rather, they propose to conceptualize critical thinking differently and to change its pedagogy accordingly. Their pedagogical proposals arise logically from their objections. They can be summarized as follows:

  • Focus on argument networks with dialectical exchanges reflecting contesting points of view rather than on atomic arguments, so as to develop “strong sense” critical thinking that transcends egocentric and sociocentric biases (Paul 1981, 1984).
  • Foster closeness to the subject-matter and feeling connected to others in order to inform a humane democracy (Martin 1992).
  • Develop “constructive thinking” as a social activity in a community of physically embodied and socially embedded inquirers with personal voices who value not only reason but also imagination, intuition and emotion (Thayer-Bacon 2000).
  • In developing critical thinking in school subjects, treat as important neither skills nor dispositions but opening worlds of meaning (Alston 2001).
  • Attend to the development of critical thinking dispositions as well as skills, and adopt the “critical pedagogy” practised and advocated by Freire (1968 [1970]) and hooks (1994) (Dalgleish, Girard, & Davies 2017).

A common thread in these proposals is treatment of critical thinking as a social, interactive, personally engaged activity like that of a quilting bee or a barn-raising (Thayer-Bacon 2000) rather than as an individual, solitary, distanced activity symbolized by Rodin’s The Thinker . One can get a vivid description of education with the former type of goal from the writings of bell hooks (1994, 2010). Critical thinking for her is open-minded dialectical exchange across opposing standpoints and from multiple perspectives, a conception similar to Paul’s “strong sense” critical thinking (Paul 1981). She abandons the structure of domination in the traditional classroom. In an introductory course on black women writers, for example, she assigns students to write an autobiographical paragraph about an early racial memory, then to read it aloud as the others listen, thus affirming the uniqueness and value of each voice and creating a communal awareness of the diversity of the group’s experiences (hooks 1994: 84). Her “engaged pedagogy” is thus similar to the “freedom under guidance” implemented in John Dewey’s Laboratory School of Chicago in the late 1890s and early 1900s. It incorporates the dialogue, anchored instruction, and mentoring that Abrami (2015) found to be most effective in improving critical thinking skills and dispositions.

What is the relationship of critical thinking to problem solving, decision-making, higher-order thinking, creative thinking, and other recognized types of thinking? One’s answer to this question obviously depends on how one defines the terms used in the question. If critical thinking is conceived broadly to cover any careful thinking about any topic for any purpose, then problem solving and decision making will be kinds of critical thinking, if they are done carefully. Historically, ‘critical thinking’ and ‘problem solving’ were two names for the same thing. If critical thinking is conceived more narrowly as consisting solely of appraisal of intellectual products, then it will be disjoint with problem solving and decision making, which are constructive.

Bloom’s taxonomy of educational objectives used the phrase “intellectual abilities and skills” for what had been labeled “critical thinking” by some, “reflective thinking” by Dewey and others, and “problem solving” by still others (Bloom et al. 1956: 38). Thus, the so-called “higher-order thinking skills” at the taxonomy’s top levels of analysis, synthesis and evaluation are just critical thinking skills, although they do not come with general criteria for their assessment (Ennis 1981b). The revised version of Bloom’s taxonomy (Anderson et al. 2001) likewise treats critical thinking as cutting across those types of cognitive process that involve more than remembering (Anderson et al. 2001: 269–270). For details, see the Supplement on History .

As to creative thinking, it overlaps with critical thinking (Bailin 1987, 1988). Thinking about the explanation of some phenomenon or event, as in Ferryboat , requires creative imagination in constructing plausible explanatory hypotheses. Likewise, thinking about a policy question, as in Candidate , requires creativity in coming up with options. Conversely, creativity in any field needs to be balanced by critical appraisal of the draft painting or novel or mathematical theory.

  • Abrami, Philip C., Robert M. Bernard, Eugene Borokhovski, David I. Waddington, C. Anne Wade, and Tonje Person, 2015, “Strategies for Teaching Students to Think Critically: A Meta-analysis”, Review of Educational Research , 85(2): 275–314. doi:10.3102/0034654314551063
  • Aikin, Wilford M., 1942, The Story of the Eight-year Study, with Conclusions and Recommendations , Volume I of Adventure in American Education , New York and London: Harper & Brothers. [ Aikin 1942 available online ]
  • Alston, Kal, 1995, “Begging the Question: Is Critical Thinking Biased?”, Educational Theory , 45(2): 225–233. doi:10.1111/j.1741-5446.1995.00225.x
  • –––, 2001, “Re/Thinking Critical Thinking: The Seductions of Everyday Life”, Studies in Philosophy and Education , 20(1): 27–40. doi:10.1023/A:1005247128053
  • American Educational Research Association, 2014, Standards for Educational and Psychological Testing / American Educational Research Association, American Psychological Association, National Council on Measurement in Education , Washington, DC: American Educational Research Association.
  • Anderson, Lorin W., David R. Krathwohl, Peter W. Airiasian, Kathleen A. Cruikshank, Richard E. Mayer, Paul R. Pintrich, James Raths, and Merlin C. Wittrock, 2001, A Taxonomy for Learning, Teaching and Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives , New York: Longman, complete edition.
  • Bailin, Sharon, 1987, “Critical and Creative Thinking”, Informal Logic , 9(1): 23–30. [ Bailin 1987 available online ]
  • –––, 1988, Achieving Extraordinary Ends: An Essay on Creativity , Dordrecht: Kluwer. doi:10.1007/978-94-009-2780-3
  • –––, 1995, “Is Critical Thinking Biased? Clarifications and Implications”, Educational Theory , 45(2): 191–197. doi:10.1111/j.1741-5446.1995.00191.x
  • Bailin, Sharon and Mark Battersby, 2009, “Inquiry: A Dialectical Approach to Teaching Critical Thinking”, in Juho Ritola (ed.), Argument Cultures: Proceedings of OSSA 09 , CD-ROM (pp. 1–10), Windsor, ON: OSSA. [ Bailin & Battersby 2009 available online ]
  • –––, 2016a, “Fostering the Virtues of Inquiry”, Topoi , 35(2): 367–374. doi:10.1007/s11245-015-9307-6
  • –––, 2016b, Reason in the Balance: An Inquiry Approach to Critical Thinking , Indianapolis: Hackett, 2nd edition.
  • –––, 2021, “Inquiry: Teaching for Reasoned Judgment”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 31–46. doi: 10.1163/9789004444591_003
  • Bailin, Sharon, Roland Case, Jerrold R. Coombs, and Leroi B. Daniels, 1999a, “Common Misconceptions of Critical Thinking”, Journal of Curriculum Studies , 31(3): 269–283. doi:10.1080/002202799183124
  • –––, 1999b, “Conceptualizing Critical Thinking”, Journal of Curriculum Studies , 31(3): 285–302. doi:10.1080/002202799183133
  • Blair, J. Anthony, 2021, Studies in Critical Thinking , Windsor, ON: Windsor Studies in Argumentation, 2nd edition. [Available online at https://windsor.scholarsportal.info/omp/index.php/wsia/catalog/book/106]
  • Berman, Alan M., Seth J. Schwartz, William M. Kurtines, and Steven L. Berman, 2001, “The Process of Exploration in Identity Formation: The Role of Style and Competence”, Journal of Adolescence , 24(4): 513–528. doi:10.1006/jado.2001.0386
  • Black, Beth (ed.), 2012, An A to Z of Critical Thinking , London: Continuum International Publishing Group.
  • Bloom, Benjamin Samuel, Max D. Engelhart, Edward J. Furst, Walter H. Hill, and David R. Krathwohl, 1956, Taxonomy of Educational Objectives. Handbook I: Cognitive Domain , New York: David McKay.
  • Boardman, Frank, Nancy M. Cavender, and Howard Kahane, 2018, Logic and Contemporary Rhetoric: The Use of Reason in Everyday Life , Boston: Cengage, 13th edition.
  • Browne, M. Neil and Stuart M. Keeley, 2018, Asking the Right Questions: A Guide to Critical Thinking , Hoboken, NJ: Pearson, 12th edition.
  • Center for Assessment & Improvement of Learning, 2017, Critical Thinking Assessment Test , Cookeville, TN: Tennessee Technological University.
  • Cleghorn, Paul. 2021. “Critical Thinking in the Elementary School: Practical Guidance for Building a Culture of Thinking”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessmen t, Leiden: Brill, pp. 150–167. doi: 10.1163/9789004444591_010
  • Cohen, Jacob, 1988, Statistical Power Analysis for the Behavioral Sciences , Hillsdale, NJ: Lawrence Erlbaum Associates, 2nd edition.
  • College Board, 1983, Academic Preparation for College. What Students Need to Know and Be Able to Do , New York: College Entrance Examination Board, ERIC document ED232517.
  • Commission on the Relation of School and College of the Progressive Education Association, 1943, Thirty Schools Tell Their Story , Volume V of Adventure in American Education , New York and London: Harper & Brothers.
  • Council for Aid to Education, 2017, CLA+ Student Guide . Available at http://cae.org/images/uploads/pdf/CLA_Student_Guide_Institution.pdf ; last accessed 2022 07 16.
  • Dalgleish, Adam, Patrick Girard, and Maree Davies, 2017, “Critical Thinking, Bias and Feminist Philosophy: Building a Better Framework through Collaboration”, Informal Logic , 37(4): 351–369. [ Dalgleish et al. available online ]
  • Dewey, John, 1910, How We Think , Boston: D.C. Heath. [ Dewey 1910 available online ]
  • –––, 1916, Democracy and Education: An Introduction to the Philosophy of Education , New York: Macmillan.
  • –––, 1933, How We Think: A Restatement of the Relation of Reflective Thinking to the Educative Process , Lexington, MA: D.C. Heath.
  • –––, 1936, “The Theory of the Chicago Experiment”, Appendix II of Mayhew & Edwards 1936: 463–477.
  • –––, 1938, Logic: The Theory of Inquiry , New York: Henry Holt and Company.
  • Dominguez, Caroline (coord.), 2018a, A European Collection of the Critical Thinking Skills and Dispositions Needed in Different Professional Fields for the 21st Century , Vila Real, Portugal: UTAD. Available at http://bit.ly/CRITHINKEDUO1 ; last accessed 2022 07 16.
  • ––– (coord.), 2018b, A European Review on Critical Thinking Educational Practices in Higher Education Institutions , Vila Real: UTAD. Available at http://bit.ly/CRITHINKEDUO2 ; last accessed 2022 07 16.
  • ––– (coord.), 2018c, The CRITHINKEDU European Course on Critical Thinking Education for University Teachers: From Conception to Delivery , Vila Real: UTAD. Available at http:/bit.ly/CRITHINKEDU03; last accessed 2022 07 16.
  • Dominguez Caroline and Rita Payan-Carreira (eds.), 2019, Promoting Critical Thinking in European Higher Education Institutions: Towards an Educational Protocol , Vila Real: UTAD. Available at http:/bit.ly/CRITHINKEDU04; last accessed 2022 07 16.
  • Ennis, Robert H., 1958, “An Appraisal of the Watson-Glaser Critical Thinking Appraisal”, The Journal of Educational Research , 52(4): 155–158. doi:10.1080/00220671.1958.10882558
  • –––, 1962, “A Concept of Critical Thinking: A Proposed Basis for Research on the Teaching and Evaluation of Critical Thinking Ability”, Harvard Educational Review , 32(1): 81–111.
  • –––, 1981a, “A Conception of Deductive Logical Competence”, Teaching Philosophy , 4(3/4): 337–385. doi:10.5840/teachphil198143/429
  • –––, 1981b, “Eight Fallacies in Bloom’s Taxonomy”, in C. J. B. Macmillan (ed.), Philosophy of Education 1980: Proceedings of the Thirty-seventh Annual Meeting of the Philosophy of Education Society , Bloomington, IL: Philosophy of Education Society, pp. 269–273.
  • –––, 1984, “Problems in Testing Informal Logic, Critical Thinking, Reasoning Ability”, Informal Logic , 6(1): 3–9. [ Ennis 1984 available online ]
  • –––, 1987, “A Taxonomy of Critical Thinking Dispositions and Abilities”, in Joan Boykoff Baron and Robert J. Sternberg (eds.), Teaching Thinking Skills: Theory and Practice , New York: W. H. Freeman, pp. 9–26.
  • –––, 1989, “Critical Thinking and Subject Specificity: Clarification and Needed Research”, Educational Researcher , 18(3): 4–10. doi:10.3102/0013189X018003004
  • –––, 1991, “Critical Thinking: A Streamlined Conception”, Teaching Philosophy , 14(1): 5–24. doi:10.5840/teachphil19911412
  • –––, 1996, “Critical Thinking Dispositions: Their Nature and Assessability”, Informal Logic , 18(2–3): 165–182. [ Ennis 1996 available online ]
  • –––, 1998, “Is Critical Thinking Culturally Biased?”, Teaching Philosophy , 21(1): 15–33. doi:10.5840/teachphil19982113
  • –––, 2011, “Critical Thinking: Reflection and Perspective Part I”, Inquiry: Critical Thinking across the Disciplines , 26(1): 4–18. doi:10.5840/inquiryctnews20112613
  • –––, 2013, “Critical Thinking across the Curriculum: The Wisdom CTAC Program”, Inquiry: Critical Thinking across the Disciplines , 28(2): 25–45. doi:10.5840/inquiryct20132828
  • –––, 2016, “Definition: A Three-Dimensional Analysis with Bearing on Key Concepts”, in Patrick Bondy and Laura Benacquista (eds.), Argumentation, Objectivity, and Bias: Proceedings of the 11th International Conference of the Ontario Society for the Study of Argumentation (OSSA), 18–21 May 2016 , Windsor, ON: OSSA, pp. 1–19. Available at http://scholar.uwindsor.ca/ossaarchive/OSSA11/papersandcommentaries/105 ; last accessed 2022 07 16.
  • –––, 2018, “Critical Thinking Across the Curriculum: A Vision”, Topoi , 37(1): 165–184. doi:10.1007/s11245-016-9401-4
  • Ennis, Robert H., and Jason Millman, 1971, Manual for Cornell Critical Thinking Test, Level X, and Cornell Critical Thinking Test, Level Z , Urbana, IL: Critical Thinking Project, University of Illinois.
  • Ennis, Robert H., Jason Millman, and Thomas Norbert Tomko, 1985, Cornell Critical Thinking Tests Level X & Level Z: Manual , Pacific Grove, CA: Midwest Publication, 3rd edition.
  • –––, 2005, Cornell Critical Thinking Tests Level X & Level Z: Manual , Seaside, CA: Critical Thinking Company, 5th edition.
  • Ennis, Robert H. and Eric Weir, 1985, The Ennis-Weir Critical Thinking Essay Test: Test, Manual, Criteria, Scoring Sheet: An Instrument for Teaching and Testing , Pacific Grove, CA: Midwest Publications.
  • Facione, Peter A., 1990a, Critical Thinking: A Statement of Expert Consensus for Purposes of Educational Assessment and Instruction , Research Findings and Recommendations Prepared for the Committee on Pre-College Philosophy of the American Philosophical Association, ERIC Document ED315423.
  • –––, 1990b, California Critical Thinking Skills Test, CCTST – Form A , Millbrae, CA: The California Academic Press.
  • –––, 1990c, The California Critical Thinking Skills Test--College Level. Technical Report #3. Gender, Ethnicity, Major, CT Self-Esteem, and the CCTST , ERIC Document ED326584.
  • –––, 1992, California Critical Thinking Skills Test: CCTST – Form B, Millbrae, CA: The California Academic Press.
  • –––, 2000, “The Disposition Toward Critical Thinking: Its Character, Measurement, and Relationship to Critical Thinking Skill”, Informal Logic , 20(1): 61–84. [ Facione 2000 available online ]
  • Facione, Peter A. and Noreen C. Facione, 1992, CCTDI: A Disposition Inventory , Millbrae, CA: The California Academic Press.
  • Facione, Peter A., Noreen C. Facione, and Carol Ann F. Giancarlo, 2001, California Critical Thinking Disposition Inventory: CCTDI: Inventory Manual , Millbrae, CA: The California Academic Press.
  • Facione, Peter A., Carol A. Sánchez, and Noreen C. Facione, 1994, Are College Students Disposed to Think? , Millbrae, CA: The California Academic Press. ERIC Document ED368311.
  • Fisher, Alec, and Michael Scriven, 1997, Critical Thinking: Its Definition and Assessment , Norwich: Centre for Research in Critical Thinking, University of East Anglia.
  • Freire, Paulo, 1968 [1970], Pedagogia do Oprimido . Translated as Pedagogy of the Oppressed , Myra Bergman Ramos (trans.), New York: Continuum, 1970.
  • Gigerenzer, Gerd, 2001, “The Adaptive Toolbox”, in Gerd Gigerenzer and Reinhard Selten (eds.), Bounded Rationality: The Adaptive Toolbox , Cambridge, MA: MIT Press, pp. 37–50.
  • Glaser, Edward Maynard, 1941, An Experiment in the Development of Critical Thinking , New York: Bureau of Publications, Teachers College, Columbia University.
  • Groarke, Leo A. and Christopher W. Tindale, 2012, Good Reasoning Matters! A Constructive Approach to Critical Thinking , Don Mills, ON: Oxford University Press, 5th edition.
  • Halpern, Diane F., 1998, “Teaching Critical Thinking for Transfer Across Domains: Disposition, Skills, Structure Training, and Metacognitive Monitoring”, American Psychologist , 53(4): 449–455. doi:10.1037/0003-066X.53.4.449
  • –––, 2016, Manual: Halpern Critical Thinking Assessment , Mödling, Austria: Schuhfried. Available at https://pdfcoffee.com/hcta-test-manual-pdf-free.html; last accessed 2022 07 16.
  • Hamby, Benjamin, 2014, The Virtues of Critical Thinkers , Doctoral dissertation, Philosophy, McMaster University. [ Hamby 2014 available online ]
  • –––, 2015, “Willingness to Inquire: The Cardinal Critical Thinking Virtue”, in Martin Davies and Ronald Barnett (eds.), The Palgrave Handbook of Critical Thinking in Higher Education , New York: Palgrave Macmillan, pp. 77–87.
  • Haran, Uriel, Ilana Ritov, and Barbara A. Mellers, 2013, “The Role of Actively Open-minded Thinking in Information Acquisition, Accuracy, and Calibration”, Judgment and Decision Making , 8(3): 188–201.
  • Hatcher, Donald and Kevin Possin, 2021, “Commentary: Thinking Critically about Critical Thinking Assessment”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 298–322. doi: 10.1163/9789004444591_017
  • Haynes, Ada, Elizabeth Lisic, Kevin Harris, Katie Leming, Kyle Shanks, and Barry Stein, 2015, “Using the Critical Thinking Assessment Test (CAT) as a Model for Designing Within-Course Assessments: Changing How Faculty Assess Student Learning”, Inquiry: Critical Thinking Across the Disciplines , 30(3): 38–48. doi:10.5840/inquiryct201530316
  • Haynes, Ada and Barry Stein, 2021, “Observations from a Long-Term Effort to Assess and Improve Critical Thinking”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 231–254. doi: 10.1163/9789004444591_014
  • Hiner, Amanda L. 2021. “Equipping Students for Success in College and Beyond: Placing Critical Thinking Instruction at the Heart of a General Education Program”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 188–208. doi: 10.1163/9789004444591_012
  • Hitchcock, David, 2017, “Critical Thinking as an Educational Ideal”, in his On Reasoning and Argument: Essays in Informal Logic and on Critical Thinking , Dordrecht: Springer, pp. 477–497. doi:10.1007/978-3-319-53562-3_30
  • –––, 2021, “Seven Philosophical Implications of Critical Thinking: Themes, Variations, Implications”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 9–30. doi: 10.1163/9789004444591_002
  • hooks, bell, 1994, Teaching to Transgress: Education as the Practice of Freedom , New York and London: Routledge.
  • –––, 2010, Teaching Critical Thinking: Practical Wisdom , New York and London: Routledge.
  • Johnson, Ralph H., 1992, “The Problem of Defining Critical Thinking”, in Stephen P, Norris (ed.), The Generalizability of Critical Thinking , New York: Teachers College Press, pp. 38–53.
  • Kahane, Howard, 1971, Logic and Contemporary Rhetoric: The Use of Reason in Everyday Life , Belmont, CA: Wadsworth.
  • Kahneman, Daniel, 2011, Thinking, Fast and Slow , New York: Farrar, Straus and Giroux.
  • Kahneman, Daniel, Olivier Sibony, & Cass R. Sunstein, 2021, Noise: A Flaw in Human Judgment , New York: Little, Brown Spark.
  • Kenyon, Tim, and Guillaume Beaulac, 2014, “Critical Thinking Education and Debasing”, Informal Logic , 34(4): 341–363. [ Kenyon & Beaulac 2014 available online ]
  • Krathwohl, David R., Benjamin S. Bloom, and Bertram B. Masia, 1964, Taxonomy of Educational Objectives, Handbook II: Affective Domain , New York: David McKay.
  • Kuhn, Deanna, 1991, The Skills of Argument , New York: Cambridge University Press. doi:10.1017/CBO9780511571350
  • –––, 2019, “Critical Thinking as Discourse”, Human Development, 62 (3): 146–164. doi:10.1159/000500171
  • Lipman, Matthew, 1987, “Critical Thinking–What Can It Be?”, Analytic Teaching , 8(1): 5–12. [ Lipman 1987 available online ]
  • –––, 2003, Thinking in Education , Cambridge: Cambridge University Press, 2nd edition.
  • Loftus, Elizabeth F., 2017, “Eavesdropping on Memory”, Annual Review of Psychology , 68: 1–18. doi:10.1146/annurev-psych-010416-044138
  • Makaiau, Amber Strong, 2021, “The Good Thinker’s Tool Kit: How to Engage Critical Thinking and Reasoning in Secondary Education”, in Daniel Fasko, Jr. and Frank Fair (eds.), Critical Thinking and Reasoning: Theory, Development, Instruction, and Assessment , Leiden: Brill, pp. 168–187. doi: 10.1163/9789004444591_011
  • Martin, Jane Roland, 1992, “Critical Thinking for a Humane World”, in Stephen P. Norris (ed.), The Generalizability of Critical Thinking , New York: Teachers College Press, pp. 163–180.
  • Mayhew, Katherine Camp, and Anna Camp Edwards, 1936, The Dewey School: The Laboratory School of the University of Chicago, 1896–1903 , New York: Appleton-Century. [ Mayhew & Edwards 1936 available online ]
  • McPeck, John E., 1981, Critical Thinking and Education , New York: St. Martin’s Press.
  • Moore, Brooke Noel and Richard Parker, 2020, Critical Thinking , New York: McGraw-Hill, 13th edition.
  • Nickerson, Raymond S., 1998, “Confirmation Bias: A Ubiquitous Phenomenon in Many Guises”, Review of General Psychology , 2(2): 175–220. doi:10.1037/1089-2680.2.2.175
  • Nieto, Ana Maria, and Jorge Valenzuela, 2012, “A Study of the Internal Structure of Critical Thinking Dispositions”, Inquiry: Critical Thinking across the Disciplines , 27(1): 31–38. doi:10.5840/inquiryct20122713
  • Norris, Stephen P., 1985, “Controlling for Background Beliefs When Developing Multiple-choice Critical Thinking Tests”, Educational Measurement: Issues and Practice , 7(3): 5–11. doi:10.1111/j.1745-3992.1988.tb00437.x
  • Norris, Stephen P. and Robert H. Ennis, 1989, Evaluating Critical Thinking (The Practitioners’ Guide to Teaching Thinking Series), Pacific Grove, CA: Midwest Publications.
  • Norris, Stephen P. and Ruth Elizabeth King, 1983, Test on Appraising Observations , St. John’s, NL: Institute for Educational Research and Development, Memorial University of Newfoundland.
  • –––, 1984, The Design of a Critical Thinking Test on Appraising Observations , St. John’s, NL: Institute for Educational Research and Development, Memorial University of Newfoundland. ERIC Document ED260083.
  • –––, 1985, Test on Appraising Observations: Manual , St. John’s, NL: Institute for Educational Research and Development, Memorial University of Newfoundland.
  • –––, 1990a, Test on Appraising Observations , St. John’s, NL: Institute for Educational Research and Development, Memorial University of Newfoundland, 2nd edition.
  • –––, 1990b, Test on Appraising Observations: Manual , St. John’s, NL: Institute for Educational Research and Development, Memorial University of Newfoundland, 2nd edition.
  • OCR [Oxford, Cambridge and RSA Examinations], 2011, AS/A Level GCE: Critical Thinking – H052, H452 , Cambridge: OCR. Past papers available at https://pastpapers.co/ocr/?dir=A-Level/Critical-Thinking-H052-H452; last accessed 2022 07 16.
  • Ontario Ministry of Education, 2013, The Ontario Curriculum Grades 9 to 12: Social Sciences and Humanities . Available at http://www.edu.gov.on.ca/eng/curriculum/secondary/ssciences9to122013.pdf ; last accessed 2022 07 16.
  • Passmore, John Arthur, 1980, The Philosophy of Teaching , London: Duckworth.
  • Paul, Richard W., 1981, “Teaching Critical Thinking in the ‘Strong’ Sense: A Focus on Self-Deception, World Views, and a Dialectical Mode of Analysis”, Informal Logic , 4(2): 2–7. [ Paul 1981 available online ]
  • –––, 1984, “Critical Thinking: Fundamental to Education for a Free Society”, Educational Leadership , 42(1): 4–14.
  • –––, 1985, “McPeck’s Mistakes”, Informal Logic , 7(1): 35–43. [ Paul 1985 available online ]
  • Paul, Richard W. and Linda Elder, 2006, The Miniature Guide to Critical Thinking: Concepts and Tools , Dillon Beach, CA: Foundation for Critical Thinking, 4th edition.
  • Payette, Patricia, and Edna Ross, 2016, “Making a Campus-Wide Commitment to Critical Thinking: Insights and Promising Practices Utilizing the Paul-Elder Approach at the University of Louisville”, Inquiry: Critical Thinking Across the Disciplines , 31(1): 98–110. doi:10.5840/inquiryct20163118
  • Possin, Kevin, 2008, “A Field Guide to Critical-Thinking Assessment”, Teaching Philosophy , 31(3): 201–228. doi:10.5840/teachphil200831324
  • –––, 2013a, “Some Problems with the Halpern Critical Thinking Assessment (HCTA) Test”, Inquiry: Critical Thinking across the Disciplines , 28(3): 4–12. doi:10.5840/inquiryct201328313
  • –––, 2013b, “A Serious Flaw in the Collegiate Learning Assessment (CLA) Test”, Informal Logic , 33(3): 390–405. [ Possin 2013b available online ]
  • –––, 2013c, “A Fatal Flaw in the Collegiate Learning Assessment Test”, Assessment Update , 25 (1): 8–12.
  • –––, 2014, “Critique of the Watson-Glaser Critical Thinking Appraisal Test: The More You Know, the Lower Your Score”, Informal Logic , 34(4): 393–416. [ Possin 2014 available online ]
  • –––, 2020, “CAT Scan: A Critical Review of the Critical-Thinking Assessment Test”, Informal Logic , 40 (3): 489–508. [Available online at https://informallogic.ca/index.php/informal_logic/article/view/6243]
  • Rawls, John, 1971, A Theory of Justice , Cambridge, MA: Harvard University Press.
  • Rear, David, 2019, “One Size Fits All? The Limitations of Standardised Assessment in Critical Thinking”, Assessment & Evaluation in Higher Education , 44(5): 664–675. doi: 10.1080/02602938.2018.1526255
  • Rousseau, Jean-Jacques, 1762, Émile , Amsterdam: Jean Néaulme.
  • Scheffler, Israel, 1960, The Language of Education , Springfield, IL: Charles C. Thomas.
  • Scriven, Michael, and Richard W. Paul, 1987, Defining Critical Thinking , Draft statement written for the National Council for Excellence in Critical Thinking Instruction. Available at http://www.criticalthinking.org/pages/defining-critical-thinking/766 ; last accessed 2022 07 16.
  • Sheffield, Clarence Burton Jr., 2018, “Promoting Critical Thinking in Higher Education: My Experiences as the Inaugural Eugene H. Fram Chair in Applied Critical Thinking at Rochester Institute of Technology”, Topoi , 37(1): 155–163. doi:10.1007/s11245-016-9392-1
  • Siegel, Harvey, 1985, “McPeck, Informal Logic and the Nature of Critical Thinking”, in David Nyberg (ed.), Philosophy of Education 1985: Proceedings of the Forty-First Annual Meeting of the Philosophy of Education Society , Normal, IL: Philosophy of Education Society, pp. 61–72.
  • –––, 1988, Educating Reason: Rationality, Critical Thinking, and Education , New York: Routledge.
  • –––, 1999, “What (Good) Are Thinking Dispositions?”, Educational Theory , 49(2): 207–221. doi:10.1111/j.1741-5446.1999.00207.x
  • Simon, Herbert A., 1956, “Rational Choice and the Structure of the Environment”, Psychological Review , 63(2): 129–138. doi: 10.1037/h0042769
  • Simpson, Elizabeth, 1966–67, “The Classification of Educational Objectives: Psychomotor Domain”, Illinois Teacher of Home Economics , 10(4): 110–144, ERIC document ED0103613. [ Simpson 1966–67 available online ]
  • Skolverket, 2018, Curriculum for the Compulsory School, Preschool Class and School-age Educare , Stockholm: Skolverket, revised 2018. Available at https://www.skolverket.se/download/18.31c292d516e7445866a218f/1576654682907/pdf3984.pdf; last accessed 2022 07 15.
  • Smith, B. Othanel, 1953, “The Improvement of Critical Thinking”, Progressive Education , 30(5): 129–134.
  • Smith, Eugene Randolph, Ralph Winfred Tyler, and the Evaluation Staff, 1942, Appraising and Recording Student Progress , Volume III of Adventure in American Education , New York and London: Harper & Brothers.
  • Splitter, Laurance J., 1987, “Educational Reform through Philosophy for Children”, Thinking: The Journal of Philosophy for Children , 7(2): 32–39. doi:10.5840/thinking1987729
  • Stanovich Keith E., and Paula J. Stanovich, 2010, “A Framework for Critical Thinking, Rational Thinking, and Intelligence”, in David D. Preiss and Robert J. Sternberg (eds), Innovations in Educational Psychology: Perspectives on Learning, Teaching and Human Development , New York: Springer Publishing, pp 195–237.
  • Stanovich Keith E., Richard F. West, and Maggie E. Toplak, 2011, “Intelligence and Rationality”, in Robert J. Sternberg and Scott Barry Kaufman (eds.), Cambridge Handbook of Intelligence , Cambridge: Cambridge University Press, 3rd edition, pp. 784–826. doi:10.1017/CBO9780511977244.040
  • Tankersley, Karen, 2005, Literacy Strategies for Grades 4–12: Reinforcing the Threads of Reading , Alexandria, VA: Association for Supervision and Curriculum Development.
  • Thayer-Bacon, Barbara J., 1992, “Is Modern Critical Thinking Theory Sexist?”, Inquiry: Critical Thinking Across the Disciplines , 10(1): 3–7. doi:10.5840/inquiryctnews199210123
  • –––, 1993, “Caring and Its Relationship to Critical Thinking”, Educational Theory , 43(3): 323–340. doi:10.1111/j.1741-5446.1993.00323.x
  • –––, 1995a, “Constructive Thinking: Personal Voice”, Journal of Thought , 30(1): 55–70.
  • –––, 1995b, “Doubting and Believing: Both are Important for Critical Thinking”, Inquiry: Critical Thinking across the Disciplines , 15(2): 59–66. doi:10.5840/inquiryctnews199515226
  • –––, 2000, Transforming Critical Thinking: Thinking Constructively , New York: Teachers College Press.
  • Toulmin, Stephen Edelston, 1958, The Uses of Argument , Cambridge: Cambridge University Press.
  • Turri, John, Mark Alfano, and John Greco, 2017, “Virtue Epistemology”, in Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy (Winter 2017 Edition). URL = < https://plato.stanford.edu/archives/win2017/entries/epistemology-virtue/ >
  • Vincent-Lancrin, Stéphan, Carlos González-Sancho, Mathias Bouckaert, Federico de Luca, Meritxell Fernández-Barrerra, Gwénaël Jacotin, Joaquin Urgel, and Quentin Vidal, 2019, Fostering Students’ Creativity and Critical Thinking: What It Means in School. Educational Research and Innovation , Paris: OECD Publishing.
  • Warren, Karen J. 1988. “Critical Thinking and Feminism”, Informal Logic , 10(1): 31–44. [ Warren 1988 available online ]
  • Watson, Goodwin, and Edward M. Glaser, 1980a, Watson-Glaser Critical Thinking Appraisal, Form A , San Antonio, TX: Psychological Corporation.
  • –––, 1980b, Watson-Glaser Critical Thinking Appraisal: Forms A and B; Manual , San Antonio, TX: Psychological Corporation,
  • –––, 1994, Watson-Glaser Critical Thinking Appraisal, Form B , San Antonio, TX: Psychological Corporation.
  • Weinstein, Mark, 1990, “Towards a Research Agenda for Informal Logic and Critical Thinking”, Informal Logic , 12(3): 121–143. [ Weinstein 1990 available online ]
  • –––, 2013, Logic, Truth and Inquiry , London: College Publications.
  • Willingham, Daniel T., 2019, “How to Teach Critical Thinking”, Education: Future Frontiers , 1: 1–17. [Available online at https://prod65.education.nsw.gov.au/content/dam/main-education/teaching-and-learning/education-for-a-changing-world/media/documents/How-to-teach-critical-thinking-Willingham.pdf.]
  • Zagzebski, Linda Trinkaus, 1996, Virtues of the Mind: An Inquiry into the Nature of Virtue and the Ethical Foundations of Knowledge , Cambridge: Cambridge University Press. doi:10.1017/CBO9781139174763
How to cite this entry . Preview the PDF version of this entry at the Friends of the SEP Society . Look up topics and thinkers related to this entry at the Internet Philosophy Ontology Project (InPhO). Enhanced bibliography for this entry at PhilPapers , with links to its database.
  • Association for Informal Logic and Critical Thinking (AILACT)
  • Critical Thinking Across the European Higher Education Curricula (CRITHINKEDU)
  • Critical Thinking Definition, Instruction, and Assessment: A Rigorous Approach
  • Critical Thinking Research (RAIL)
  • Foundation for Critical Thinking
  • Insight Assessment
  • Partnership for 21st Century Learning (P21)
  • The Critical Thinking Consortium
  • The Nature of Critical Thinking: An Outline of Critical Thinking Dispositions and Abilities , by Robert H. Ennis

abilities | bias, implicit | children, philosophy for | civic education | decision-making capacity | Dewey, John | dispositions | education, philosophy of | epistemology: virtue | logic: informal

Copyright © 2022 by David Hitchcock < hitchckd @ mcmaster . ca >

  • Accessibility

Support SEP

Mirror sites.

View this site from another server:

  • Info about mirror sites

The Stanford Encyclopedia of Philosophy is copyright © 2024 by The Metaphysics Research Lab , Department of Philosophy, Stanford University

Library of Congress Catalog Data: ISSN 1095-5054

Cognition and Instruction/Problem Solving, Critical Thinking and Argumentation

We are constantly surrounded by ambiguities, falsehoods, challenges or situations in our daily lives that require our Critical Thinking , Problem Solving Skills , and Argumentation skills . While these three terms are often used interchangeably, they are notably different. Critical thinking enables us to actively engage with information that we are presented with through all of our senses, and to think deeply about such information. This empowers us to analyse, critique, and apply knowledge, as well as create new ideas. Critical thinking can be considered the overarching cognitive skill of problem solving and argumentation. With critical thinking, although there are logical conclusions we can arrive at, there is not necessarily a 'right' idea. What may seem 'right' is often very subjective. Problem solving is a form of critical thinking that confronts learners with decisions to be made about best possible solutions, with no specific right answer for well-defined and ill-defined problems. One method of engaging with Problem Solving is with tutor systems such as Cognitive Tutor which can modify problems for individual students as well as track their progress in learning. Particular to Problem Solving is Project Based Learning which focuses the learner on solving a driving question, placing the student in the centre of learning experience by conducting an extensive investigation. Problem Based Learning focuses on real-life problems that motivate the student with experiential learning. Further, Design Thinking uses a specific scaffold system to encourage learners to develop a prototype to solve a real-world problem through a series of steps. Empathy, practical design principles, and refinement of prototyping demonstrate critical thought throughout this process. Likewise, argumentation is a critical thinking process that does not necessarily involve singular answers, hence the requirement for negotiation in argumentative thought. More specifically, argumentation involves using reasoning to support or refute a claim or idea. In comparison problem solving may lead to one solution that could be considered to be empirical.

This chapter provides a theoretical overview of these three key topics: the qualities of each, their relationship to each other, as well as practical classroom applications.

Learning Outcomes:

  • Defining Critical Thought and its interaction with knowledge
  • Defining Problem Solving and how it uses Critical Thought to develop solutions to problems
  • Introduce a Cognitive Tutor as a cognitive learning tool that employs problem solving to enhance learning
  • Explore Project Based Learning as a specific method of Problem Solving
  • Examine Design Thinking as a sub-set of Project Based Learning and its scaffold process for learning
  • Define Argumentation and how it employs a Critical Though process
  • Examine specific methodologies and instruments of application for argumentation
  • 1.1 Defining critical thinking
  • 1.2 Critical thinking as a western construct
  • 1.3 Critical thinking in other parts of the world
  • 1.4 Disposition and critical thinking
  • 1.5 Self-regulation and critical thinking
  • 1.6.1 Venn Diagrams
  • 1.6.2.1 The classroom environment
  • 1.6.3.1 Socratic Method
  • 1.6.3.2 Bloom’s Taxonomy
  • 1.6.3.3 Norman Webb’s Depth of Knowledge
  • 1.6.3.4 Williams Model
  • 1.6.3.5 Wiggins & McTighe’s Six Facets of Understanding
  • 2.1.1.1.1 Structure Of The Classroom
  • 2.2.1.1 Instructional Implications
  • 2.2.2.1 Instructional Implications
  • 2.3.1 Mind set
  • 2.3.2.1.1 Instructional Implications
  • 2.4 Novice Versus Expert In Problem Solving
  • 2.5.1 An overview of Cognitive Tutor
  • 2.5.2.1 ACT-R theory
  • 2.5.2.2 Production rules
  • 2.5.2.3 Cognitive model and model tracing
  • 2.5.2.4 Knowledge tracing
  • 2.5.3.1 Cognitive Tutor® Geometry
  • 2.5.3.2 Genetics Cognitive Tutor
  • 2.6.1 Theorizing Solutions for Real World Problems
  • 2.6.2 Experience is the Foundation of Learning
  • 2.6.3 Self-Motivation Furthers Student Learning
  • 2.6.4 Educators Find Challenges in Project Based Learning Implementation
  • 2.6.5 Learner Need for Authentic Results through Critical Thought
  • 2.7.1 Using the Process of Practical Design for Real-World Solutions
  • 2.7.2 Critical Thought on Design in the Artificial World
  • 2.7.3 Critical Thinking as Disruptive Achievement
  • 2.7.4 Designers are Not Scientific?
  • 2.7.5 21st Century Learners and the Need for Divergent Thinking
  • 3.1 Educators Find Challenges in Project Based Learning Implementation
  • 3.2 Learner Need for Authentic Results through Critical Thought
  • 3.3 Critical Thinking as Disruptive Achievement
  • 3.4.1 Argumentation Stages
  • 3.5 The Impact of Argumentation on Learning
  • 4.1.1 Production, Analysis, and Evaluation
  • 4.2 How Argumentation Improves Critical Thinking
  • 5.1 Teaching Tactics
  • 5.2.1 The CoRT Thinking Materials
  • 5.2.2 The Feuerstein Instrumental Enrichment Program (FIE)
  • 5.2.3 The Productive Thinking Program
  • 5.2.4 The IDEAL Problem Solver
  • 5.3.1 Dialogue and Argumentation
  • 5.3.2 Science and Argumentation
  • 5.3.3.1 Historical Thinking - The Big Six
  • 5.4 Instructing through Academic Controversy
  • 7.1 External links
  • 8 References

Critical thinking [ edit | edit source ]

Critical thinking and its relationship to other cognitive skills

Critical thinking is an extremely valuable aspect of education. The ability to think critically often increases over the lifespan as knowledge and experience is acquired, but it is crucial to begin the process of this development as early on as possible. Research has indicated that critical thinking skills are correlated with better transfer of knowledge, while a lack of critical thinking skills has been associated with biased reasoning [1] . Before children even begin formal schooling, they develop critical thinking skills at home because of interactions with parents and caregivers [2] . As well, critical thinking appears to improve with explicit instruction [3] . Being able to engage in critical thought is what allows us to make informed decisions in situations like elections, in which candidates present skewed views of themselves and other candidates. Without critical thinking, people would fall prey to fallacious information and biased reasoning. It is therefore important that students are introduced to critical thought and are encouraged to utilize critical thinking skills as they face problems.

Defining critical thinking [ edit | edit source ]

In general, critical thinking can be defined as the process of evaluating arguments and evidence to reach a conclusion that is the most appropriate and valid among other possible conclusions. Critical thinking is a dynamic and reflective process, and it is primarily evidence-based [4] . Thinking critically involves being able to criticize information objectively and explore opposing views, eventually leading to a conclusion based on evidence and careful thought. Critical thinkers are skeptical of information given to them, actively seek out evidence, and are not hesitant to take on decision-making and complex problem solving tasks [5] . Asking questions, debating topics, and critiquing the credibility of sources are all activities that involve thinking critically. As outlined by Glaser (1941), critical thinking involves three main components: a disposition for critical thought, knowledge of critical thinking strategies, and some ability to apply the strategies [6] . Having a disposition for critical thought is necessary for applying known strategies.

Critical thinking, which includes cognitive processes such as weighing and evaluating information, leads to more thorough understanding of an issue or problem. As a type of reflection, critical thinking also promotes an awareness of one's own perceptions, intentions, feelings and actions. [7]

Critical thinking as a western construct [ edit | edit source ]

Critical thinking is considered to be essential for all democratic citizens

In modern education, critical thinking is taken for granted as something that people universally need and should acquire, especially at a higher educational level [8] [9] . However, critical thinking is a human construct [10] - not a scientific fact - that is tied to Ancient Greek philosophy and beliefs [11] .

The link to Ancient Greece relates both to Ancient Greek priorities of logic over emotion [11] , as well as its democratic principles. Various authors, including Elder & Paul [12] , Moon [8] , and Stanlick & Strawser [13] share the view that critical thinking questioning back to the time of Socrates . Likewise, Morgan & Saxton (2006) associate critical thinking with a fundamental requirement of all democratic citizens [14] .

An additional connection with Ancient Greece involves the Socratic Method. The Socratic Method involves a conversation between two or more people in which they ask and answer questions to challenge each other’s theses using logic and reason [15] . Such debates are subject to the issue of objective/subjective dualism in that the purpose of debate is the belief that there is a ‘right answer’, yet the ability to conduct such a debate demonstrates the subjectivity of any thesis [15] .

Because of this strong connection to Ancient Greece, critical thinking is generally considered to be a western construct. This is further amplified another western construct called Bloom’s Taxonomy , which is considered to be the essence of critical thinking in modern education [16] .

Since critical thinking is a human construct, notions of what constitutes critical thinking vary considerably from person to person. Moon (2007) lists 21 common notions of critical thinking provided by people from her workshops, and then provides her own 2-page definition of the term [8] . One view of critical thinking is that it involves a set of skills that enables one to reach defensible conclusions and make decisions in a domain or context in which one has some prior knowledge [10] . Another view is that critical thinking involves the use of systematic logic and reasoning, which while not necessarily producing empirical answers nevertheless uses a rational and scientific approach [17] . Ultimately, Moon concludes that there is no right or wrong definition [8] .

Critical thinking in other parts of the world [ edit | edit source ]

Scholars argue that while the critical thinking construct is linked to western, democratic nations, that does not mean that other non-western cultures do not possess or use similar constructs that involve critical thinking [18] . Instead, “there are different ways or forms of reasoning” [19] ; for example, Asian approaches to debates involve finding connections between conflictive arguments in order for such ideas to coexist [18] . This is due to eastern values regarding face-saving [8] . In contrast, western approaches are often viewed as being competitive: attacking the views of others while defending one's own position. Despite this dichotomous generalisation, eastern and western approaches have more similarities than they would first seem. With regards to the diplomatic Asian approach to debating, western approaches also involve compromise and negotiation for the very reason that ideas are often complex and that there can be many ‘right’ answers [14] . Similarly, the extent to which other cultures adopt western notions of critical thinking is determined by cultural values. In Muslim cultures, for example, the value of critical thinking is link to views on the appropriateness of voicing one’s views [20] .

Disposition and critical thinking [ edit | edit source ]

It has been suggested that critical thinking skills alone are not sufficient for the application of critical thinking – a disposition for critical thinking is also necessary [5] . A disposition for critical thought differs from cognitive skills. A disposition is better explained as the ability to consciously choose a skill, rather than just the ability to execute the skill [4] . Having a disposition for critical thinking can include such things as genuine interest and ability in intellectual activities. Perkins et al. (2000) expand on the idea of the necessity for a critical thinking disposition, and indicate three aspects involved in critical thinking disposition: an inclination for engaging in intellectual behaviours; a sensitivity to opportunities, in which such behaviours may be engaged; and a general ability for engaging in critical thought [5] . Halpern (1998) suggests that this critical thinking disposition must include a willingness to continue with tasks that seem difficult, openmindedness, and a habit of planning [5] . In fact, in a cognitive skills study conducted by Clifford et al. (2004), they discovered that a disposition for critical thinking was associated with better overall critical thinking skills [4] .

These are characteristics of one's attitude or personality that facilitate the process of developing CT skills:

  • Inquisitive
  • Truthseeking
  • Open-minded
  • Confidence in reasoning

There are many factors that can influence one's disposition towards CT; the first of these is culture [5] . There are many aspects of culture that can impact the ability for people to think critically. For instance, religion can negatively impact the development of CT [5] . Many religions are founded upon faith, which often requires wholehearted belief without evidence or support. The nature of organized religion counters the very premise of CT, which is to evaluate the validity and credibility of any claim. Growing up in an environment such as this can be detrimental to the development of CT skills. This kind of environment can dampen dispositions that question religious views or examine the validity of religion. Another cultural factor that can be detrimental to a CT disposition is that of authority [5] . When a child is raised under the conditions of an authoritarian parenting style, it can be detrimental to many aspects of their lives, but especially to their CT skills, as they are taught not to question the credibility of authority and often receive punishment if they do. This is also applicable in the classroom [5] . Classroom environments that foster a disposition for critical thinking in which teachers who do not foster an atmosphere of openness or allow students to question what they are taught can impact CT development as well. Classrooms where questions are rejected or home environments in which there is a high level of parental power and control can all affect the ability of students to think critically. What is more, students will have been conditioned not to think this way for their entire lives [5] . However, despite these cultural limitations, there are ways in which a disposition for CT can be fostered in both the home and the classroom.

Classroom structure is a primary way in which CT dispositions can be highlighted. Fostering a classroom structure in which students are a part of the decision making process of what they are studying can be very helpful in creating CT dispositions [5] . Such structures help students become invested in what they are learning as well as promote a classroom atmosphere in which students may feel free to question the teacher, as well as other students' opinions and beliefs about different subjects. Allowing the freedom to scrutinize and evaluate information that has been given to students is an effective way of creating a classroom environment that can encourage students to develop CT dispositions. This freedom allows for the students to remain individuals within the larger classroom context, and gives them the power to evaluate and make decisions on their own. Allowing the students to share power in the classroom can be extremely beneficial in helping the students stay motivated and analytical of classroom teachings [5] . Teachers can also employ a variety of techniques that can help students become autonomous in the classroom. Giving students the opportunity to take on different roles can be effective in creating CT dispositions, such as making predictions and contemplating problems [5] . Allowing students to engage with problems that are presented, instead of just teaching them what the teacher or textbook believes to be true, is essential for students to develop their own opinions and individual, though. In addition to this, gathering data and information on the subject is an important part of developing CT dispositions. Doing so allows for students to go out and find resources that they themselves can analyze and come to conclusions on their own [5] . Using these aspects of CT students can most effectively relate to the predictions that were first made and critique the validity of the findings [5] .

Self-regulation and critical thinking [ edit | edit source ]

In conjunction with instructing CT, teachers also need to keep in mind the self-regulation of their students. Students need to be able to maintain motivation and have a proactive attitude towards their own learning when learning a new skill. In an article by Phan (2010), he argues that self-regulated students that have better goal setting have more personal responsibility for their learning, can maintain their motivation, are more cognitively flexible, and hence are more inclined to utilize CT. Since CT skills are highly reflective, they help in self-regulated learning (SRL), and in turn, self-regulatory strategies aid in developing CT skills. These two cognitive practices are assets to students’ growth and development [7] .

Self-Regulation provides students with the basic meta-cognitive awareness required for proactive learning. This pro-activity allows students to engage in the cognitive processes of CT, such as evaluation, reflection and inference. Through one’s meta-cognitive ability to assess one’s own thoughts, one develops the capability to become autonomous in one’s learning [7] . Instead of having a supervisor overlook every task, the learner can progress at their own pace while monitoring their performance, thereby engaging in SRL. Part of this process would include periodic reflection upon the strategies that one uses when completing a task. This reflection can facilitate the student’s learning by using CT to evaluate which strategies best suit their own learning based on their cognitive needs.

The complex nature of CT suggests that it requires a long developmental process requiring guidance, practice and reinforcement. To facilitate this process, self-monitoring as a first step to self-regulation can jump-start reflective thought through assessing one’s own educational performance. This assessment promotes self-efficacy through generating motivational beliefs about one’s academic capabilities [7] . From there, through practice, students can extend their CT skills beyond themselves and into their educational contexts. With practice, students use their meta-cognitive strategies as a basis for developing CT in the long run.

Critical thinking strategies [ edit | edit source ]

discuss the relationship between critical thinking and problem solving

Psychologists and educators have discovered many different strategies for the development of critical thinking. Among these strategies are some that may be very familiar, such as concept maps or Venn diagrams , as well as some that may be less familiar, such as appeal-question stimuli strategies [21] . Concept mapping is particularly useful for illustrating the relationships between ideas and concepts, while Venn diagrams are often used to represent contrasting ideas [21] .

Venn Diagrams [ edit | edit source ]

Venn diagrams are used frequently in elementary grade levels and continue to be used as a contrast/compare tool throughout secondary school. An example of a situation in which a Venn diagram activity may be appropriate is during a science class. Instructors may direct students to develop a Venn diagram comparing and contrasting different plants or animals. Concept maps may be introduced in elementary grades, although they are most often used in the secondary and post-secondary levels. Concept maps are an interactive and versatile way to encourage students to engage with the course material. A key aspect of concept mapping is how it requires students to reflect on previously learned information and make connections. In elementary grades, concept maps can be introduced as a project, while later, possibly in college or university, students may use them as a study strategy. At the elementary level, students can use concept maps to make connections about the characters, settings, or plot in a story they have read. When introducing concept maps, teachers may provide students with a list of words or phrases and instruct the students to illustrate the connections between them in the form of a concept map. Asking questions can also be a simple and engaging way to develop critical thought. Teachers may begin by asking the students questions about the material, and then encouraging students to come up with their own questions. In secondary and post-secondary education, students may use questions as a way to assess the credibility of a source. At the elementary school level, questions can be used to assess students' understanding of the material, while also encouraging them to engage in critical thought by questioning the actions of characters in a story or the validity of an experiment. Appeal-question stimuli, founded by Svobodová, involves a process of students asking questions regarding their reading comprehension [21] .

Discussions [ edit | edit source ]

Using discussions as a way to develop students’ critical thinking skills can be a particularly valuable strategy for teachers. Peer interactions provide a basis for developing particular critical thinking skills, such as perspective taking and cooperation, which may not be as easily taught through instruction. A large part of discussions, of course, is language. Klooster (2002) suggested that critical thinking begins with asking questions [21] . Similarly, Vygotsky has claimed that language skills can be a crucial precursor for higher level thought processes [2] . As children develop larger vocabularies, they are better able to understand reading material and can then begin to think abstractly about the material and engage in thoughtful discussions with peers about what they understood [2] .

Studies have indicated that cross-age peer discussions may be particularly helpful in facilitating the development of critical thinking. Cross-age peer groups can be effective because of the motivation children tend to have when working with peers of different ages [2] . Younger children often look up to the older children as mentors and valuable sources of knowledge and experience, while older children feel a sense of maturity and a responsibility to share their knowledge and experience with younger students [2] . These cross-age peer discussions also provide students with the challenge of tailoring their use of language to the other group members in order to make their points understandable [2] . An example of cross-age peer groups that is relatively common in Canadian schools is the big buddy programs, where intermediate grade students are assigned a primary grade buddy to help over the course of the school year. Big buddies may help their little buddies with projects, advice, or school events. The big buddy/little buddy programs can be effective as younger students look up to their big buddies, and the big buddies feel a responsibility to help their little buddy. One important factor to be considered with cross-age peer discussions, as noted by Hattie (2006), is that these discussions should be highly structured activities facilitated by a teacher in order to ensure that students understand their group responsibilities [2] .

The classroom environment [ edit | edit source ]

Having an environment that is a safe place for students to ask questions and share ideas is extremely valuable for creating a classroom that encourages critical thinking. It has been suggested that students are more likely to develop a disposition for critical thinking when they are able to participate in the organization and planning of their classroom and class activities [5] . In these classrooms, students are legitimately encouraged by their teacher to engage in the decision making process regarding the functioning of the classroom [5] . It is also important for teachers to model the desired types of critical thought, by questioning themselves and other authorities in a respectful and appropriate manner [5] . Studies have indicated higher levels of cognitive engagement among students in classrooms with teachers who are enthusiastic and responsive [22] . Therefore, teachers should be encouraging and inclusive, and allow student engagement in classroom planning processes when possible.

Critical questions [ edit | edit source ]

Research is increasingly supporting the idea that critical thinking can be explicitly taught [23] . The use of critical questioning in education is of particular importance, because by teaching critical questioning, educators are actively modelling critical thinking processes. One of the key issues with teaching critical thinking in education is that students merely witness the product of critical thinking on the part of the teacher, i.e. they hear the conclusions that the teacher has reached through critical thinking [9] . Whereas an experienced critical thinker uses critical questions, these questions are implicit and not normally verbalised. However, for students to understand critical questioning and critical thinking strategies, the students must see the process of critical thinking. Modelling the formation and sequencing of critical questions explicitly demonstrates the thought process of how one can reach a logical conclusion.

There various methods of teaching critical questioning. The frameworks discussed below are among the most famous of these. All have their own strengths and weaknesses in terms of ease-of-use, complexity, and universality. Each of these methods approaches critical thinking with a specific definition of this human concept. As such, one’s own definition of critical thinking will likely affect one’s receptiveness to a specific critical questioning framework.

 Socrates

Socratic Method [ edit | edit source ]

One of the key features of western approaches to critical thinking involves the importance of critical questioning, which is linked to the Socratic Method from Ancient Greece traditions. Whether answering existing questions posed or creating new questions to be considered, critical thinking involves questions, whether explicitly / implicitly, consciously / unconsciously [13] . Browne & Keeley (2006) base their definition of critical thinking specifically on the involvement of critical questions [24] .

Answers to critical questions are not necessarily empirical. They may involve reasoning and be logical, but are nevertheless subject to alternative views from others, thus making all views both subjective and objective at the same time. Elder & Paul (2009) separate such critical questions into three categories [12] :

  • Questions that have a correct answer, which can be determined using knowledge
  • Questions that are open to subjective answers that cannot be judged
  • Questions that produce objective answers that are judged based the quality of evidence and reasoning used

Books on critical questioning tend to be influenced heavily by the Socratic Method, and they make a distinction between ‘good’ and ‘bad’ questions. Good questions are those that are relevant to the topic at hand and that take a logical, systematic approach [14] [13] , while bad questions are those that are not relevant to the topic, are superficial, and are sequenced haphazardly. Elder & Paul (2009) argue that “[i]t is not possible to be a good thinker and a poor questioner.” [25] In other words, if a person cannot thinking of relevant and logical questions, they will be unable to reach any rational conclusions.

Additionally, as indicated above, critical thinking requires more than just asking the right questions. There is a direct relationship between critical thinking and knowledge [23] . One can possess knowledge, but not know how to apply it. Conversely, one can have good critical questioning skills, but lack the knowledge to judge the merits of an answer.

In terms of teaching critical questioning using the Socratic Method, it is essential to appreciate that there is no set of questions that one can follow, since the type of critical questions needed is based on the actual context. Consequently, the examples presented by different authors vary quite considerably. Nevertheless, there are specific guidelines one can follow [26] :

  • Use critical questions to identify and understand the situation, issues, viewpoints and conclusions
  • Use critical questions to search for assumptions, ambiguity, conflicts, or fallacies
  • Use critical questions to evaluate the effects of the ideas

Part 1 of the Socratic Method is more of an information gathering stage, using questions to find out essential details, to clarify ideas or opinions, and to determine objectives. Part 2 uses the information from Part 1 and then uses questions to probe for underlying details that could provide reasons for critiquing the accuracy of the idea. Part 3 uses questions to reflect upon the consequences of such ideas.

Conklin (2012) separates the above three parts into six parts [27] :

  • Using questions to understand
  • Using questions to determine assumptions
  • Using questions to discover reasons / evidence
  • Using questions to determine perspectives
  • Using questions to determine consequences
  • Using questions to evaluate a given question

Here are some sample questions for each part [28] :

Questions for understanding:

  • Why do you think that?
  • What have you studied about this topic so far?
  • How does this relate to what you are studying now?

Questions that determine assumptions

  • How could you check that assumption?
  • What else could be assumed?
  • What are your views on that? Do you agree or disagree?

Questions that discover reasons / evidence

  • How can you be sure?
  • Why is this happening?
  • What evidence do you have to back up your opinion?

Questions that determine perspectives

  • How could you look at this argument another way?
  • Which perspective is better?

Questions that determine consequences

  • How does it affect you?
  • What impact does that have?

Questions that evaluate a given question

  • Why was I asked this question?
  • Which questions led to the most interesting answers?
  • What other questions should be asked?

Depending on the text, the Socratic Method can be extraordinarily elaborate, making it challenging for educators to apply. Conklin (2012) states that a teacher would need to spend time planning such questions in advance, rather than expect to produce them during a lesson [27] .

Bloom’s Taxonomy [ edit | edit source ]

Bloom’s Taxonomy was originally designed in 1956 to determine cognitive educational objectives and assess students’ higher-order thinking skills [29] . Since then, though, it has become adapted and used as a useful tool for promoting critical thinking skills, particularly through critical questioning [30] . These critical questions involve Bloom’s categories of understanding, applying, analysing, synthesising and evaluating. Such categories can be seen to relate to the Socratic Method promoted by other authors, i.e. the importance of questioning to understanding, analyse and evaluate. Moon (2007) believes that “‘evaluation’, ‘reflection’ and ‘understanding’” are key aspects of critical thinking [8] , which should therefore appear in any notion of critical thinking. At the same time, Bloom’s Taxonomy generates a natural set of questions that can be adapted to various contexts [31] .

In one example, a teacher uses a picture of a New York speakeasy bar. Using Bloom’s Taxonomy, the teacher could ask and model the following critical questions [14] :

  • KNOWLEDGE: What do you see in the picture?
  • COMPREHENSION: What do people do in places like that?
  • ANALYSIS: Why are there so many policemen in the picture?
  • APPLICATION: What similar situations do we see nowadays?
  • SYNTHESIS: What if there were no laws prohibiting such behaviour?
  • EVALUATION: How would you feel if you were one of these people? Why?

Norman Webb's Depth of Knowledge

Norman Webb’s Depth of Knowledge [ edit | edit source ]

Webb’s Depth of Knowledge (DOK) taxonomy was produced in 2002 in response to Bloom’s Taxonomy [32] . In contrast with Bloom’s Taxonomy, Webb’s DOK focuses on considering thinking in terms of complexity of thinking rather than difficulty [32] .

Webb’s DOK has four levels:

  • Recall & reproduction
  • Working with skills & concepts
  • Short-term strategic thinking
  • Extended strategic thinking

Level 1 aligns with Bloom’s level of remembering and recalling information. Example critical questions in this level would include:

  • What is the name of the protagonist?
  • What did Oliver Twist ask Fagin?

Level 2 involves various skills, such as classifying, comparing, predicting, gathering, and displaying. Critical questions can be derived from these skill sets, including the following:

  • How do these two ideas compare?
  • How would you categorise these objects?
  • How would you summarize the text?

Level 3 involves analysis and evaluation, once again aligning with Bloom’s Taxonomy.

  • What conclusions can you reach?
  • What theory can you generate to explain this?
  • What is the best answer? Why?

At the same time, Level 3 of DOK shares similarities with the Socratic Method in that the individual must defend their views.

Level 4 is the most elaborate and challenging level. It involves making interdisciplinary connections and the creation of new ideas / solutions.

Since DOK becomes increasingly elaborate with levels and leads to the requirement to defend one’s position using logic and evidence, there are parallels with the Socratic Method. At the same time, because is used to develop standards in assessing critical thinking, it shares similarities with Bloom’s Taxonomy.

Williams Model [ edit | edit source ]

 The KWL method shares some similarities to the 'wonder' aspect of the Williams Model

The Williams Model was designed by Frank Williams in the 1970s [27] . Unlike other methods, the Williams Model was designed specifically to promote creative thinking using critical questioning [27] . This model involves the following aspects:

  • Flexibility
  • Elaboration
  • Originality
  • Risk taking
  • Imagination

Critical questions regarding fluency follow a sort of brainstorming approach in that the questions are designed to generates ideas and options [27] . For ‘flexibility’, the questions are designed to produce variations on existing ideas. ‘Elaboration’ questions are about building upon existing ideas and developing the level of detail. As the name suggests, critical questions for ‘originality’ are for promoting the development of new ideas. The ‘curiosity’ aspect of the Williams Model bears a similarity with that of the ‘Wonder’ stage of the Know Wonder Learn (KWL) system [33] . ‘Risk taking’ questions are designed to provoke experimentation. Although the name ‘complexity’ may sound similar to ‘elaboration’, it is instead about finding order among chaos, making connections, and filling in gaps of information. The final aspect is ‘Imagination’, which involves using questions to visualise.

Wiggins & McTighe’s Six Facets of Understanding

Wiggins & McTighe’s Six Facets of Understanding [ edit | edit source ]

Wiggins & McTighe’s ‘Six Facets of Understanding’ are all based on deep understanding aspects of critical thinking [34] . The method is used for teachers to design questions for students to promote critical thinking [34] . The six facets are Explanation, Interpretation, Application, Perspective, Empathy, and Self-Knowledge [35] .

‘Why’ and ‘How’ questions dominate the ‘Explanation’ facet in developing theory and reasoning [36] :

  • How did this happen? Why do you think this?
  • How does this connect to the other theory?

Interpretation questions encourage reading between the lines, creating analogies or metaphors, and creating written or visual scenarios to illustrate the idea. Questions include:

  • How would you explain this idea in other words?
  • Why do you think that there is conflict between the two sides?
  • Why is it important to know this?

Application questions are about getting students to use knowledge. Part of this comes from predicting what will happen based on prior experience. Another aspect involves learning from the past. Critical questions in this facet include:

  • How might we prevent this happening again?
  • What do you think will happen?
  • How does this work?

Perspective questions involves not only looking at ideas from other people’s perspectives, but also determining what people’s points of views are. In comparison with Empathy questions, though, Perspective questions involve more of an analytical and critical examination [35] . Here are some example questions:

  • What are the different points of view concerning this topic?
  • Whose is speaking in the poem?
  • Whose point of view is being expressed?
  • How might this look from the other person’s perspective?

Empathy questions involve perspective-taking, including empathy, in order to show an open mind to considering what it would feel like to walk in another person’s shoes.

  • How would you feel in the same situation?
  • What would it be like to live in those conditions?
  • How would you react if someone did that your family?

Self-knowledge questions are primarily designed to encourage self reflection and to develop greater self awareness [35] . In particular, Self-Knowledge questions reveal one’s biases, values, and prejudices and how they influence our judgment of others. Critical questions in this facet include:

  • How has my life shaped my view on this topic?
  • What do I really know about the lives of people in that community?
  • What knowledge or experience do I lack?
  • How do I know what I know? Where did that information / idea come from?

Questions within the Six Facets of Understanding all incorporate the following attributes [36] :

  • They are open ended
  • They require deep thought
  • They require critical thinking
  • They promote transfer of knowledge
  • They are designed to lead to follow-up questions
  • They require answers that are substantiated

For examples of critical questioning in action in a classroom environment, view the External Link section at the bottom of this page.

Problem Solving [ edit | edit source ]

In everyday life we are surrounded by a plethora of problems that require solutions and our attention to resolve them to reach our goals [37] . We may be confronted with problems such as: needing to determine the best route to get to work, what to wear for an interview, how to do well on an argumentative essay or needing to find the solution to a quadratic equation. A problem is present in situations where there is a desire to solve the problem, however the solution is not obvious to the solver [38] . Problem solving is the process of finding the solutions to these problems. [39] . Although they are related, critical thinking differs fundamentally from problem solving. Critical thought is actually a process that can be applied to problem solving. For example, students may find themselves engaging in critical thought when they encounter ill-defined problems that require them to consider many options or possible answers. In essence, those who are able to think critically are able to solve problems effectively [40] .

discuss the relationship between critical thinking and problem solving

This chapter on problem solving will first differentiate between Well-defined Problems and Ill-defined Problems , then explain uses of conceptualizing and visually representing problems within the context of problem solving and finally we will discuss how mental set may impede successful problem solving.

Well-defined and Ill-defined Problems [ edit | edit source ]

Problems can be categorized into two types: ill-defined or well-defined [37] Cognitive Psychology and Instruction (5th Ed). New York: Pearson.</ref> to the problem at hand. An example of a well-defined problem is an algebraic problem (ex: 2x - 29 = 7) where one must find the value of x. Another example may be converting the weight of the turkey from kilograms to pounds. In both instances these represent well-defined problems as there is one correct solution and a clearly defined way of finding that solution.

In contrast, ill-defined problems represent those we may face in our daily lives, the goals are unclear and they have information that is conflicting, incomplete or inconclusive [41] . An example of an ill-defined problem may be “how do we solve climate change?” or “how should we resolve poverty” as there is no one right answer to these problems. These problems yield the possibility to many different solutions as there isn’t a universally agreed upon strategy for solving them. People approach these problems differently depending on their assumptions, application of theory or values that they use to inform their approach [42] . Furthermore, each solution to a problem has its own unique strengths and weaknesses. [42] .

Table 1. Summarizes the difference between well-defined and ill-defined problems.

Differences in Solving Ill-defined and Well-defined Problems [ edit | edit source ]

In earlier times, researchers assumed both types of problems were solved in similar ways [44] , more contemporary research highlights some distinct differences between processes behind finding a solution.

Kitchener (1983) proposed that well-defined problems did not involve assumptions regarding Epistemological Beliefs [37] because they have a clear and definite solution, while ill-defined problems require these beliefs due to not having a clear and particular solution [45] . In support of this idea, Schraw, Dunkle and Bendixen conducted an experiment with 200 participants, where they found that performance in well-defined problems is not predictive of one's performance on ill-defined problems, as ill-defined problems activated different beliefs about knowledge. [46]

Furthermore Shin, Jonassen and McGee (2003), [43] found that solving ill-defined problems brought forth different skills than those found in well-structured problems. In well-structured problems domain knowledge and justification skills highly predicted problem-solving scores, whereas scores on ill-structured tasks were predictive of argumentation, attitudes and metacognition in an astronomy simulation.

Aligned with these findings, Cho and Jonassen (2002) [47] found that groups solving ill-structured problems produced more argumentation and problem solving strategies due to the importance of considering a wide variety of solutions and perspectives. In contrast, the same argumentation technique distracted the participant's activities when they dealt with well-defined problems. This research highlights the potential differences in the processes behind solving ill-defined and well-defined problems.

Implications Of The Classroom Environment [ edit | edit source ]

The fundamental differences between well-structured and ill-structured problems implicate that solving ill-structured problems calls for different skills, strategies, and approaches than well-structured problems [43] . Meanwhile, most tasks in the educational setting are designed around engaging learners in solving well-structured problems that are found at the end of textbook chapters or on standardized tests. [48] . Unfortunately the strategies used for well-defined problems have little application to ill-defined problems that are likely to be encountered day to day [49] as simplified problem solving strategies used for the well-structured designs have been found to have almost no similarities to real-life problems [48]

This demonstrates the need to restructure classrooms in a way that facilitates the student problem solving of ill-structured problems. One way we may facilitate this is through asking students questions that exemplify the problems found in everyday life [50] . This type of approach is called problem based learning and this type of classroom structure students are given the opportunity to address questions by collecting and compiling evidence, data and information from a plethora of sources [51] . In doing so students learn to analyze the information,data and information, while taking into consideration the vast interpretations and perspectives in order to present and explain their findings [51] .

Structure Of The Classroom [ edit | edit source ]

In problem-based learning, students work in small groups to where they explore meaningful problems, identify the information needed to solve the given problem, and devise effective approaches for the solution [50] . Students utilize these strategies, analyze and consider their results to devise new strategies until they have come up with an effective solution [50] . The teacher’s role in this classroom structure is to guide the process, facilitate participation and pose questions to elicit reflections and critical thinking about their findings [50] . In addition teachers may also provide traditional lectures and explanations that are intended to support student inquiry [50] .

In support of the argument to implement a problem-based approach to problem solving, a meta-analysis conducted by Dochy, Segers, Van den Bossche, & Gijbels (2003), found problem-based learning to be superior to traditional styles of learning though in supporting flexible problem solving, application of knowledge, and hypothesis generation. [52] Furthermore, Williams, Hemstreet, Liu, and Smith (1998) found that this approach fostered greater gains in conceptual understanding in science [53] . Lastly Gallagher, Stepien, & Rosenthal (1992), found that in comparing traditional vs. project-based approaches students in problem-based learning demonstrate an ability to define problems. [54] These findings highlight the benefits of problem-based learning on understanding and defining problems in science. Given the positive effects of defining problems this education approach may also be applied to our next sub-topic of conceptualizing problems.

Steps to Problem Solving [ edit | edit source ]

There have been five stages consistently found within the literature of problem solving: (1) identifying the problem, (2) representing the problem, (3) choosing the appropriate strategy, (4) implementing the strategy, and (5) assessing the solutions [37] . This overview will focus on the first two stages of problem solving and examine how they influence problem solving.

discuss the relationship between critical thinking and problem solving

Conceptualizing Problems [ edit | edit source ]

One of the most tedious and taxing aspects of problem solving is identifying the problem as it requires one to consider the problem through multiple lenses and perspectives without being attached to one particular solution to early on in the task [39] . In addition it is also important to spend time clearly identifying the problem due to the association between time spent "conceptualizing a particular problem and the quality of one's solutions". [37] For example consider the following problem:

Becka baked a chocolate cake in her oven for twenty five minutes. How long would it take her to bake three chocolate cakes?

Most people would jump to the conclusion to multiply twenty five by three, however if we place all three cakes in the oven at a time we find it would take the same time to bake three cakes as it would take to bake one. This example highlights the need to properly conceptualize the problem and look at it from different viewpoints, before rushing to solutions.

Taking this one step further, break down the five steps as the would be used to conceptualize the problem:

Stage 1 - Define the Problem

Stage 2 - Brainstorm Solutions

Stage 3 - Pick a Solution

Stage 4 - Implement the Solution

Stage 5 - Review the Result

Research also supports the importance of taking one's time to clearly identifying the problem before proceeding to other stages. In support of this argument, Getzel and Csikszentmihalyi found that artist students that spend more time identifying the problem when producing their art were rated as having more creative and original pieces than artists who spent less time at this stage [37] . These researchers postulated that in considering a wider scope of options during this initial stage they were able to come up with more original and dynamic solutions.

Furthermore, when comparing the approaches of experienced teachers and novice post-secondary students studying to be teachers, it was found that experienced teachers spent a greater amount of time lesson planning in comparison to post-secondary students when in a placed in a hypothetical classroom. [37] In addition these teachers offered significantly more solutions to problems posed in both ill-defined and well-defined problems. Therefore it is implicated that successful problem solving is associated with the time spent finding the correct problem and the consideration of multiple solutions.

Instructional Implications [ edit | edit source ]

One instructional implication we may draw from the literature that supports that the direct relationship between time spent on conceptualizing a problem and the quality of the solution, is that teachers should encourage students to spend as much time as possible at this stage [37] . In providing this knowledge and by monitoring student’s problem solving processes to ensure that they “linger” when conceptualizing problems, we may facilitate effective problem solving [37] .

Representing the Problem [ edit | edit source ]

Problem Representation refers to how the known information about a particular problem is organized [37] . In abstract representation of a problem, we merely think or speak about the problem without externally visually representing [37] . In representing a problem tangibly this is done by creating a visual representation on paper, computer, etc. of the data though graphs, stories, symbols, pictures or equations. These visual representations [37] may be helpful they can help us keep track of solutions and steps to a problem, which can particularly be useful when encountering complex problems.

discuss the relationship between critical thinking and problem solving

For example if we look at Dunker's Buddhist Monk example [37]  :

In the morning a Buddhist monk walks outside at sunrise to climb up the mountain to get to the temple at the peak. He reaches the temple just prior to sunset. A couple days later, he departs from the temple at sunrise to climb back down the mountain, travelling quicker than he did during his ascent as he is going down the mountain. Can you show a location along the path that the monk would have passed on both at the exact time of the day? [37]

In solely using abstraction, this problem is seemingly impossible to solve due to the vast amount of information, how it is verbally presented and the amount of irrelevant information present in the question. In using a visual representation we are able to create a mental image of where the two points would intersect and are better able to come up with a solution [55] .

Research supports the benefits of visual representation when confronted with difficult problems. Martin and Schwartz [56] found greater usage of external representations when confronted with a difficult task and they had intermittent access to resources, which suggests that these representations are used as a tool when problems are too complex without external aids. Results found that while creating the initial visual representation itself took up time, those who created these visual representations solved tasks with greater efficiency and accuracy.

Another benefit is that these visual representations may foster problem solving abilities by enabling us to overcome our cognitive biases. In a study conducted by Chambers and Reisberg [57] , participants were asked to look at the image below then close their eyes and form a mental image. When asked to recall their mental image of the photo and see if there were any alternate possibilities of what the photo could be, none of the participants were able to do so. However when participants were given the visual representation of the photo they were quickly able to manipulate the position of the photo to come up with an alternate explanation of what the photo could be. This shows how visual representations may be used in education by learners to counteract mental sets, which will be discussed in the next section.

As shown above, relying on abstraction can often overload one’s cognitive resources due to short- term memory being limited to seven items of information at a time [37] . Many problems surpass these limits disabling us being able to hold all the relevant information needed to solve a problem in our working memory [37] . Therefore it is implicated that in posing problems teachers should represent them written or visually in order to reduce the cognitive load. Lastly another implication is that as teachers we may increase problem-solving skills through demonstrating to students different types of external representations that can be used to show the relevant information pertaining to the problem. These representations may include different types of graphs, charts and imagery, which all can serve as tools for students in coming up with an effective solution, representing relevant information and reducing cognitive load

Challenges of Problem Solving [ edit | edit source ]

As discussed above there are many techniques to facilitate the problem solving process, however there are factors that can also hinder this process. For example: one’s past experiences can often impede problem solving as they can provide a barrier in looking at novel solutions, approaches or ideas [58] .

Mind set [ edit | edit source ]

A mind set refers to one's tendency to be influenced by one's past experiences in approaching tasks. [58] Mental set refers to confining ourselves to using solutions that have worked in the past rather than seeking out alternative approaches. Mental sets can be functional in certain situation as in using strategies that have worked before we are quickly able to come up with solutions. However, they can also eliminate other potential and more effective solutions.

discuss the relationship between critical thinking and problem solving

Functional Fixedness [ edit | edit source ]

Functional Fixedness is a type of mental set that refers to our tendency to focus on a specific function of an object (ie. what we traditionally use it for) while overlooking other potential novel functions of that object. [37]

A classic example of functional fixedness is the candle problem [59] . Consider you are at a table with a box full of tacks, one candle, and matches, you are then asked to mount the lit candle on the wall corkscrew board wall as quickly as possible, and make sure that this doesn't cause any wax to melt on the table. Due to functional fixedness you might first be inclined to pin the candle to the wall as that is what tacks are typically used for, similar to participants in this experiment. However, this is the incorrect solution as it would cause the wax to melt on the table.

The most effective solution requires you to view the box containing the tacks as a platform for the candle rather than it's traditional use as a receptacle. In emptying the box, we may use it as a platform for the candle and then use the tacks inside to attach the box to the wall. It is difficult to initially arrive at this solution as we tend to fixate on the function of the box of holding the tacks and have difficulty designating an alternate function to the box (ie. as a platform as opposed to a receptacle). This experiment demonstrates how prior knowledge can lead to fixation and can hinder problem solving.

Techniques to Overcome Functional Fixedness [ edit | edit source ]

As proposed by McCaffrey (2012), [60] one way to overcome functional fixedness is to break the object into parts. In doing so we may ask two fundamental questions “can it be broken down further” and “does my description of the part imply a use”. To explain this we can use McCaffrey’s steel ring figure-8 example. In this scenario the subject is given two steel rings, a candle and a match, they are asked to make the two steel rings into a figure 8. Looking at the tools provided to the subject they might decide that the wax from the candle could potentially hold the two pieces of steel together when heated up. However the wax would not be strong enough. It leaves them with a problem, how do they attach the two steel rings to make them a figure eight.

In being left with the wick as a tool, and labelling it as such we become fixated on seeing the primary function of the wick as giving off light, which hinders our ability to come up with a solution for creating a figure-8. In order to effectively solve problem we must break down our concept of the wick down further. In seeing a wick as just a waxed piece of string, we are able to get past functional fixedness and see the alternate functions of the string. In doing so we may come to the conclusion and see the waxed string as being able to be used to tie the two rings together. In showing the effectiveness of this approach McCaffrey (2012) found that people trained to use this technique solved 67% more problems than the control group [60] .

Given the effectiveness of this approach, it is implicated that one way we may promote Divergent Thinking is through teaching students to consider: "whether the object may be broken down further" [60] and "whether the description of the part imply a use" in doing so we may teach students to break down objects to their purest form and make salient the obscure features of a problem. This connects to the previously discussed idea of conceptualization where problem solving effectiveness can be increased through focusing time on defining the problem rather than jumping to conclusions based on our own preconceptions. In the following section we will discuss what strategies experts use when solving problems.

Novice Versus Expert In Problem Solving [ edit | edit source ]

Many researchers view effective problem solving as being dependent on two important variables: the amount of experience we have in trying to solve a particular category of problems [61] , which we addressed earlier by demonstrating that in practicing problem solving through engaging in a problem-based approach we may increase problem solving skills. However, the second factor to consider is the amount of domain-specific knowledge that we have to draw upon [61] . Experts possess a vast amount of domain knowledge, which allows them to efficiently apply their knowledge to relevant problems. Experts have a well-organized knowledge of their domain, which impacts they notice and how they arrange, represent and interpret information, this in turn enables them to better recall, reason and solve problems in comparison to novices. [62]

In comparing experts to novices in their problem strategies, experts are able to organize their knowledge around the deep structure in important ideas or concepts in their domain, such as what kind of solution strategy is required to solve the problem [63] . In contrast novices group problems based on surface structure of the problems, such as the objects that appear in the problem. [63]

Experts also spend more time than novices analyzing and identifying problems at the beginning of the problem-solving process. Experts take more time in thinking and planning before implementing solutions and use a limited set of strategies that are optimal in allowing them to richer and more effective solutions to the given problem. [64]

In addition experts will engage in deeper and more complete problem representation novices, in using external representations such as sketches and diagrams to represent information and solve problems. In doing so they are able to solve problems quicker and come up with better solutions. [65]

Given the literature above it is evident that problem solving and expertise overlap as the key strategies that experts utilize are also provided as effective problem solving strategies. Therefore, we may conclude that experts not only have a vast knowledge of their domain, they also know and implement the most effective strategies in order to solve problem more efficiently and effectively in comparison to novices. [65] In the next section we will discuss the connection between problem solving and critical thinking.

Cognitive Tutor for Problem Solving [ edit | edit source ]

Cognitive Tutor is a kind of Intelligent Tutoring Systems. [66] It can assign different problems to students according to their individual basis, trace users’ solution steps, provide just-in-time feedback and hint, and implement mastery learning criteria. [67]

According to Anderson and colleague, [67] the students who worked with LISP tutors completed the problems 30% faster and 43% outperformed than their peers with the help of teachers in mini-course. Also, college students who employed ACT Programming Tutor (APT) with the function of immediate feedback finished faster on a set of problems and 25% better on tests than the students who received the conventional instruction. [68] In addition, in high school geometry school settings, students who used Geometry Proof Tutor (GPT) for in- class problem solving had a letter grade scores higher than their peers who participated in traditional classroom problem-solving activities on a subsequent test. [69]

An overview of Cognitive Tutor [ edit | edit source ]

In 1985, Anderson, Boyle, and Reigser added the discipline of cognitive psychology to the Intelligent Tutoring Systems. Since then, the intelligent tutoring system adopted this approach to construct cognitive models for students to gain knowledge was named Cognitive Tutors. [67] The most widely used Cognitive Tutor is Cognitive Tutor® Algebra I. [69] Carnegie Learning, Inc., the trademark owner, is developing full- scale Cognitive Tutor®, including Algebra I, II, Bridge to Algebra, Geometry, and Integrated Math I, II, III. Cognitive Tutor® now includes Spanish Modules, as well.

Cognitive Tutors support the idea of learning by doing, an important part of human tutoring, which to provide students the performance opportunities to apply the objective skills or concepts and content related feedback. [69] To monitor students’ performance, Cognitive Tutors adopt two Algorithms , model tracing and knowledge tracing. Model tracing can provide immediate feedback, and give content-specific advice based on every step of the students’ performance trace. [67] Knowledge tracing can select appropriate tasks for every user to achieve mastery learning according to the calculation of one’s prior knowledge. [67] [69]

Cognitive Tutors can be created and applied to different curriculum or domains to help students learn, as well as being integrated into classroom learning as adaptive software. The curriculum and domains include mathematics in middle school and high school, [66] [68] [70] genetics in post-secondary institutions, [71] and programming. [67] [68] [72] [73]

Cognitive Tutors yielded huge impacts on the classroom, student motivation, and student achievement. [74] Regarding the effectiveness of Cognitive Tutors, research evidence supports more effectiveness of Cognitive Tutors than classroom instruction. [67] [75] [76] [68]

The Theoretical Background of Cognitive Tutor [ edit | edit source ]

Act-r theory [ edit | edit source ].

The theoretical background of Cognitive Tutors is ACT-R theory of learning and performance, which distinguishes between procedural knowledge and declarative knowledge. [67] According to the ACT-R theory, procedural knowledge cannot be directly absorbed into people’s heads, and it can be presented in the notation of if-then Production rules. The only way to acquire procedural knowledge is learning by doing.

Production rules [ edit | edit source ]

Production rules characterize how students, whether they beginning learners or advanced learners, think in a domain or subject. [67] Production rules can represent students' informal or intuitive thinking. [77] The informal or intuitive forms of thinking are usually different from what textbook taught, and students might gain such patterns of thinking outside from school. [78] Heuristic methods, such us providing a plan of actions for problem-solving instead of giving particular operation; [79] and non-traditional strategies, such as working with graphics rather than symbols when solving equation, [69] can be represented in production rules as well.

Cognitive model and model tracing [ edit | edit source ]


Cognitive model is constructed on both ACT-R theory and empirical studies of learners. [69] All the solutions and typical misconceptions of learners are represented in the production system of the cognitive model.

For example, there are three strategies of solving an algebra equation, 2(3+X)=10. Strategy 1 is multiplying 2 across the sum (3+X); Strategy 2 is dividing both sides of the equation by 2; Strategy 3 shows the misconception of failing to multiply 2 across the sum (3+X). Since there are various methods of each task, students can choose their way of solving problems.

Model tracing is an algorithm that can run forward along every student’s learning steps and provide instant context-specific feedback. If a student chooses the correct answer, for example, using strategy 1 or strategy 2 to solve the equation, the Cognitive Tutor® will accept the action and provide the student next task. If the student’s mistake match a common misconception, such as using strategy 3, the Cognitive Tutor will highlight this step as incorrect and provide a just-in- time feedback, such as you also need to multiply X by 2. If the student’s mistake does not match any of the production rule in the cognitive model, which means that the student does not use any of the strategies above, the Cognitive Tutor® will flag this step as an error in red and italicized. Students can ask for advice or hint any time when solving problems. According to Corbett, [68] there are three levels of advice. The first level is to accomplish a particular goal; the second level is to offer general ideas of achieving the goal, and the third level is to give students detailed advice on how to solve the problem in the current context.

Knowledge tracing [ edit | edit source ]

Knowledge tracing can monitor the growing number of production rules during the problem solving process. Every student can choose one production rule every step of his or her way of solving problems, and Cognitive Tutors can calculate an updated estimate of the probability of the student has learned the particular rule. [68] [69] The probability estimates of the rules are integrated into the interface and displayed in the skill-meter. Using probability estimates, the Cognitive Tutors can select appropriate tasks or problems according to students’ individual needs.

Effectiveness [ edit | edit source ]

Cognitive tutor® geometry [ edit | edit source ].

Aleven and Koedinger conducted two experiments to examine whether Cognitive Tutor® can scaffold self-explanation effectively in high school geometry class settings. [66] The findings suggested that “problem-solving practice with a Cognitive Tutor® is even more effective when the students explain their steps by providing references to problem-solving principles.” [80]

In geometry learning, it could happen when students have over-generalized production rules in their prior knowledge, and thus leading shallow encoding and learning. For instance, a student may choose the correct answer and go to next step base on the over-generalized production rule, if an angle looks equal to another, then it is , instead of real understanding. According to Aleven & Koedinger, self-explanation can promote more general encoding during problem-solving practice for it can push students to think more and reflect explicitly on the rules in the domain of geometry. [66]

All the geometry class in the experiments includes classroom discussion, small-group activities, lectures, and solving problems with Cognitive Tutor®. In both of the experiments, students are required to solve problems with the help of the Cognitive Tutor®. However, the Cognitive Tutor® were provided with two different versions, the new version can support self-explanation which is also called guided learning by doing and explaining, [66] and the other cannot. Theses additional features of the new version required students to justify each step by entering geometry principles or referring the principles to an online glossary of geometry knowledge, as well as providing explanations and solutions according to students’ individual choice. Also, the form of explanation in the new version is different from speech-based explanations mentioned in another experiment on self-explanation. The researchers found that students who use the new version of the Cognitive Tutor® were not only better able to give accurate explanation, but also able to deeper understand the domain rules. Thus, the students were able to transfer those learned rules to new situations better, avoiding shallow encoding and learning.

Genetics Cognitive Tutor [ edit | edit source ]

Corbett et al. (2010) conducted two evaluations of the Genetics Cognitive Tutor in seven different kinds of biology courses in 12 universities in America. The findings suggested the effectiveness of implementing Genetics Cognitive Tutor in post-secondary institution genetic problem-solving practice settings. [81]

In the first evaluation, the participants used the Genetics Cognitive Tutor with their class activities or homework assignments. The software has 16 modules with about 125 problems in five general genetic topics. Genetics Cognitive Tutor utilized the cognitive model of genetics problem solving knowledge to provide step-by-step help, and both model tracing and knowledge tracing. With the average correctness of pretest (43%) and post-test (61%), the average improvements of using Genetic Cognitive Tutors was 18%. In the second empirical evaluations, the researchers examined whether the knowledge tracing can correctly predict students’ knowledge. The finding suggested that the algorithm of knowledge tracing is capable of accurately estimating every student performance on the paper- and-pencil post-test.

Project Based Learning and Design Thinking [ edit | edit source ]

Theorizing solutions for real world problems [ edit | edit source ].

Project Based Learning is a concept that is meant to place the student at the center of learning. The learner is expected to take on an active role in their learning by responding to a complex challenge or question through an extended period of investigation. Project Based Learning is meant for students to acknowledge the curriculum of their class, but also access the knowledge that they already have to solve the problem challenge. At its roots, project-based learning is an activity in which students develop an understanding of a topic based on a real-life problem or issue and requires learners to have a degree of responsibility in designing their learning activity [82] . Blummenfeld et al. (1991) states that Project Based Learning allows students to be responsible for both their initial question, activities, and nature of their artifacts [83] .

Project based learning is based on five criteria [84]

discuss the relationship between critical thinking and problem solving

Challenges are based on authentic, real-world problems that require learners to engage through an inquiry process and demonstrate understanding through active or experiential learning. An example would be elementary or secondary students being asked by their teacher to solve a school problem – such as how to deal with cafeteria compost. Students would be encouraged to work in groups to develop solutions for this problem within specific criteria for research, construction, and demonstration of their idea as learners are cognitively engaged with subject matter over an extended period of time keeping them motivated [83] . The result is complex learning that defines its success is more than as more than the sum of the parts [85] . Project Based Learning aims at learners coordinating skills of knowledge, collaboration, and a final project presentation. This type of schema construction allows learners to use concrete training to perform concrete results. The learner uses previous knowledge to connect with new information and elaborate on their revised perception of a topic [85] . In Project Based Learning this would constitute the process of information gathering and discussing this information within a team to decide on a final solution for the group-instructed problem.

Unlike Problem-Based Learning, experiential learning within a constructivist pedagogy, is the basis of Project Based Learning, and learners show their knowledge, or lack there of, by working towards a real solution through trial and error on a specific driving question. The philosophy of Experiential experiential learning education comes from the theories developed by John Dewey in his work Education and Experience. Dewey argues that experience is shown to be a continuous process of learning by arousing curiosity, strengthen initiative, and is a force in moving the learner towards further knowledge [86] . The experiential aspect of Project Based Learning through working towards solutions for real world problems ties learner’s solutions to practical constructs. Learners must make up the expected gap in their knowledge through research and working together in a collaborative group. The experiential learning through Project Based Learning is focused on a driving question usually presented by the teacher. It is this focus that students must respond to with a designed artifact to show acquired knowledge.

The constructivist methodology of Project Based Learning is invoked through the guided discovery process set forth by the instructor, unlike pure discovery which has been criticised for student having too much freedom [87] , Project Based Learning involves a specific question driven by the instructor to focus the process of investigation. This form of constructivist pedagogy has shown to promote cognitive processing that is most effective in this type of learning environment [87] . Project Based Learning provides a platform for learners to find their own solutions to the teacher driven question, but also have a system in which to discover, analyze, and present. Therefore, Project Based Learning delivers beneficial cognitive meaningful learning by selecting, organizing, and integrating knowledge [87] .

Experience is the Foundation of Learning [ edit | edit source ]

Project Based Learning is a branch of education theory that is based on the idea of learning through doing. John Dewey indicated that teachers and schools should help learners to achieve greater depth in correlation between theory and real-world through experiential and constructivist methods. Dewey stated that education should contain an experiential continuum and a democratization of education to promote a better quality of human experience [86] . These two elements are consistent with Project Based Learning through the application of authentic, real world problems and production of artifacts as solutions, and the learner finding their own solutions through a collaborative effort with in a group. Blumenfeld et al. mentions that the value in Project Based Learning comes from questions that students can relate to including personal health and welfare, community concerns, or current events [83] .

Project Based Learning has basis also in the work of Jean Piaget who surmised that the learner is best served to learn in a constructivist manner – using previous knowledge as a foundation for new learning and connections. The learner’s intelligence is progressed from the assimilation of things in the learner’s environment to alter their original schema by accommodating multiple new schema and assimilating all of this experienced knowledge [88] . Piaget believed in the learner discovering new knowledge for themselves, but that without collaboration the individual would not be able to coherently organize their solution [87] . Project Based Learning acknowledges Piaget’s beliefs on the need for collective communication and its use in assembling new knowledge for the learner.

Self-Motivation Furthers Student Learning [ edit | edit source ]

Project Based Learning is perceived as beneficial to learners in various ways including gained knowledge, communication, and creativity. While engaging on a single challenge, learners obtain a greater depth of knowledge. Moreover, abilities in communication, leadership, and inter-social skills are strengthened due to the collaborative nature of Project Based Learning. Students retain content longer and have a better understanding of what they are learning. There are at least four strands of cognitive research to support Project Based Learning [84] – motivation, expertise, contextual factors, and technology.

Motivation of students that is centred on the learning and mastery of subject matter are more inclined to have sustained engagement with their work [89] . Therefore, Project Based Learning discourages public competition in favour of cooperative goals to reduce the threat to individual students and increase focus on learning and mastery [84] . Project Based Learning is designed to allow students to reach goals together, without fear of reprisal or individual criticism. For instance, Helle, et al. completed a study of information system design students who were asked to work on a specific assignment over a seven-month timeline. Students were given questionnaires about their experience during this assignment to determine their motivation level. Helle, et al. examined the motivation of learners in project groups and found intrinsic motivation increased by 0.52 standard deviations, showing that Project Based learner groups used self-motivation more often to complete assignments. Further, the study implied intrinsic motivation increase substantially for those who were lowest in self-regulation [90] .

Learner metacognitive and self-regulation skills are lacking in many students and these are important to master in student development in domains [84] . In the Project Based Learning system the relationship between student and teacher allows the instructor to use scaffolding to introduce more advance forms of inquiry for students to model, thus middle school students and older are very capable of meaningful learning and sophisticated results [91] . Learners would then become experts over time of additional skills sets that they developed on their own within this system.

Contextually, situated cognition is best realized when the material to be used resembles real-life as much as possible [84] , therefore, Project Based Learning provides confidence in learners to succeed in similar tasks outside of school because they no longer associate subjects as artificial boundaries to knowledge transfer. Gorges and Goke (2015) investigated the relationship between student perception of their abilities in major high school subjects and their relating these skills to real-world problem application through an online survey. Learners showed confidence in problem-solving skills and how to apply their learning to real-life situations, as Gorges and Goke [92] report, and that students who used Project Based Learning style learning have increased self-efficacy and self-concepts of ability in math (SD .77), history (SD .72), etc. [92] . Therefore, students are more likely to use domain-specific knowledge outside of an academic setting through increased confidence. Further, a comparison between students immediately after finishing a course and 12 weeks to 2 years provided effect sizes that showed Project Based Learning helped retain much knowledge [92] .

Technology use allows learners to have a more authentic experience by providing users with an environment that includes data, expanded interaction and collaboration, and emulates the use of artifacts [84] . The learner, in accessing technology, can enhance the benefits of Project Based Learning by having more autonomy is finding knowledge and connecting with group members. Creativity is enhanced as students must find innovative solutions to their authentic problem challenges. For instance, using digital-story-telling techniques through Project Based Learning, as stated by Hung and Hwang [93] , to collect data (photos) in elementary class to help answer a specific project question on global warming in science provided a significant increase in tests results (SD 0.64). As well, in order to find answers, learners must access a broad range of knowledge, usually crossing over various disciplines. The end result is that projects are resolved by student groups that use their knowledge and access to additional knowledge (usually through technology) to build a solution to the specific problem.

Educators Find Challenges in Project Based Learning Implementation [ edit | edit source ]

One of the main arguments against this type of learning is that the project can become unfocused and not have the appropriate amount of classroom time to build solutions. Educators themselves marginalized Project Based Learning because they lack the training and background knowledge in its implementation. Further financial constraints to provide effective evaluation through technology dissuades teachers as well [94] . The information gained by students could be provided in a lecture-style instruction and can be just as effective according to critics. Further, the danger is in learners becoming off-task in their time spent in the classroom, and if they are not continually focused on the task and the learning content, then the project will not be successful. Educators with traditional backgrounds in teaching find Project Based Learning requires instructors to maintain student connection to content and management of their time – this is not necessarily a style that all teachers can accomplish [94] .Blumenfeld et al. (1998) state that real success from Project Based Learning begins and ends with a focused structure that allows teacher modelling, examples, suggested strategies, distributing guidelines, giving feedback during the activity, and allowing for revision of work [91] .

Learner Need for Authentic Results through Critical Thought [ edit | edit source ]

discuss the relationship between critical thinking and problem solving

Project Based Learning is applicable to a number of different disciplines since it has various applications in learning, and is specifically relevant with the 21st century redefinition of education (differentiated, technologically-focused, collaboration, cross-curricular). STEM (Science, Technology, Engineering, Mathematics) is one form of 21st century education that benefits from instructors using Project Based Learning since it natural bridges between domains. The focus of STEM is to prepare secondary students for the rigors of post-secondary education and being able to solve complex problems in teams as would be expected when performing these jobs in the real world after graduation. Many potential occupational areas could benefit from Project Based Learning including medical, engineering, computer design, and education. Project Based Learning allows secondary students the opportunity to broaden their knowledge and become successful in high-stakes situation [95] . Moreover, these same students then develop a depth in knowledge when it comes to reflecting upon their strengths and limitations [95] . The result would be a learner who has developed critical thinking and has had a chance to apply it to real situations. Further the construction of a finished product is a realistic expectation in presenting an authentic result from learning. The product result demands accountability, and learner adherent to instructor expectations as well as constraints for the project [95] .

The learner is disciplined to focus on specific outcomes, understand the parameters of the task, and demonstrate a viable artifact. The implication is that students will be ready to meet the challenges of a high-technology, fast-paced work world where innovation, collaboration, and results-driven product is essential for success. Technology is one area where Project Based Learning can be applied by developing skills in real-world application, thus cognitive tools aforded by new technology will be useful if perceived as essential for the project (as is the case in many real-world applications) [83] .. For example, designers of computer systems with prior knowledge may be able to know how to trouble-shoot an operating system, but they do not really understand how things fit or work together, and they have a false sense of security about their skills [96] .

Design-Thinking as a Sub-set of Project-Based Learning [ edit | edit source ]

Using the process of practical design for real-world solutions [ edit | edit source ].

discuss the relationship between critical thinking and problem solving

Design Thinking is a pedagogical approach to teaching through a constructionist methodology of challenge-based problem solving branching off of Project Based learning. It should be understood as a combination of sub-disciplines having design as the subject of their cognitive interests [97] .

An example of design-thinking would be learners engaged with finding a solution to a real-world problem. However, unlike Project Based Learning, design-thinking asks the learner to create a practical solution within a scaffolding process (Figure 3) such as finding a method to deliver clean drinking water to a village. Designers would consider social, economic, and political considerations, but would deliver a final presentation of a working prototype that could be marketable. Hence a water system could be produced to deliver water to villagers, but within the limits of the materials, finances, and local policies in mind. It designates cores principles of empathy, define, ideate, prototype, and test to fulfill the challenges of design. Starting with a goal (solution) in mind, empathise is placed upon creative and practical decision making through design to achieve an improved future result. It draws upon a thinking that requires investigation into the details of a problem to find hidden parameters for a solution-based result. The achieved goal then becomes the launching point for further goal-setting and problem solving. [97]

This type of approach to education is based on the premise that the modern world is full of artificial constructs, and that our civilization historically has relied upon these artifacts to further our progress in technological advances. Herbert Simon, a founder of design-thinking, states that the world that students find themselves in today is much more man-made and artificial that it is a natural world [98] . The challenge of design-thinking is to foster innovation by enhancing student creative thinking abilities [99] . Design-thinking is a tool for scaffolding conventional educational projects into Project Based thinking. Van Merrienbroer (2004) views design-learning as a scaffolding for whole-task practice. It decreases intrinsic cognitive load while learners can practice on the simplest of worked-out examples [87] . Therefore, Design-thinking is currently becoming popular due to its ability to bridge between the justification of what the learner knows and what the learner discovers within the context of 21st century skills and learning. A further example of this process is the design of a product that children will use to increase their physical activity (see video on Design Thinking) and can be explained using the scaffold of Design Thinking:

Critical Thought on Design in the Artificial World [ edit | edit source ]

Design-thinking is can be traced back to a specific scholars including Herbert Simon, Donald Schon, and Nigel Cross. Simon published his findings on the gap he found in education of professions in 1969. He observed that techniques in the natural sciences and that just as science strove to show simplicity in the natural world of underlying complex systems, and Simon determined the it was the same for the artificial world as well [100] . Not only should this include the process behind the sciences, but the arts and humanities as well since music, for example involves formal patterns like mathematics (Simon, 136). Hence, the creative designs of everyone is based upon a common language and its application. While Schon builds upon the empathetic characteristics of design-thinking as a Ford Professor of Urban Planning and Education at MIT, referring to this process as an artistic and intuitive process for problem-solving [101] . Schon realized that part of the design process was also the reflection-in-action that must be involved during critical thinking and ideating. Moreover, the solutions for problems do not lie in text-books, but in the designer’s ability to frame their own understanding of the situation [100] . Cross fuses these earlier ideas into a pedagogy surrounding education stating that design-thinking should be part of the general education of both sciences and humanities [97] . He implies that students encouraged to use this style of thinking will improve cognitive development of non-verbal thought and communication [97] .

Critical Thinking as Disruptive Achievement [ edit | edit source ]

Design-thinking follows a specific flow from theoretical to practical. It relies upon guided learning to promote effective learner solutions and goes beyond inquiry which has been argued does not work because it goes beyond the limits of long-term memory [97] . Design-thinking requires the learner to have a meta-analysis of their process. Creativity (innovative thought) is evident in design thinking through studies in defocused and focused attention to stimuli in memory activation [97] . Hu et al. (2010) developed a process of disrupted thinking in elementary students by having them use logical methods of critical thought towards specific design projects, over a four-year period, through specific lesson techniques. The results show that these students had increased thinking ability (SD .78) and that these effects have a long-term transfer increasing student academic achievement [102] . This shows use of divergent and convergent thinking in the creative process, and both of these process of thought has been noted to be important in the process of creativity (Goldschmidt, 2016, p 2) and demonstrates the Higher Order Thinking that is associated with long-term memory. Design-thinking specifically demonstrates the capability of having learners develop

Designers are Not Scientific? [ edit | edit source ]

Design-thinking critics comment that design is in itself not a science or cognitive method of learning, and is a non-scientific activity due to the use of intuitive processes [97] . The learner is not truly involved within a cognitive practice (scientific process of reasoning). However, the belief of Cross is that design itself is a science to be studied, hence it can be investigated with systematic and reliable methods of investigation [97] . Further, Schon states that there is connection between theory and practice that in design thinking means that there is a loyalty to developing a theoretical idea into a real world prototype [101] . Design-thinking is a process of scientific cognitive practice that does constitute technical rationality [101] and using this practice to understand the limits of their design that includes a reflective practice and meta. Further, this pedagogy is the application for the natural gap between theory and practice for most ideas, by allowing the learner to step beyond normal instruction and practice to try something new and innovative to come up with a solution. Design-thinking rejects heuristically-derived responses based on client or expert appreciation to take on an unforeseen form [101] .

21st Century Learners and the Need for Divergent Thinking [ edit | edit source ]

Design-thinking is exceptionally positioned for use with 21st century skills based around technological literacy. Specifically, it is meant to assist the learner in developing creative and critical skills towards the application of technology. Designing is a distinct form of thinking that creates a qualitative relationship to satisfy a purpose [103] . Moreover, in a world that is rapidly becoming technologized, design-thinking the ability to make decisions based upon feel, be able to pay attention to nuances, and appraise the consequences of one’s actions [103] . The designer needs to be able to think outside the perceived acceptable solution and look to use current technology. Therefore, learners using design thinking are approaching all forms of technology as potential applications for a solution. Prototyping might include not just a hardware application, but also the use of software. Cutting-edge technologies such as Augmented Reality and Virtual Reality would be acceptable forms of solutions for design challenges. Specific application of design-thinking is, therefore applicable to areas of study that require technological adaptation and innovation. Specifically, the K-12 BC new curriculum (2016) has a specific focus on Applied Design, Skills, and Technologies that calls for all students to have knowledge of design-thinking throughout their entire education career and its application towards the advancement of technology. Therefore, Design Thinking is a relative and essential component to engaging student critical thought process.

Argumentation [ edit | edit source ]

Argumentation is the process of assembling and communicating reasons for or against an idea, that is, the act of making and presenting arguments. CT in addition to clear communication makes a good argument. It is the process through which one rationally solves problems, issues and disputes as well as resolving questions [104] .

The practice of argumentation consists of two dimensions: dialogue and structure [105] . The dialogue in argumentative discussions focus on specific speech acts – actions done through language (i.e. accept, reject, refute, etc.) – that help advance the speaker’s position. The structure of an argument helps distinguish the different perspectives in discussion and highlight positions for which speakers are arguing [105] .

One of the main arguments against this type of learning is that the project can become unfocused and not have the appropriate amount of classroom time to build solutions. Educators themselves marginalize PBL* because they lack the training and background knowledge in its implementation. Further financial constraints to provide effective evaluation through technology dissuades teachers as well (Efstratia, 2014, p 1258). The information gained by students could be provided in a lecture-style instruction and can be just as effective according to critics. Further, the danger is in learners becoming off-task in their time spent in the classroom, and if they are not continually focused on the task and the learning content, then the project will not be successful. Educators with traditional backgrounds in teaching find Project Based Learning requires instructors to maintain student connection to content and management of their time – this is not necessarily a style that all teachers can accomplish (Efstratia, 2014, p 1258).

Project Based Learning is applicable to a number of different disciplines since it has various applications in learning, and is specifically relevant with the 21st century redefinition of education (differentiated, technologically-focused, collaboration, cross-curricular). STEM (Science, Technology, Engineering, Mathematics) is one form of 21st century education that benefits from instructors using Project Based Learning since it natural bridges between domains. The focus of STEM is to prepare secondary students for the rigors of post-secondary education and being able to solve complex problems in teams as would be expected when performing these jobs in the real world after graduation. Many potential occupational areas could benefit from Project Based Learning including medical, engineering, computer design, and education.

Project Based Learning allows secondary students the opportunity to broaden their knowledge and become successful in high-stakes situation (Capraro, et al., 2013, p 2). Moreover, these same students then develop a depth in knowledge when it comes to reflecting upon their strengths and limitations (Capraro, et al., 2013, p 2). The result would be a learner who has developed critical thinking and has had a chance to apply it to real situations. Further the construction of a finished product is a realistic expectation in presenting an authentic result from learning. The product result demands accountability, and learner adherent to instructor expectations as well as constraints for the project (Capraro, et al., 2013, p 2). The learner is disciplined to focus on specific outcomes, understand the parameters of the task, and demonstrate a viable artifact. The implication is that students will be ready to meet the challenges of a high-technology, fast-paced work world where innovation, collaboration, and results-driven product is essential for success. Technology is one area where Project Based Learning can be applied by developing skills in real-world application. For example, designers of computer systems with prior knowledge may be able to know how to trouble-shoot an operating system, but they do not really understand how things fit or work together, and they have a false sense of security about their skills (Gary, 2013, p 1).

Design-thinking follows a specific flow from theoretical to practical. It relies upon guided learning to promote effective learner solutions and goes beyond inquiry which has been argued does not work because it goes beyond the limits of long-term memory (Lazonder and Harmsen, 2016, p 2). Design-thinking requires the learner to have a meta-analysis of their process. Creativity (innovative thought) is evident in design thinking through studies in defocused and focused attention to stimuli in memory activation (Goldschmidt, 2016, p 1). Hu et al. (2010) developed a process of disrupted thinking in elementary students by having them use logical methods of critical thought towards specific design projects, over a four-year period, through specific lesson techniques. The results show that these students had increased thinking ability (SD .78) and that these effects have a long-term transfer increasing student academic achievement (Hu, et al. 2010, p 554). This shows use of divergent and convergent thinking in the creative process, and both of these process of thought has been noted to be important in the process of creativity (Goldschmidt, 2016, p 2) and demonstrates the Higher Order Thinking that is associated with long-term memory. Design-thinking specifically demonstrates the capability of having learners develop.

The Process of Argumentation [ edit | edit source ]

Argumentation stages [ edit | edit source ].

The psychological process of argumentation that allows one the produce, analyze and evaluate arguments [106] . These stages will be discussed in more detail later in this chapter.

The Impact of Argumentation on Learning [ edit | edit source ]

Argumentation does not only impact the development of CT and vice versa, it affects many other aspects of learning as well. For instance, a study conducted in a junior high school science class showed that when students engaged in argumentation, they drew heavily on their prior knowledge and experiences [107] . Not only did argumentation enable the students to use their prior knowledge, it also helped them consolidate knowledge and elaborate on their understanding of the subject at a higher level [107] . These are just a few of the ways in which argumentation can be seen to impact aspects of learning other than the development of CT.

Video: Argumentation in Education: https://www.youtube.com/watch?v=YHm5xUZmCDg

The Relationship between Critical Thinking and Argumentation [ edit | edit source ]

Argumentation and CT appear to have a close relationship in instruction. Many studies have shown the impact that both of these elements can have on one another. Data suggests that when CT is infused into instruction it impacts the ability of students to argue [108] tasks that involve both critical thinking and creative thinking must be of an argumentative nature [109] , and that argument analysis and storytelling can improve CT [110] . In other words it would appear that both CT and argumentation impact the development of each other in students and that both impact other aspects of learning and cognition.

How Critical Thinking Improves Argumentation [ edit | edit source ]

CT facilitates the evaluation of the information necessary to make an argument. It aids in the judgement of the validity of each position. It is used to assess the credibility of sources and helps in approaching the issue from multiple points of view. The elements of CT and argumentation have many common features. For example, examining evidence and counter-evidence of a statement and the information that backs up these claims are both facets of creating a sound argument and thinking critically.

The impact of how CT explicitly impacts one’s ability to argue and reason with reference to the aforementioned four CT components will be examined in this section. First, there needs to be an examination of the aspects of CT and how they can be impacted by argumentation. The first component, knowledge, as stated by Bruning et. al (2011), actively shapes the way in which one resolves problems [111] . Therefore, it is essential that students have a solid foundation of knowledge of whatever it is that they are arguing. The ability to use well founded information in order to effectively analyze the credibility of new information is imperative for students who wish to increase their argumentative abilities. The second component of CT that is important for argumentation is inference . As Chesñevar and Simari (2007) discuss in their examination of how we develop arguments, inference and deduction are essential aspects of reaching new conclusions from knowledge that is already known or proven [112] .

discuss the relationship between critical thinking and problem solving

In other words, the ability to reach conclusions from known information is pivotal in developing and elaborating an argument. As well, the use of induction , a part of the CT process, is important to argumentation. As Bruning et al. suggest, the ability to make a general conclusion from known information is an essential part of the CT process [111] . Ontañón and Plaza (2015) make the argument that induction can be used in argumentation through communication with one another. Moreover, making inductions of general conclusions using the complete information that every member of the group can provide shows how interaction can be helpful through the use of induction in argumentation [113] . Therefore, it can be seen how induction, an important part of CT, can have a significant impact on argumentation and collaboration. The final component of CT, that may be the most important in its relationship to argumentation, is evaluation . The components of Evaluation indicated by Bruning et al. are analyzing, judging and weighing. These are three essential aspects of creating a successful argument [111] . Hornikx and Hahn (2012) provide a framework for three key elements of argumentation that are heavily attached in these Bruning et al.'s three aspects of CT [106] .

Production, Analysis, and Evaluation [ edit | edit source ]

The three aspects of argumentation that Hornikx and Hahn focus on in their research is the production , analysis and evaluation of arguments [106] . Producing an argument uses the key aspects of CT; there must be evaluation, analysis, judgement and weighing of the argument that one wishes to make a stand on. Analysis of arguments and analysis in CT go hand in hand, there must be a critical analysis of information and viewpoints in order to create a successful and fully supported argument. As well, evaluation is used similarly in argumentation as it is derived from CT. Assessing the credibility of sources and information is an essential part in finding articles and papers that can assist someone in making an informed decision. The final aspect of evaluation in critical thinking is metacognition, thinking about thinking or monitoring one's own thoughts [111] . Monitoring one's own thoughts and taking time to understand the rationality of the decisions that one makes is also a significant part of argumentation. According to Pinto et al.’s research, there is a strong correlation between one's argumentation ability and metacognition. [114] In other words, the ability to think about one’s own thoughts and the validity of those thoughts correlates positively with the ability to formulate sound arguments. The transfer of thoughts into speech/argumentation shows that CT influences argumentation dramatically, however some research suggests that the two interact in different ways as well. It can clearly be seen through the research presented that argumentation is heavily influenced by CT skills, such as knowledge, inference, evaluation and metacognition. However there are also strong implications that instruction of CT in a curriculum can bolster argumentation. A study conducted by Bensley et. al (2010) suggests that when CT skills are directly infused into a course compared to groups that received no CT instruction, those who received CT instruction showed significant gains in their ability of argument analysis [115] . There can be many arguments made for the implication of specific CT skills to impact argumentation, but this research shows that explicit teaching of CT in general can increase the ability of students to more effectively analyze arguments as well. This should be taken into account that Skills Programs mentioned later in this chapter should be instituted if teachers wish to foster argumentation as well as CT in the classroom.

How Argumentation Improves Critical Thinking [ edit | edit source ]

Argumentation is a part of the CT process, it clarifies reasoning and the increases one's ability to assess viable information. It is a part of metacognition in the sense that one needs to evaluate their own ideas. CT skills such as induction and/or deduction are used to create a structured and clear argument.

Research by Glassner and Schwarz (2007) shows that argumentation lies at the intersection of critical and creative thinking. They argue that reasoning, which is both critical and creative, is done through argumentation in adolescents. They suggest that reasoning is constantly being influenced by other perspectives and information. The ability to think creatively as well as critically about new information is managed by argumentation [116] . The back and forth process of accommodating, evaluating, and being open minded to new information can be argued as critical and creative thinking working together. However, the way in which one reaches conclusions from information is created from the ability to weigh this information, and then to successfully draw a conclusion regarding the validity of the solution that students come to. There is also a clear correlation of how argumentation helps students to nurture CT skills as well.

It is clear that CT can directly impact argumentation, but this relationship can also be seen as bidirectional, with argumentation instruction developing the CT skills. A study by Gold et al. shows that CT skills can be fostered through the use of argument analysis and storytelling in instruction [117] . This research suggests that argumentation and argument analysis are not only be beneficial to students, but also to older adults. This study was conducted using mature adult managers as participants. The article outlines four skills of CT that can be impacted by the use of argument analysis and storytelling: critique of rhetoric, tradition, authority, and knowledge. These four skills of CT are somewhat deeper than many instructed in high schools and extremely important to develop. The ability of argumentation to impact CT in a way that enables a person to gain a better perspective on their view about these things is essential to developing personal values as well as being able to use argumentation and CT to critique those values when presented with new information. The ability of argumentation to influence the ability of individuals to analyze their own traditions and knowledge is important for all students as it can give them better insight into what they value.

Argumentation is beneficial to CT skills as well as creative thinking skills in high school students. Research done by Demir and İsleyen (2015) shows that argumentation based a science learning approach in 9th graders improves both of types of thinking [118] . The ability of students to use argumentation to foster CT as well as creative thinking can be seen as being very beneficial, as mentioned earlier creative and CT skills use argumentation as a means of reasoning to draw conclusions, it is therefore not surprising that argumentation in instruction also fosters both of these abilities. In summation, it can clearly be seen that there is a link between both argumentation and CT along with many skills in the subset of CT skills. Explicit instruction of both of these concepts seems to foster the growth of the other and can be seen as complementary. In the next sections of this chapter how these aspects can be beneficial if taught within the curriculum and how they go hand in hand in fostering sound reasoning as well as skills that will help students throughout their lives will be examined.

Instructional Application of Argumentation and Critical Thinking [ edit | edit source ]

discuss the relationship between critical thinking and problem solving

Teaching Tactics [ edit | edit source ]

An effective method for structuring the instruction of CT is to organize the thinking skills into a clear and sequential steps. The order in which these steps aid in guiding the student towards internalizing those steps in order to apply them in their daily lives. By taking a deductive approach, starting from broader skills and narrowing them down to task-specific skills helps the student begin from what they know and generate something that they hadn't known before through CT. In the spirit of CT, a student's awareness of their own skills also plays an important role in their learning. In the classroom, they should be encouraged to reflect upon the process through which they completed a goal rather than just the result. Through the encouragement of reflection, students can become more aware of the necessary thinking skills necessary for tasks, such as Argumentation.

Instructing CT and Argumentation predisposes the instruction to using CT skills first. In designing a plan to teach CT, one must be able to critically evaluate and assess different methods and make an informed decision on which would work best for one's class. There are a variety of approaches towards instructing CT. Descriptive Models consist of explanations of how "good" thinking occurs. Specifically, it focuses on thinking strategies such as heuristics to assess information and how to make decisions. Prescriptive Models consist of explanations of what good thinking should be. In a sense, these models give a prototype, a "prescription", of what good thinking is. This approach is comparatively less applicable and sets a high standard of what is expected of higher order thinking. In addition to evaluating which approach would work best for them, prior to teaching CT, instructors need to carefully select the specific types of CT skills that they want students to learn. This process involves assessing factors such as age range, performance level as well as cognitive ability of one's class in order to create a program that can benefit most of, if not all, the students. A final aspect of instruction to consider as an educator is whether direct or indirect instruction will be used to teach CT. Direct Instruction refers to the explicit teaching of CT skills that emphasizes rules and steps for thinking. This is most effective when solutions to problems are limited or when the cognitive task is easy. In contrast, Indirect Instruction refers to a learner-oriented type of teaching that focuses on the student building their own understanding of thinking. This is most effective when problems are ambiguous, unclear or open to interpretation such as moral or ethical decisions [111] .

One example of indirect CT instruction is through the process of writing literature reviews. According to Chandler and Dedman, having the skills to collect, assess and write literature reviews as well as summarize results of studies requires CT. In a teaching note, they evaluated a BSW (Baccalaureate of Social Work) program that strived to improve CT in undergraduate students. Specifically, they assert that practical writing assignments, such as creating literature reviews, help students combine revision and reflection while expanding their thinking to evaluate multiple perspectives on a topic. They found that upon reframing the assignment as a tool to facilitate students in becoming critical reviewers, students viewed the literature review as a summation of course material in addition to an opportunity to improve critical reading and writing skills. Through questioning during discussions, students were guided to analyze the authority and credibility of their articles. The students actively sought for more evidence to support articles on their topics. They found that students successfully created well synthesized literature reviews at the end of the BSW program [119] . This program used implicit instruction of CT skills through dialogue between instructor and students as well as peer engagement. Instead of explicitly stating specific skills or steps to learn CT, the instructors lead the students to practice CT through an assignment. As students worked on the assignment, they needed to use reasoning, analysis and inferential skills in order to synthesize and draw conclusions around the evidence they found on their topics. Practical application of CT skills through an assignment helped students develop CT through indirect instruction.

discuss the relationship between critical thinking and problem solving

Argument mapping is a way to visualize argumentation. The following are links to argument mapping software: https://www.rationaleonline.com/ http://www.argunet.org/editor/ http://debategraph.org/planet https://www.truthmapping.com/map/1021/#s7164

Skills Programs for CT [ edit | edit source ]

These programs aid in the formulation of critical thinking skills through alternative methods of instruction such as problem-solving. They are usually targeted towards special populations such as students with learning disabilities or cognitive deficits.

The CoRT Thinking Materials [ edit | edit source ]

The CoRT (Cognitive Research Trust) program is based on de Bono’s idea that thinking skills should be taught in school as a subject [120] . The Thinking Materials are geared towards the improvement of thinking skills. This skills program takes on a Gestalt approach and emphasizes the perceptual factor of problem solving. It usually spans over the course of 2 years and is suitable for a wide age range of children. The lessons strive to develop creative thinking, problem-solving as well as interpersonal skills. The materials are split into 6 units and cover topics such as planning, analyzing, comparing, selecting, evaluating and generating alternatives. A typical unit has leaflets covering a single topic, followed by examples using practice items. The leaflets are usually effective in group settings. The focus of these units are to practice thinking skills, therefore much of the instructional time is spent on practicing the topics brought up in the leaflets [111] .

Much of the empirical research on this stand-alone program revolves around the development of creative thinking, however, it is relatively more extensive in comparison to the other programs mentioned in this chapter. The CoRT program has been shown to improve creativity in gifted students. Al-Faoury and Khwaileh (2014) assessed the effectiveness of the CoRT on gifted students’ creative writing abilities. The students were given a pretest that evaluated the fluency, flexibility and originality in writing creative short stories [120] . Students in the experimental group were taught 20 CoRT lessons in total with 10 from CoRT 1 “Breadth” and 10 from CoRT 4 “Creativity” over the course of three months while the control group received traditional lessons on creative writing. The posttest followed the same parameters as the pretest and the results were analyzed by comparing pre and posttest scores. The researchers found a statistically significant effect of CoRT on the experimental group’s fluency, flexibility and originality scores. The mean scores of the experimental groups in all three elements were higher than the control group [120] . These findings suggest that the CoRT program aids gifted students in creative writing skills as indicated through the use of rhetorical devices (metaphor, analogy, etc.), developing characters through dialogue and the control of complex structures [120] . The flexibility and fluency of writing is also applicable to the practice of argumentation and CT. In developing the ability to articulate and modify ideas, students can transfer these skills from creative writing towards higher-order cognitive processes such as CT and argumentation.

The Feuerstein Instrumental Enrichment Program (FIE) [ edit | edit source ]

The FIE is a specialized program focused on mediated learning experiences that strives to develop critical thinking and problem solving skills. Mediation is learning through interaction between the student and the mediator. Similar to Vygotsky's scaffolding, mediation is student-oriented and hinges upon 4 parameters: Intentionality, Reciprocity, Transcendence and Meaning. [121] Intentionality emphasizes the differences between mediation and interaction where the student and mediator have a common goal in mind. Reciprocity involves the student-oriented mentality of mediation, the response of the student hold most importance over academic results. Transcendence focuses on the connectivity of the mediation, it encourages the formation of associations and applications that stretch beyond the scope of the immediate material. Lastly, Meaning in mediation is where the student and mediator explicitly identify "why" and "what for" which promotes dialogue between the two during mediation. [121] [122]

The "instruments" used to facilitate instruction are a series of paper and pencil exercises geared towards practicing internalizing higher order thinking strategies. The instruments cover domains such as analytic perception, spatial organization, categorization, comparison and many more. The implementation of this program varies across countries and is also dependent on the targeted population. A typical program contains 14 units with 3-4 sessions for a few hours every week administered by trained IE staff and teachers. [121]

The Productive Thinking Program [ edit | edit source ]

The Productive Thinking Program consists of the development of planning skills, generating and checking hypotheses as well as creating new ideas. This program is designed as a set of 15 lessons aimed at being completed over one semester. The target population of the program is upper-level elementary school students. The lessons are administered through the use of narrative booklets, often taking a detective-like approach to problem solving where the student is the detective solving a mystery. A structured sequence of steps guides the student to attain an objective specific to the lesson at hand. [123] Following the booklet or story, supplementary problems are given in order for students to apply and practice learned skills. [111]

The IDEAL Problem Solver [ edit | edit source ]

The IDEAL Problem Solver structures problem-solving as 5 steps using the acronym IDEAL. First, (I)dentify the problem, the solver needs to find out what the problem is. Second, (D)efine the problem involves having a clear picture of the entire problem before trying to solve it. Third, (E)xplore the alternatives, meaning that the solver needs to assess the potential solutions available. Fourth, (A)cting on a plan, that is, applying the solution and doing the act of solving. Lastly, (L)ooking at the effects which encompasses the evaluation of the consequences of the chosen solution. IDEAL is flexible in that it can be adapted to suit a wide age range and different levels of ability in its application. It can also be applied to different domains such as composition or physics. [111]

Instructing Argumentation [ edit | edit source ]

Research on argumentation is a comparatively new field of study for education, but has been noted to be of significant importance to almost all educational settings. Grade schools, high schools, and colleges now emphasize the use of argumentation in the classroom as it is seen as the best way for communication and debate in a both vocational and educational settings around the world. [124] A longitudinal study done by Crowell and Kuhn showed that an effective way to help students gain argumentative skills was through consistent and dense application of argumentation in the classroom and as homework. [124] During this longitudinal study, students were exposed to a variety of different methods from which they gained argumentative abilities. The activities employed such as peer collaboration, using computers, reflection activities, individual essays, and small group work all have implications for being valuable in teaching argumentation although it is not clear which ones are the most effective. [124] Data also showed that students all rose to a similar level of argumentative ability, no matter what they scored on argumentative tests before the study began. This shows that even students with seemingly no argumentative skills can be instructed to become as skilled or more skilled than their peers who tested higher than them at the beginning of the study. [124]

Dialogue and Argumentation [ edit | edit source ]

Research by Crowell and Kuhn (2011) highlights collaborative dialogical activities as practical interventions in the development of argumentative skills. The researchers implemented a longitudinal argumentative intervention that used topic cycles to structure a middle school philosophy class [125] . The students had class twice a week for 50 minutes each class over the span of three years. The intervention is as follows: first, students were split into small groups on the same side of the argument to generate ideas around the topic (“for” and “against” teams). Then individuals from either side argue with an opponent through an electronic medium. Finally, the students engage in a whole class debate. These three stages were termed Pregame, Game and Endgame, respectively. After the intervention, students were required to write individual essays regarding the topic through which their argumentative skills would be assessed [125] . The results showed an increased in the generation of dual perspective arguments in the intervention group. Such arguments require the arguer to assume the opposing stance to one’s own and reason its implications. This type of argument reflects a higher-order reasoning that requires critical assessment of multiple perspectives. These results did not begin to appear until year two and was only found statistically significant in year three suggesting that argumentative skills have a longer development trajectory than other lower-level cognitive skills [125] . Through this stand-alone intervention, the collaborative aspect of dialogical activities facilitates the development of intellectual dispositions necessary for good argumentation [125] .

discuss the relationship between critical thinking and problem solving

Further research suggests that teaching through the use of collaborative discussions and argumentative dialogue is an effective teaching strategy [105] . Through argumentation, students can acquire knowledge of concepts as well as the foundational ideas behind these concepts. In formulating arguments, students need to generate premises that provide structure to an argument through accepted definitions or claims. Argumentation helps students reveal and clarify misconceptions as well as elaborate on background knowledge. The two aforementioned dimensions of argumentation – dialogue and structure – are often used in assessing and measuring argumentative performance [105] . Specifically, through student-expert dialogue, the students can be guided to give certain arguments and counterarguments depending on the expert’s dialectical decisions [105] . This scaffolding helps the student engage in more critical evaluations that delve deeper into the topic in discussion.

In a study using content and functional coding schemes of argumentative behavior during peer-peer and peer-expert dialogue pairings, Macagno, Mayweg-Paus and Kuhn (2014) found that through student-expert dialogues, students were able to later formulate arguments that dealt with abstract concepts at the root of the issue at hand (i.e. ethical principles, conflict of values) in comparison to peer-peer dialogues [105] . The expert used more specific and sophisticated ways of attacking the student’s argument, such as suggesting an alternative solution to the problem at hand, which in turn enhanced the performance of the student in later meta-dialogues [105] . The results suggest that the practical application of argumentation through collaborate activities facilitates the development of argumentation skills. Similar to CT skills development, rather than teaching, implicit instruction through the practice of argumentation in interactive settings helps its development.

Science and Argumentation [ edit | edit source ]

Much of the literature surrounding the application of argumentation in the classroom revolves around the scientific domain. Argumentation is often used as a tool in scientific learning to enhance CT skills, improve class engagement and activate prior knowledge and beliefs around the subject [105] . In order to articulate and refine scientific theories and knowledge, scientists themselves utilize argumentation [104] . Jonassen and Kim (2010) assert that science educators often emphasize the role of argumentation more than other disciplines [126] . Argumentation supports the learning of how to solve well-structures problems as well as ill-structured ones in science, and from there by extension, in daily life. Specifically, the ill-structured ones reflect more practical everyday problems where goals and limitations are unclear and there are multiple solution pathways as well as multiple factors for evaluating possible solutions [104] .

Through argumentation, students learn to use sound reasoning and CT in order to assess and justify their solution to a problem. For example, a well-structured problem would be one posed in a physics class where concrete laws and formulas dictate the solution pathway to a problem or review questions found at the end textbook chapters which require the application of a finite set of concepts and theories. An ill-structured problem would be finding the cause of heart disease in an individual. Multiple developmental and lifestyle factors contribute to this one problem in addition to the various different forms of heart disease that need to be evaluated. This sort of problem requires the application of knowledge from other domains such as nutrition, emotional well-being and genetics. Since ill-structured problems do not have a definite answer, students are provided with an opportunity to formulate arguments that justify their solutions [104] . Through the practice of resolving problems in science, such as these, students can use CT to develop their argumentative ability.

One’s willingness to argue as well as one's ability to argue also play a significant role in learning science [127] . For one science is at its core, extremely argumentative.

If students have to ability to engage in argumentation at an early age then there knowledge of specific content such as science can grow immensely. The main reason for this is argumentative discourse, being able to disagree with others is extremely important because for adolescents they are at an age which is fundamentally social (ie junior to senior high) using this social ability is pivotal as students at this point may have the confidence to disagree with one another. When a student disagrees with another in argument in a classroom setting it gives them an opportunity to explain the way in which they think about the material. This verbalization of one’s own thoughts and ideas on a subject can help with learning the subject immensely [127] . It also allows for the student to reflect upon and expand their ideas as they have to present them to the class which helps with learning. This also provides the opportunity for the student to identify any misconceptions they have about the subject at hand as more than likely they will receive rebuttal arguments from others in their class [127] . All these factors are aspects of CT and contribute to the learning of the concept and conceptual change in the student which is what learning is all about. The nature of adolescent social behaviour could provide a window through which argumentation could benefit their learning in dramatic ways in learning science [127] .

Argumentation, Problem Solving and Critical Thinking in History Education [ edit | edit source ]

History education offers learners an abundant opportunity to develop their problem solving and critical thinking skills while broadening their perspective on the human condition. The study of history addresses a knowledge gap; specifically, it is the difference between our knowledge of present day and the “infinite, unorganized and unknowable everything that ever happened”. [128] It has long been understood that the study of history requires critical thought and analytical problem-solving skills. In order to become proficient at the study of history, learners must interpret and construct how we come to know about the past and navigate the connection between the past and the body of knowledge we call history. [129] Unfortunately, history education has been demoted to simply recalling factual information - via the overuse of rote memorization and multiple-choice testing - all of which is placed outside the context of present day. This approach does little to inspire a love of history nor does it support the learner’s ability to construct an understanding of how the past and present are connected.

On the other hand, the study of science and mathematics has for many years been centred around developing skills through problem-solving activities. Students learn basic skills and build upon these skills through a progression of increasingly complex problems in order to further their understanding of scientific theory and mathematical relationships. Specific to science education, learners are taught to think like scientists and approach problems using the scientific method. If this approach works well for science and math education, why should it not be utilized for the teaching of history? [128] . Therefore, to develop historical thinking skills it is necessary for instructors to teach the strategies and problem-solving approaches that are used by professional historians. However, unlike science and mathematics, the problems we solve in history are often ill-defined and may be unanswerable in a definitive sense making it more challenging for students to learn and transfer these skills. The following section will address these challenges and provide support for teaching historical thinking via The Big Six Historical Thinking Concepts (2013).

Historical Thinking - The Big Six [ edit | edit source ]

Based upon years of research and first-hand classroom experience, Seixas and Morton (2013) established a set of six competencies essential to the development of historical thinking skills. Much like science and mathematics education discussed above, the Big Six approach to history education allows the learner to progress from simplistic to advanced tasks. Moreover, the Big Six approach is intended to help the learner “move from depending on easily available, commonsense notions of the past to using the culture’s most powerful intellectual tools for understanding history”. (pg 1) [128] Additionally, the Big Six concepts reveal to the learner the difficulties we encounter while attempting to construct a history of the past. The Big Six competencies include the following: historical significance, evidence, continuity and change, cause and consequence, historical perspectives, and the ethical dimension.

Historical Significance

To develop a critical view of history the learner must recognize and define the qualities that makes something (e.g., person, event, social change) historically significant and why they should spend their time learning about this thing. Behaviourist approaches to history education, focusing on the textbook as the main source of information, have caused learners to become passive in their approach to learning about the past. The textbook becomes the authority on what they need to know. Moreover, the sole use of textbooks to teach national history may contribute to the creation of a “master narrative” that limits a student’s access to what is controversial about their country’s past. [130] By shifting the focus away from the textbook, learners may be able to further their critical thinking skills by following the steps historians take to study the past and constructing their own “reasoned decisions about historical significance”. [128] However, even if a learner is provided primary source evidence to construct a narrative of the past but is not taught to recognize the subjective side to historical thinking - why these pieces of evidence were selected, why this topic was selected, and why they are both historically significant - they may not recognize the impacts of human motivation on the construction of historic understanding. Unlike scientific inquiry that relies on a “positivistic definition of rationality”, historical thinking requires learners to acknowledge human motivation - their own motivation in studying the past, their instructors motivation for selecting certain topics of study, and the motivation of those living in the past [131]

Seixas & Morton (2013) cite two elements involved in constructing historical significance: “big, compelling concerns that exist in our lives today, such as environmental sustainability, justice, power, [and] welfare” and “particular events, objects, and people whose historical significance is in question” (pg 16) [128] The intersection between these two elements is where historical significance is found. It is useful here to add Freedman’s (2015), definition of critical historical reasoning . Critical historical reasoning requires us to recognize that the study of history is not objective. Historians “frame their investigations through the questions they pose and the theories they advance” and therefore, learners of history must analyze the “integrity of historical narratives and their pattern of emphasis and omission” (pg 360). [131] Critical historical reasoning aims towards “conscious awareness of the frame one has adopted and the affordances and constraints it imposes” (pg 360) [131] . Therefore, both historians and learners of history must recognize that historical significance is assigned and not an inherent feature of the past, and, importantly, is subject to change.

The second set of competencies described by Seixas and Morton (2013) are based on using evidence to address an inquiry about the past. In a study of the cognitive processes involved in evaluating source documents, Wineburg (1991) lists three heuristics: corroboration, sourcing, and contextualization. Corroboration refers to comparing one piece of evidence to another, sourcing is identifying the author(s) of the evidence prior to reading or viewing the material, and contextualization refers to situating evidence in a specific time and place (pg 77). [132]

This study utilized an expert/novice design to compare how historians and high school students make sense of historic documents. Wineburg (1991) argues that the historians were more successful in the task not because of the “schema-driven processing” common to science and mathematics, but by building a model of the [historic] event through the construction of “context-specific schema tailored to this specific event” (pg 83). [132] Additionally, historians demonstrated greater appreciation for the source of the historic documents compared to the students. This suggests that the students did not make the connection between a document's author and the reliability of the source. As Wineburg states, the historian understands “that there are no free-floating details, only details tied to witnesses, and if witnesses are suspect, so are their details” (pg. 84). [132] This study suggests the potential for historical understanding to be improved by teaching the cognitive strategies historians use to construct history.

Multiple narratives of the past exist as individuals bring their own values and experiences to their interpretations of historical evidence. Recognizing this may push learners beyond accepting historic accounts at face value and pull them towards a more critical approach to history. Inquiry-based guided discovery activities, such as Freedman’s (2015) Vietnam war narrative study, suggest that students may gain an awareness of the way they and others “frame” history through exploring primary source documents and comparing their accounts with standardized accounts (i.e. a textbook). [133] By allowing learners to view history as an interpretation of evidence rather than a fixed body of knowledge, we can promote critical thought through the learners’ creation of inferences based on evidence and construction of arguments to support their inferences.

Continuity and Change

Developing an understanding of continuity and change requires the learner to recognize that these two elements overlap over the chronology of history; some things are changing at the same time that other things remain the same. If students are able to recognize continuity and the processes of change in their own lives they should be able to transfer this understanding to their study of the past. [134] Students should be encouraged to describe and question the rate and depth of historic change as well as consider whether the change should be viewed as progress or decline. [134] The evaluation of historic change as positive or negative is, of course, dependent on the perspective taken by the viewer. An example of continuity through history is the development of cultural identity. Carretero and van Alphen (2014), explored this concept in their study of master narratives in Argentinian high school students. They suggest that identity can be useful to facilitate history education, but could also create misconceptions by the learner confounding past with present (or, presentism), as demonstrated when using “we” to discuss people involved in victorious battles or revolutions of the past which gave shape to a nation (pg 308-309). [130] It is useful, then to teach students to differentiate between periods of history. However, periodization of history, much like everything else in the knowledge domain, is based on interpretation and is dependent on the questions historians ask [134]

Educational technology such as interactive timelines, narrative history games, and online discussion groups may help learners make connections between the past and present. For example, the Museum of Civilization offers a teaching tool on the history of Canadian medicare ( http://www.museedelhistoire.ca/cmc/exhibitions/hist/medicare/medic01e.shtml ). Interactive timelines allow students to see connections between continuity, change, cause, and consequences by visually representing where these elements can be found over historic time. Also, guiding the learners’ exploration of interactive timelines by selecting strong inquiry questions may improve students understanding and facilitate the development of historical thinking. For example, an investigation into the European Renaissance could be framed by the following question: “Did everyone in Europe experience the Renaissance the same way?” Questions such as this are open-ended so as to not restrict where the students takes their inquiry but also suggest a relationship between the changes of the Renaissance and the continuity of European society. Other examples of educational technology that support historical thinking include the “Wold History for us All” ( http://worldhistoryforusall.sdsu.edu/ ) project. This website offers world history units separated into large-scale and local-scale topics and organized by historic period. The lesson plans and resources may allow the learner to making connections between local issues and the broader, global conditions affecting world history. Finally, a case study by Blackenship (2009) suggests that online discussion groups are a useful for developing critical thinking by allowing the teacher to view the students’ thought processes and thereby facilitating formative assessment and informing the type of instructional interventions required by the teacher. Blackenship (2009) cites additional research supporting the use of online discussion because it allows the learners to collect their thoughts before responding to a discussion prompt; they have more time to access prior knowledge and consider their own ideas. [135]

Cause and Consequence

The historical thinking competencies of cause and consequence require learners to become proficient at identifying direct and indirect causes of historic events as well as their immediate and long-term consequences. Effective understanding of the causes of historic change requires the recognition of both the actions of individuals as well as the prevailing conditions of the time. Historical thinking requires students to go beyond simplistic immediate causes and think of history as web of “interrelated causes and consequences, each with various influences” (pg 110). [134] In addition to improving understanding of the past, these competencies may help learners to better understand present-day conflicts and issues. Shreiner (2014) used the novice/expert format to evaluate how people utilize their knowledge of history to make reasoned conclusions about events of the present. Similar to the Wineburg (1991) study discussed above, Shreiner (2014) found the experts were better at contextualizing and using sourcing to critically analyze documents for reliability and utility in establishing a reasoned judgement. Additionally, the study found that while students would use narrative to construct meaning, they typically created schematic narrative templates - general statements about the past which lack specific details & events. [136] Seixas and Morton (2013) caution the use of overly-simplistic timelines of history because they could create a misconception that history is nothing more than a list of isolated events.The study indicates that historical narratives that follow periodization schemes and are characterized by cause-and-effect relationships, as well as change over time, are helpful for understanding contemporary issues. [134] Therefore, it is important that educators work to develop these competencies in students. Much like historic change, the consequences of certain actions in history can be viewed as positive and negative, depending on perspective. This will be discussed in further detail below.

Historical Perspectives and Ethics

The final two historical thinking competencies proposed by Seixas and Morton are historical perspectives and ethics. Historical perspectives refers to analyzing the historical context for conditions that would influence a historic figure to view an event or act in a particular way. This could include religious beliefs, social status, geographic location, time period, prevailing economic and political conditions, and social/cultural conditions. This again requires some interpretation of evidence as oftentimes we do not have evidence that explicitly describes a historic figure’s attitudes and reasons for acting. Primary source documents, such as letters and journals can provide insight but still require the historian to use inference to make sense of the documents and connect the information to a wider historical narrative or biographical sketch of an individual. Additionally, “[h]ard statistics, such as birth and death rates, ages of marriage, literacy rates, and family size... can all help us make inferences about people's experiences, thoughts, and feelings” (pg 143). [134] There are, of course, limitations to how much we can infer about the past; however, Seixas and Morton (2013) suggest that acknowledging the limitations of what we can know about the past is part of “healthy historical thinking” (pg 143). [134] Learners can develop their understanding of historical perspective by observing the contrast between past and present ways of life and worldviews, identifying universal human traits that transcend time periods (e.g., love for a child), and avoiding presentism and anachronism . [134] A greater understanding of historical perspective will be useful for students when encountering conflicting historical accounts as they will be able to see where the historical actors are “coming from” and therefore better understand their actions. Historical perspective and ethics are related. Seixas and Morton (2013) argue that “the ethical dimension of historical thinking helps to imbue the study of history with meaning” (pg 170). [134] To understand the moral reasons for an individual's actions we need to understand the influence of historical, geographical, and cultural context. Additionally, to understand ethical consequences of the past we make moral judgments which require “empathetic understanding[;] an understanding of the differences between our moral universe and theirs” (Seixas and Peck, 2004, pg 113). [137] People with little experience with historical thinking have difficulty separating the moral standards of today’s society with the societies of the past. Additionally, students tend to judge other cultures more critically than their own; oftentimes defending or justifying actions of their own nations. [138] Therefore, Lopez, Carretero and Rodriguez-Moneo (2014) suggest using national narratives of nations different from the learner’s own nation to more effectively develop critical historical thinking. As the learner becomes proficient at analyzing the ethical decisions of the past, they can translate these skills to analyzing present-day ethical questions. Role playing is a useful instructional strategy for teaching historical perspective. Traditional, face-to-face classrooms allow for dramatic role play activities, debates, and mock trials where students can take on the role of an individual or social group from history. Additionally, educational games and websites allow for the integration of technology while using the role play strategy. Whitworth and Berson (2003) found that, in the 1990-2000s, technology in the social studies classroom was focused mostly on using the internet as a digital version of material that would have otherwise been presented in the classroom. They suggest that alternative uses of technology - such as inquiry-based webquests, simulations, and collaborative working environments - promote interaction and critical thinking skills. [139] One example of a learning object that promotes critical thinking through role playing is the Musee-Mccord’s online game collection ( http://www.mccord-museum.qc.ca/en/keys/games/ ). Specifically, the Victorian Period and the Roaring Twenties games allow the learner to progress through the time period and make decisions appropriate to the historic context of the period. These games are paired with relevant resources from the museum collections which can enhance the learner’s depth of understanding of the period. In terms of teaching strategies for the ethical component of history can be explored through historical narratives, debating ethical positions on historic events, and evaluating and critiquing secondary sources of information for ethical judgements.

To summarize, introducing professional historians’ strategies for studying history is widely regarded as a way to improve historical thinking in students. Professional historian’s cognitive processes of corroborating accounts, critically analyzing sources, and establishing historic context are reflected well by Seixas and Morton’s Big Six Historical Thinking Concepts (2013). Historical thinking gives students the skills to problem solve within the context of history and make sense of the past and connect it to the present in order to broaden the learner’s perspective, understand prevailing social conditions, and influence how they interact with the world. See the Historical Thinking Project’s webpage ( http://historicalthinking.ca/lessons ) for instructional ideas for all the historical competencies.

Instructing through Academic Controversy [ edit | edit source ]

Using the technique of Academic Controversy could be an effective way of teaching both argumentation and CT skills to students. Academic controversy involves dividing a cooperative group of four in two pairs of students and assigning them opposing positions of an argument or issue, after which the two pairs each argue for their position. The groups then switch their positions and argue again, finally the group of four is asked to come up with an all-around solution to the problem [140] . This activity can be effective in instructing both aspects of argumentation and CT, though it may be a bit dated. The activity is argumentative by nature, making students come up with reasons and claims for two sets of arguments. This equilibrium is important to the argumentative process because provides the students with an opportunity to evaluate the key points of their argument and the opposition's which could be beneficial in any debate. As well, this activity is geared to engage students in a few aspects of CT such as evaluation, since the students must assess each side of the argument. It also engages metacognitive processes as the students must come up with a synthesized conclusion with their peers of their own arguments, a process which requires them to be both analytical and open minded. This activity is a good way of increasing both CT skills and argumentation as it requires students to be open-minded, but also engage in analytical debate.

Glossary [ edit | edit source ]

Suggested readings [ edit | edit source ].

  • Abrami, P.C., Bernard, R.M., Borokhovski, E., Wade, A., Surkes, M.A., Tamim, R., & Zhang, D. (2008). Instructional Interventions Affecting Critical Thinking Skills and Dispositions: A Stage 1 Meta-Analysis. Review of Educational Research, 78(4). 1102-1134. DOI: 10.3102/0034654308326084.
  • Phan, H.P. (2010). Critical thinking as a self-regulatory process component in teaching and learning. Psicothema, 22(2). 284-292.
  • Kozulin, A. & Presseisen, B.Z. (1995). Mediated Learning Experience and Psychological Tools: Vygotsky’s and Feuerstein’s Perspective in a Study of Student Learning. Educational Psychologist, 30(2), 67-75.
  • Crowell, A., & Kuhn, D. (2011). Dialogic Argumentation as a Vehicle for Developing Young Adolescents’ Thinking. Psychological Science, 22(4), 545-552. DOI: 10.1177/0956797611402512.

External links [ edit | edit source ]

  • Critical Thinking: How Children Can Start Thinking Deeply, Part 1
  • Critical Thinking for Kids In Action, Part 2
  • Critical Thinking for Kids In Action, Part 3
  • Critical Thinking for Kids In Action, Part 4
  • Critical Thinking Exercises for Kids

References [ edit | edit source ]

  • ↑ Heijltjes, A., Van Gog, T., & Paas, F. (2014). Improving Students' Critical Thinking: Empirical Support for Explicit Instructions Combined with Practice. Applied Cognitive Psychology, 28(4), 518-530.
  • ↑ a b c d e f g Murphy, K. P., Rowe, M. L., Ramani, G., & Silverman, R. (2014). Promoting Critical-Analytic Thinking in Children and Adolescents at Home and in School. Educational Psychology Review, 26(4), 561-578.
  • ↑ Gick, M. L. (1986). Problem-Solving Strategies. Educational Psychologist, 21(1/2), 99-121.
  • ↑ a b c Ku, K. Y., Ho, I. T., Hau, K., & Lau, E. C. (2014). Integrating direct and Inquiry_Based Instruction in the teaching of critical thinking: An intervention study. Instructional Science, 42(2), 251-169.
  • ↑ a b c d e f g h i j k l m n o p q Mathews, S. R., & Lowe, K. (2011). Classroom environments that foster a Disposition for Critical Thinking . Learning Environments Research, 14(1), 59-73.
  • ↑ Glaser, E. M. (1941). An Experiment in the Development of Critical Thinking. Columbia University.
  • ↑ a b c d Phan, H.P. (2010). Critical thinking as a self-regulatory process component in teaching and learning. Psicothema, 22(2). 284-292.
  • ↑ a b c d e f Moon, J. (2007). Critical Thinking: An Exploration of Theory and Practice (1st ed.). London ; New York: Routledge.
  • ↑ a b Kurfiss, J. G. (1988). Critical Thinking: Theory, Research, Practice, and Possibilities: ASHE-ERIC/Higher Education Research Report, Volume 17, Number 2, 1988 (1st ed.). Washington, D.C: Jossey-Bass.
  • ↑ a b Board on Testing and Assessment, Division of Behavioral and Social Sciences and Education, & National Research Council. (2011). Assessing 21st Century Skills: Summary of a Workshop. National Academies Press.
  • ↑ a b Mason, M. (2009). Critical Thinking and Learning. John Wiley & Sons.
  • ↑ a b Elder, L., & Paul, R. (2009). The Art of Asking Essential Questions (5th Edition). Dillon Beach, CA: Foundation for Critical Thinking
  • ↑ a b c Paul, R., & Elder, L. (2007). The Thinker’s Guide to The Art of Socratic Questioning. Dillon Beach, CA: The Foundation for Critical Thinking.
  • ↑ a b c d Morgan, N., & Saxton, J. (2006). Asking Better Questions (2nd ed.). Markham, ON: Pembroke Publishers.
  • ↑ a b Cain, R. B. (2007). The Socratic Method: Plato’s Use of Philosophical Drama. A&C Black.
  • ↑ Harmon, D. A., & Jones, T. S. (2005). Elementary Education: A Reference Handbook. ABC-CLIO.
  • ↑ Stanley, T., & Moore, B. (2013). Critical Thinking and Formative Assessments: Increasing the Rigor in Your Classroom. Routledge.
  • ↑ a b Jung, I., Nishimura, M., & Sasao, T. (2016). Liberal Arts Education and Colleges in East Asia: Possibilities and Challenges in the Global Age. Springer.
  • ↑ Mason, M. (2009). Critical Thinking and Learning. John Wiley & Sons. p. 8.
  • ↑ Davies, M., & Barnett, R. (2015). The Palgrave Handbook of Critical Thinking in Higher Education. Springer.
  • ↑ a b c d Cibáková, D. (2015). Methods of developing critical thinking when working with educative texts. E-Pedagogium, (2), 135-145.
  • ↑ Garcia, T., & Pintrich, P. R. (1992). Critical Thinking and Its Relationship to Motivation, Learning Strategies, and Classroom Experience. 2-30.
  • ↑ a b Halpern, D. F. (2013). Thought and Knowledge: An Introduction to Critical Thinking. Psychology Press.
  • ↑ Browne, M. N., & Keeley, S. M. (2006). Asking the Right Questions: A Guide to Critical Thinking (8th ed.). Upper Saddle River, N.J: Prentice Hall.
  • ↑ Elder, L., & Paul, R. (2009). The Art of Asking Essential Questions (5th Edition). Dillon Beach, CA: Foundation for Critical Thinking. p. 3.
  • ↑ Micarelli, A., Stamper, J., & Panourgia, K. (2016). Intelligent Tutoring Systems: 13th International Conference, ITS 2016, Zagreb, Croatia, June 7-10, 2016. Proceedings. Springer.
  • ↑ a b c d e Conklin, W., & Teacher Created Materials. (2012). Strategies for Developing Higher-Order Thinking Skills, Grades 6 - 12. Shell Education.
  • ↑ http://www.janinesmusicroom.com/socratic-questioning-part-i-the-framework.html
  • ↑ Bloom, B. S., Krathwohl, D. R., & Masia, B. B. (1984). Taxonomy of educational objectives: the classification of educational goals. Longman.
  • ↑ Blosser, P. E. (1991). How to Ask the Right Questions. NSTA Press.
  • ↑ Wang, J.-F., & Lau, R. (2013). Advances in Web-Based Learning -- ICWL 2013: 12th International Conference, Kenting, Taiwan, October 6-9, 2013, Proceedings. Springer.
  • ↑ a b Gregory, G., & Kaufeldt, M. (2015). The Motivated Brain: Improving Student Attention, Engagement, and Perseverance. ASCD.
  • ↑ Carol, K., & Sandi, Z. (2014). Q Tasks, 2nd Edition: How to empower students to ask questions and care about answers. Pembroke Publishers Limited.
  • ↑ a b Doubet, K. J., & Hockett, J. A. (2015). Differentiation in Middle and High School: Strategies to Engage All Learners. ASCD.
  • ↑ a b c http://www.educ.kent.edu/fundedprojects/tspt/units/sixfacets.htm
  • ↑ a b McTighe, J., & Wiggins, G. (2013). Essential Questions: Opening Doors to Student Understanding (1st ed.). Alexandria, Va. USA: Association for Supervision & Curriculum Development.
  • ↑ a b c d e f g h i j k l m n o p q Bruning, G.J. Schraw & M.M. Norby (2011) Cognitive Psychology and Instruction (5th Ed). New York: Pearson.
  • ↑ Anderson, J. R. Cognitive Psychology and Its Implications. New York: Freeman, 1980
  • ↑ a b Mayer, R. E., & Wittrock, R. C. (2006). Problem solving. In P. A. Alexander & P. H. Winne (Eds.), Handbook of educational psychology (2nd ed., pp. 287–304). Mahwah, NJ: Erlbaum.
  • ↑ Snyder, M. J., & Snyder, L. G. (2008). Teaching critical thinking and Problem solving skills. Delta Pi Epsilon Journal, L(2), 90-99.
  • ↑ a b c Voss, J. F. (1988). Problem solving and reasoning in ill-structured domains. In C. Antaki (Ed.), Analyzing everyday explanation: A casebook of methods (pp. 74-93). London: SAGE Publications.
  • ↑ a b Pretz, J. E., Naples, A. J., & Sternberg, R. J. (2003). Recognizing, defining, and representing problems. In J. E. Davidson and R. J. Sternberg (Eds.), The psychology of problem solving (pp. 3–30). Cambridge, UK: Cam- bridge University Press.
  • ↑ a b c d Shin, N., Jonassen, D. H., & McGee, S. (2003). Predictors of Well-Structured and Ill-Structured Problem Solving in an Astronomy Simulation. Journal Of Research In Science Teaching, 40(1), 6-33.8
  • ↑ Simon, D. P. (1978). Information processing theory of human problem solving. In W. K. Estes (Ed.), Handbook of learning and cognitive process. Hillsdale, NJ: Lawrence Erlbau
  • ↑ Kitchener, K.S., Cognition, metacognition, and epistemic cognition. Human Development, 1983. 26: p. 222-232.
  • ↑ Schraw G., Dunkle, M. E., & Bendixen L. D. (1995). Cognitive processes in well-structured and ill-structured problem solving. Applied Cognitive Psychology, 9, 523–538.
  • ↑ Cho, K. L., & Jonassen, D. H. (2002) The effects of argumentation scaffolds on argumentation and problem solving. Educational Technology: Research & Development, 50(3), 5-22.
  • ↑ a b Jonassen, D. H. (2000). Revisiting activity theory as a framework for designing student-centered learning environments. In D. H. Jonassen, S. M. Land, D. H. Jonassen, S. M. Land (Eds.) , Theoretical foundations of learning environments (pp. 89-121). Mahwah, NJ, US: Lawrence Erlbaum Associates Publishers.
  • ↑ Spiro, R. J., Feltovich, P. J., Jacobson, M. J., & Coulson, R. L. (1992). Cognitive flexibility, constructivism, and hypertext: Random access instruction for advanced knowledge acquisition in ill-structured domains. In T. M. Duffy, D. H. Jonassen, T. M. Duffy, D. H. Jonassen (Eds.) , Constructivism and the technology of instruction: A conversation (pp. 57-75). Hillsdale, NJ, England: Lawrence Erlbaum Associates, Inc.
  • ↑ a b c d e Barrows, H. S. (1996). “Problem-based learning in medicine and beyond: A brief overview.” In L. Wilkerson & W. H. Gijselaers (Eds.), Bringing Problem-Based Learning to higher education: Theory and practice (pp. 3-12). San Francisco: Jossey Bass.
  • ↑ a b (Barron, B., & Darling-Hammond, L. (2008). Teaching for meaningful learning: A review of research on inquiry-based and cooperative learning. Powerful Learning: What We Know About Teaching for Understanding (pp. 11-70). San Francisco, CA: Jossey-Bass.)
  • ↑ Dochy, F., Segers, M., Van den Bossche, P., & Gijbels, D. (2003). Effects of problembased learning: A meta-analysis. Learning and Instruction, 13, 533–568.
  • ↑ Williams, D., Hemstreet, S., Liu, M.& Smith, V. (1998). Examining how middle school students use problem-based learning software. Unpublished paper presented at the ED-Media/ED Telecom ‘98 world Conference on Educational Multimedia and Hypermedia & World Conference on Educational Telecommunications, Freiberg, Germany.
  • ↑ Gallagher, S. A., Stepien, W. J., & Rosenthal, H. (1992). The effects of problem based learning on problem solving. Gifted Child Quarterly, 36, 195–200. Gertzman, A., & Kolodner, J. L.
  • ↑ Bruning, G.J. Schraw & M.M. Norby (2011) Cognitive Psychology and Instruction (5th Ed). New York: Pearson.
  • ↑ Martin, L., and D. L. Schwartz. 2009. “Prospective Adaptation in the Use of External Representations.” Cognition and Instruction 27 (4): 370–400. doi:10.1080/
  • ↑ Chambers, D., and D. Reisberg. 1985. “Can Mental Images Be Ambiguous?” Journal of Experimental Psychology: Human A pragmatic perspective on visual representation and creative thinking Perception and Performance 11 (3): 317–328.
  • ↑ a b Öllinger, M., Jones, G., & Knoblich, G. (2008). Investigating the effect of mental set on insight problem solving. Experimental Psychology, 55(4), 269-282. doi:10.1027/1618-3169.55.4.269
  • ↑ Duncker, K. (1945). On Problem Solving. Psychological Monograph, Whole No. 270.
  • ↑ a b c McCaffrey, T. (2012). Innovation relies on the obscure: A key to overcoming the classic problem of functional fixedness. Psychological Science, 23(3), 215-218.
  • ↑ a b Taconis, R., Ferguson-Hessler, M. G. M., & Broekkamp, H. (2002). Teaching science problem solving: An overview of experimental work. Journal of Research in Science Teaching, 38, 442–46
  • ↑ Bransford, J. D., Brown, A. L., & Cocking, R. R. (Eds.). (2000). How people learn: Brain, mind, experience, and school. Washington, DC: National Academies Press.
  • ↑ a b Novick, L. R., & Bassok, M. (2005). Problem solving. In K. Holyoak & R. Morrison (Eds.), The Cambridge handbook of thinking and reasoning (pp. 321–350). Cambridge, UK: Cambridge University Press.
  • ↑ Fuchs, L. S., Fuchs, D., Stuebing, K., Fletcher, J. M., Hamlett, C. L., & Lambert, W. (2008). Problem solving and computational skills: Are they shared or distinct aspects of mathematical cognition? Journal of Educa- tional Psychology, 100, 30–
  • ↑ a b McNeill, K., & Krajcik, J. (2008). Scientific explanations: Characterizing and evaluating the effects of teachers’ instructional practices on student learning. Journal of Research in Science Teaching, 45, 53–78.
  • ↑ a b c d e Aleven, V. A. W. M. M., & Koedinger, K. R. (2002) An effective meta-cognitive strategy: learning by doing and explaining with a computer-based Cognitive Tutor. Cognitive Science, 26(2), 147–179.
  • ↑ a b c d e f g h i Anderson, J. R., Corbett, A. T., Koedinger, K. R., & Pelletier, R. (1995). Cognitive Tutors: Lessons learned. Journal of the Learning Sciences, 4(2), 167.
  • ↑ a b c d e f Corbett, A. T., & Anderson, J. R. (2001). Locus of feedback control in computer-based tutoring: Impact on learning rate, achievement and attitudes. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 245–252). New York, NY, USA: ACM.
  • ↑ a b c d e f g Koedinger, K. R., & Corbett, A. (2006). Cognitive tutors. The Cambridge handbook of the learning sciences, 61-77.
  • ↑ Koedinger, K. R. (2002). Toward evidence for instructional design principles: Examples from Cognitive Tutor Math 6. In Proceedings of PME-NA XXXIII (The North American Chapter of the International Group for the Psychology of Mathematics Education).
  • ↑ Corbett, A., Kauffman, L., Maclaren, B., Wagner, A., & J,.ones, E. (2010). A Cognitive Tutor for genetics problem solving: Learning gains and student modeling. Journal of Educational Computing Research, 42(2), 219–239.
  • ↑ Corbett, A. T., & Anderson, J. R. (2008). Knowledge decomposition and sub-goal reification in the ACT programming tutor. Department of Psychology, 81.
  • ↑ Corbett, A. T., & Bhatnagar, A. (1997). Student modelling in the ACT programming tutor: Adjusting a procedural learning model with declarative knowledge. In User modelling (pp. 243-254). Springer Vienna.
  • ↑ Corbett, A. (2002). Cognitive tutor algebra I: Adaptive student modelling in widespread classroom use. In Technology and assessment: Thinking ahead. proceedings from a workshop (pp. 50-62).
  • ↑ Koedinger, K. R. & Anderson, J. R. (1993). Effective use of intelligent software in high school math classrooms. In Proceedings of the World Conference on Artificial Intelligence in Education, (pp. 241-248). Charlottesv
  • ↑ Koedinger, K., Anderson, J., Hadley, W., & Mark, M. (1997). Intelligent tutoring goes to school in the big city. Human-Computer Interaction Institute.
  • ↑ Koedinger, K. R., & Corbett, A. (2006). Cognitive tutors. The Cambridge handbook of the learning sciences, 61-77.
  • ↑ Resnick, L. B. (1987). Learning in school and out. Educational Researcher, 16(9), 13-20.
  • ↑ Polya, G. (1957). How to Solve It: A New Aspect of Mathematical Method. (2nd ed.). Princeton, NJ: Princeton University Press.
  • ↑ <Aleven, V. A. W. M. M., & Koedinger, K. R. (2002). An effective metacognitive strategy: learning by doing and explaining with a computer-based Cognitive Tutor. Cognitive Science, 26(2), 147–179. p. 173
  • ↑ Corbett, A., Kauffman, L., Maclaren, B., Wagner, A., & Jones, E. (2010). A Cognitive Tutor for genetics problem solving: Learning gains and student modelling. Journal of Educational Computing Research, 42(2), 219–239.
  • ↑ Morgan, Alistar. (1983). Theoretical Aspects of Project-Based Learning in Higher Education. British Journal of Educational Technology, 14(1), 66-78.
  • ↑ a b c d Blumenfeld, Phyllis C., Elliot Soloway, Ronald W. MArx, Joseph S. Krajick, Mark Guzdial, Annemarie Palincsar.. (1991). Motivating Project-Based Learning: Sustaining the Doing, Supporting the Learning. Educational Psychologist, 26(3&4), 369-398.
  • ↑ a b c d e f Thomas, John W. (2000). A Review of Research on Project-Based Learning. San Rafael: Autodesk Press.
  • ↑ a b van Merriënboer, J. J. G., Clark, R. E., & de Croock, M. B. M. (2002). Blueprints for complex learning: The 4C/ID-model. Educational Technology Research and Development, 50(2), 39–61. doi:10.1007/bf02504993.
  • ↑ a b Dewey, J. (1938). Experience and education. New York: Macmillan.
  • ↑ a b c d e van Merrienboer, J. J. G., Paul A. Kirschner, & Liesbeth Kester. (2004). Taking the Load off a Learner’s Mind: Instructional Design for Complex Learning. Amsterdam: Open University of the Netherlands.
  • ↑ Piaget, Jean; Cook, Margaret (Trans). The construction of reality in the child. New York, NY, US: Basic Books The construction of reality in the child. (1954). xiii 386 pp. http://dx.doi.org/10.1037/11168-000 .
  • ↑ Ames, C. (1992). Classrooms: goals, structures, and student motivation. Journal of Educational Psychology, 84, 261-271.
  • ↑ Helle, Laura, Paivi Tynjara, Erkki Olkinora, Kristi Lonka. (2007). “Aint nothin’ like the real thing”. Motivation and study processes on a work-based project course in information systems design. British Journal of Educational Psychology, 70(2), 397-411.
  • ↑ a b Joseph Krajcik , Phyllis C. Blumenfeld , Ronald W. Marx , Kristin M. Bass , Jennifer Fredricks & Elliot Soloway (1998) Inquiry in Project-Based Science Classrooms: Initial Attempts by Middle School Students, Journal of the Learning Sciences, 7:3-4, 313-350, DOI: 10.1080/10508406.1998.9672057
  • ↑ a b c Gorges, Julia, Thomas Goke. (2015). How do I Know what I can do? Anticipating expectancy of success regarding novel academic tasks. British Journal of Educational Psyschology, 85(1), 75-90.
  • ↑ Hung, C.-M., Hwang, G.-J., & Huang, I. (2012). A Project-based Digital Storytelling Approach for Improving Students' Learning Motivation, Problem-Solving Competence and Learning Achievement. Educational Technology & Society , 15 (4), 368–379. 
  • ↑ a b Efstratia, Douladeli. (2014). Experiential education through project based learning. Procedia – Social and Behavioral Sciences . 152, 1256-1260.
  • ↑ a b c Capraro, R. M., Capraro, M. M., & Morgan, J. (2013). STEM Project-Based Learning: An Integrated Science, Technology, Engineering, and Mathematics (STEM) Approach (2nd Edition). New York, NY: Sense. 
  • ↑ Gary, Kevin. (2013), Project-Based Learning. Computer. (Vol 48:9). Tempe: Arizona State. 
  • ↑ a b c d e f g h Cross, N. (2007). Designerly ways of knowing . Basel, Switzerland: Birkha¨user.
  • ↑ Simon, H. A. (1996). The sciences of the artificial . Cambridge, MA: MIT Press.
  • ↑ Aflatoony, Leila & Ron Wakkary, (2015). Thoughtful Thinkers: Secondary Schooler’s Learning about Design Thinking. Simon Fraser University: Surrey, BC.
  • ↑ a b Koh, Joyce Hwee Ling, Chin Sing Chai, Benjamin Wong, & Huang-Yao Hong. (2015) Design Thinking for Education. Singapore: Springer Science + Business Media.
  • ↑ a b c d Schon, D. A. (1983). The reflective practitioner: How professionals think in action (Vol. 5126). New York, NY: Basic Books.
  • ↑ Hu, Weiping, Philip Adey, Xiaojuan Jia, Jia Liu, Lei Zhang, Jing Li, Xiaomei Dong. (2010). Effects of a “Learn to Think” Intervention programme on primary school students. British Journal of Educational Psychology . 81(4) 537-557.
  • ↑ a b Wells, Alastair. (2013). The importance of design thinking for technological literacy: a phenomenological perspective . International Journal Technol Des Educ. (23:623-636). DOI 10.1007/s10798-012-9207-7.
  • ↑ a b c d Jonassen, D.H., & Kim, B. (2010). Arguing to learn ad learning to argue: design justifications and guidelines. Education Technology & Research Development, 58(4), 439-457. DOI 10.1007/s11423-009-9143-8.
  • ↑ a b c d e f g h Macagno, F., Mayweg-Paus, W., & Kuhn, D. (2014). Argumentation theory in Education Studies: Coding and Improving Students’ Argumentative Strategies. Topoi, 34, 523-537.
  • ↑ a b c Hornikx, J., & Hahn, U. (2012). Reasoning and argumentation: Towards an integrated psychology of argumentation. Thinking & Reasoning, 18(3), 225-243. DOI: 10.1080/13546783.2012.674715.
  • ↑ a b Aufschnaiter, C., Erduran, S., Osborne, J., & Simon, S. (2008). Arguing to learn and learning to argue: Case studies of how students' argumentation relates to their scientific knowledge. Journal of Research in Science Teaching, 45(1), 101-131. doi:10.1002/tea.20213
  • ↑ Bensley, A., Crowe, D., Bernhardt, P., Buckner, C., & Allman, A. (2010). Teaching and assessing CT skills for argument analysis in psychology. Teaching of Psychology, 37(2), 91-96. doi:10.1080/00986281003626656
  • ↑ Glassner, A., & Schwarz, B. B. (2007). What stands and develops between creative and critical thinking? argumentation?. Thinking Skills and Creativity, 2(1), 10-18. doi:10.1016/j.tsc.2006.10.001
  • ↑ Gold J., Holman D., & Thorpe R. (2002). The role of argument analysis and story telling in facilitating critical thinking. Management Learning, 33(3), 371-388. doi:10.1177/1350507602333005
  • ↑ a b c d e f g h Bruning, R. H., Schraw, G. J., & Norby, M. M. (2011). Cognitive psychology and instruction (5th ed.) Pearson.
  • ↑ Chesñevar, I., & Simari, G. (2007). Modelling inference in argumentation through labelled deduction: Formalization and logical properties. Logica Universalis, 2007, Volume 1, Number 1, Page 93, 1(1), 93-124. doi:10.1007/s11787-006-0005-4
  • ↑ Ontañón, S., & Plaza, E. (2015). Coordinated inductive learning using argumentation-based communication. Autonomous Agents and Multi-Agent Systems, 29(2), 266-304. doi:10.1007/s10458-014-9256-2
  • ↑ Pinto, M., Iliceto, P., & Melagno, S. (2012). Argumentative abilities in metacognition and in metalinguistics: A study on university students. European Journal of Psychology of Education, 27(1), 35-58. doi:10.1007/s10212-011-0064-7
  • ↑ Bensley, A., Crowe, D., Bernhardt, P., Buckner, C., & Allman, A. (2010). Teaching and assessing critical thinking skills for argument analysis in psychology. Teaching of Psychology, 37(2), 91-96. doi:10.1080/00986281003626656
  • ↑ Demir, B., & İsleyen, T. (2015). The effects of argumentation based science learning approach on creative thinking skills of students. Educational Research Quarterly, 39(1), 49-82.
  • ↑ Chandler, S. & Dedman, D.E. (2012). Writing a Literature Review: An Essential Component of Critical Thinking. The Journal of Baccalaureate Social Work, 17. 160-165.
  • ↑ a b c d Al-Faoury, O.H., & Khwaileh, F. (2014). The Effect of Teaching CoRT Program No. (4) Entitles “Creativity” on the Gifted Learners’ Writing in Ein El-Basha Center for Gifted Students. Theory and Practice in Language Studies, 4(11), 2249-2257. doi:10.4304/tpls.4.11.2249-2257.
  • ↑ a b c Kozulin, A. & Presseisen, B.Z. (1995). Mediated Learning Experience and Psychological Tools: Vygotsky’s and Feuerstein’s Perspective in a Study of Student Learning. Educational Psychologist, 30(2), 67-75.
  • ↑ Presseisen, B.Z. & Kozulin, A. (1992). Mediated Learning – The Contributions of Vygotsky and Feuerstein in Theory and Practice.
  • ↑ Schuler, G. (1974). The Effectiveness of the Productive Thinking Program. Paper presented at the Annual Meeting of the American Educational Research Association. Retrieved from: http://www.eric.ed.gov/contentdelivery/servlet/ERICServlet?accno=ED103479 .
  • ↑ a b c d Crowell, A., & Kuhn, D. (2014). Developing dialogic argumentation skills: A 3-year intervention study. Journal of Cognition and Development, 15(2), 363-381. doi:10.1080/15248372.2012.725187
  • ↑ a b c d Crowell, A., & Kuhn, D. (2011). Dialogic Argumentation as a Vehicle for Developing Young Adolescents’ Thinking. Psychological Science, 22(4), 545-552. DOI: 10.1177/0956797611402512.
  • ↑ Jonassen, D.H., & Kim, B. (2010). Arguing to learn ad learning to argue: design justifications and guidelines. Education Technology & Research Development, 58(4), 439-457. DOI 10.1007/s11423-009-9143-8.
  • ↑ a b c d Bathgate, M., Crowell, A., Schunn, C., Cannady, M., & Dorph, R. (2015). The learning benefits of being willing and able to engage in scientific argumentation. International Journal of Science Education, 37(10), 1590-1612. doi:10.1080/09500693.2015.1045958
  • ↑ a b c d e Seixas, P., Morton, T., Colyer, J., & Fornazzari, S. (2013). The Big Six: Historical thinking Concepts. Toronto: Nelson Education.
  • ↑ Osborne, K. (2013). Forward. Seixas, P., Morton, T., Colyer, J., & Fornazzari, S. The Big Six: Historical thinking Concepts. Toronto: Nelson Education.
  • ↑ a b Carretero, M., & van Alphen, F. (2014). Do Master Narratives Change Among High School Students? A Characterization of How National History Is Represented. Cognition and Instruction, 32(3), 290–312. http://doi.org/10.1080/07370008.2014.919298
  • ↑ a b c Freedman, E. B. (2015). “What Happened Needs to Be Told”: Fostering Critical Historical Reasoning in the Classroom. Cognition and Instruction, 33(4), 357–398. http://doi.org/10.1080/07370008.2015.1101465
  • ↑ a b c Wineburg, S. S. (1991). Historical problem solving: A study of the cognitive processes used in the evaluation of documentary and pictorial evidence. Journal of Educational Psychology, 83(1), 73–87. http://doi.org/10.1037/0022-0663.83.1.73
  • ↑ Freedman, E. B. (2015). “What Happened Needs to Be Told”: Fostering Critical Historical Reasoning in the Classroom. Cognition and Instruction, 33(4), 357–398. http://doi.org/10.1080/07370008.2015.1101465
  • ↑ a b c d e f g h i Seixas, P., Morton, T., Colyer, J., & Fornazzari, S. (2013). The Big Six: Historical thinking Concepts. Toronto: Nelson Education.
  • ↑ Blackenship, W. (2009). Making connections: Using online discussion forums to engage students in historical inquiry. Social Education, 73(3), 127-130.
  • ↑ Shreiner, T. L. (2014). Using Historical Knowledge to Reason About Contemporary Political Issues: An Expert–Novice Study. Cognition and Instruction, 32(4), 313–352. http://doi.org/10.1080/07370008.2014.948680
  • ↑ Seixas, P., & Peck, C. (2004). Teaching Historical Thinking. Challenges and Prospects for Canadian Social Studies, 109–117.
  • ↑ Lopez, C., Carretero, M., & Rodriguez-Moneo, M. (2014). Telling a national narrative that is not your own. Does it enable critical historical consumption? Culture & Psychology , 20 (4 ), 547–571. http://doi.org/10.1177/1354067X14554156
  • ↑ Whitworth, S. A., & Berson, M. J. (2003). Computer technology in the social studies: An examination of the effectiveness literature (1996-2001). Contemporary Issues in Technology and Teacher Education [Online serial], 2(4). Retrieved from http://www.citejournal.org/volume-2/issue-4-02/social-studies/computer-technology-in-the-social-studies-an-examination-of-the-effectiveness-literature-1996-2001
  • ↑ Johnson, D. W., & Johnson, R. T. (1993). Creative and critical thinking through academic controversy. The American Behavioral Scientist, 37(1), 40-53. Retrieved from https://www.proquest.com/docview/1306753602

discuss the relationship between critical thinking and problem solving

  • Book:Cognition and Instruction

Navigation menu

SLO campus power restored at 8:13am

  • Cuesta College Home
  • Current Students
  • Student Success Centers

Study Guides

  • Critical Thinking

Decision-making and Problem-solving

Appreciate the complexities involved in decision-making & problem solving.

Develop evidence to support views

Analyze situations carefully

Discuss subjects in an organized way

Predict the consequences of actions

Weigh alternatives

Generate and organize ideas

Form and apply concepts

Design systematic plans of action

A 5-Step Problem-Solving Strategy

Specify the problem – a first step to solving a problem is to identify it as specifically as possible.  It involves evaluating the present state and determining how it differs from the goal state.

Analyze the problem – analyzing the problem involves learning as much as you can about it.  It may be necessary to look beyond the obvious, surface situation, to stretch your imagination and reach for more creative options.

seek other perspectives

be flexible in your analysis

consider various strands of impact

brainstorm about all possibilities and implications

research problems for which you lack complete information. Get help.

Formulate possible solutions – identify a wide range of possible solutions.

try to think of all possible solutions

be creative

consider similar problems and how you have solved them

Evaluate possible solutions – weigh the advantages and disadvantages of each solution.  Think through each solution and consider how, when, and where you could accomplish each.  Consider both immediate and long-term results.  Mapping your solutions can be helpful at this stage.

Choose a solution – consider 3 factors:

compatibility with your priorities

amount of risk

practicality

Keys to Problem Solving

Think aloud – problem solving is a cognitive, mental process.  Thinking aloud or talking yourself through the steps of problem solving is useful.  Hearing yourself think can facilitate the process.

Allow time for ideas to "gel" or consolidate.  If time permits, give yourself time for solutions to develop.  Distance from a problem can allow you to clear your mind and get a new perspective.

Talk about the problem – describing the problem to someone else and talking about it can often make a problem become more clear and defined so that a new solution will surface.

Decision Making Strategies

Decision making is a process of identifying and evaluating choices.  We make numerous decisions every day and our decisions may range from routine, every-day types of decisions to those decisions which will have far reaching impacts.  The types of decisions we make are routine, impulsive, and reasoned.  Deciding what to eat for breakfast is a routine decision; deciding to do or buy something at the last minute is considered an impulsive decision; and choosing your college major is, hopefully, a reasoned decision.  College coursework often requires you to make the latter, or reasoned decisions.

Decision making has much in common with problem solving.  In problem solving you identify and evaluate solution paths; in decision making you make a similar discovery and evaluation of alternatives.  The crux of decision making, then, is the careful identification and evaluation of alternatives.  As you weigh alternatives, use the following suggestions:

Consider the outcome each is likely to produce, in both the short term and the long term.

Compare alternatives based on how easily you can accomplish each.

Evaluate possible negative side effects each may produce.

Consider the risk involved in each.

Be creative, original; don't eliminate alternatives because you have not heard or used them before.

An important part of decision making is to predict both short-term and long-term outcomes for each alternative.  You may find that while an alternative seems most desirable at the present, it may pose problems or complications over a longer time period.

  • Uses of Critical Thinking
  • Critically Evaluating the Logic and Validity of Information
  • Recognizing Propaganda Techniques and Errors of Faulty Logic
  • Developing the Ability to Analyze Historical and Contemporary Information
  • Recognize and Value Various Viewpoints
  • Appreciating the Complexities Involved in Decision-Making and Problem-Solving
  • Being a Responsible Critical Thinker & Collaborating with Others
  • Suggestions
  • Read the Textbook
  • When to Take Notes
  • 10 Steps to Tests
  • Studying for Exams
  • Test-Taking Errors
  • Test Anxiety
  • Objective Tests
  • Essay Tests
  • The Reading Process
  • Levels of Comprehension
  • Strengthen Your Reading Comprehension
  • Reading Rate
  • How to Read a Textbook
  • Organizational Patterns of a Paragraph
  • Topics, Main Ideas, and Support
  • Inferences and Conclusions
  • Interpreting What You Read
  • Concentrating and Remembering
  • Converting Words into Pictures
  • Spelling and the Dictionary
  • Eight Essential Spelling Rules
  • Exceptions to the Rules
  • Motivation and Goal Setting
  • Effective Studying
  • Time Management
  • Listening and Note-Taking
  • Memory and Learning Styles
  • Textbook Reading Strategies
  • Memory Tips
  • Test-Taking Strategies
  • The First Step
  • Study System
  • Maximize Comprehension
  • Different Reading Modes
  • Paragraph Patterns
  • An Effective Strategy
  • Finding the Main Idea
  • Read a Medical Text
  • Read in the Sciences
  • Read University Level
  • Textbook Study Strategies
  • The Origin of Words
  • Using a Dictionary
  • Interpreting a Dictionary Entry
  • Structure Analysis
  • Common Roots
  • Word Relationships
  • Using Word Relationships
  • Context Clues
  • The Importance of Reading
  • Vocabulary Analogies
  • Guide to Talking with Instructors
  • Writing Help

Cuesta College Celebrates 59th Commencement Ceremony

Miossi art gallery presents annual student art showcase, cuesta college's book of the year presents acclaimed author myriam gurba.

Construction student

Build Your Future

Register for Summer and Fall

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • CBE Life Sci Educ
  • v.17(1); Spring 2018

Understanding the Complex Relationship between Critical Thinking and Science Reasoning among Undergraduate Thesis Writers

Jason e. dowd.

† Department of Biology, Duke University, Durham, NC 27708

Robert J. Thompson, Jr.

‡ Department of Psychology and Neuroscience, Duke University, Durham, NC 27708

Leslie A. Schiff

§ Department of Microbiology and Immunology, University of Minnesota, Minneapolis, MN 55455

Julie A. Reynolds

Associated data.

This study empirically examines the relationship between students’ critical-thinking skills and scientific reasoning as reflected in undergraduate thesis writing in biology. Writing offers a unique window into studying this relationship, and the findings raise potential implications for instruction.

Developing critical-thinking and scientific reasoning skills are core learning objectives of science education, but little empirical evidence exists regarding the interrelationships between these constructs. Writing effectively fosters students’ development of these constructs, and it offers a unique window into studying how they relate. In this study of undergraduate thesis writing in biology at two universities, we examine how scientific reasoning exhibited in writing (assessed using the Biology Thesis Assessment Protocol) relates to general and specific critical-thinking skills (assessed using the California Critical Thinking Skills Test), and we consider implications for instruction. We find that scientific reasoning in writing is strongly related to inference , while other aspects of science reasoning that emerge in writing (epistemological considerations, writing conventions, etc.) are not significantly related to critical-thinking skills. Science reasoning in writing is not merely a proxy for critical thinking. In linking features of students’ writing to their critical-thinking skills, this study 1) provides a bridge to prior work suggesting that engagement in science writing enhances critical thinking and 2) serves as a foundational step for subsequently determining whether instruction focused explicitly on developing critical-thinking skills (particularly inference ) can actually improve students’ scientific reasoning in their writing.

INTRODUCTION

Critical-thinking and scientific reasoning skills are core learning objectives of science education for all students, regardless of whether or not they intend to pursue a career in science or engineering. Consistent with the view of learning as construction of understanding and meaning ( National Research Council, 2000 ), the pedagogical practice of writing has been found to be effective not only in fostering the development of students’ conceptual and procedural knowledge ( Gerdeman et al. , 2007 ) and communication skills ( Clase et al. , 2010 ), but also scientific reasoning ( Reynolds et al. , 2012 ) and critical-thinking skills ( Quitadamo and Kurtz, 2007 ).

Critical thinking and scientific reasoning are similar but different constructs that include various types of higher-order cognitive processes, metacognitive strategies, and dispositions involved in making meaning of information. Critical thinking is generally understood as the broader construct ( Holyoak and Morrison, 2005 ), comprising an array of cognitive processes and dispostions that are drawn upon differentially in everyday life and across domains of inquiry such as the natural sciences, social sciences, and humanities. Scientific reasoning, then, may be interpreted as the subset of critical-thinking skills (cognitive and metacognitive processes and dispositions) that 1) are involved in making meaning of information in scientific domains and 2) support the epistemological commitment to scientific methodology and paradigm(s).

Although there has been an enduring focus in higher education on promoting critical thinking and reasoning as general or “transferable” skills, research evidence provides increasing support for the view that reasoning and critical thinking are also situational or domain specific ( Beyer et al. , 2013 ). Some researchers, such as Lawson (2010) , present frameworks in which science reasoning is characterized explicitly in terms of critical-thinking skills. There are, however, limited coherent frameworks and empirical evidence regarding either the general or domain-specific interrelationships of scientific reasoning, as it is most broadly defined, and critical-thinking skills.

The Vision and Change in Undergraduate Biology Education Initiative provides a framework for thinking about these constructs and their interrelationship in the context of the core competencies and disciplinary practice they describe ( American Association for the Advancement of Science, 2011 ). These learning objectives aim for undergraduates to “understand the process of science, the interdisciplinary nature of the new biology and how science is closely integrated within society; be competent in communication and collaboration; have quantitative competency and a basic ability to interpret data; and have some experience with modeling, simulation and computational and systems level approaches as well as with using large databases” ( Woodin et al. , 2010 , pp. 71–72). This framework makes clear that science reasoning and critical-thinking skills play key roles in major learning outcomes; for example, “understanding the process of science” requires students to engage in (and be metacognitive about) scientific reasoning, and having the “ability to interpret data” requires critical-thinking skills. To help students better achieve these core competencies, we must better understand the interrelationships of their composite parts. Thus, the next step is to determine which specific critical-thinking skills are drawn upon when students engage in science reasoning in general and with regard to the particular scientific domain being studied. Such a determination could be applied to improve science education for both majors and nonmajors through pedagogical approaches that foster critical-thinking skills that are most relevant to science reasoning.

Writing affords one of the most effective means for making thinking visible ( Reynolds et al. , 2012 ) and learning how to “think like” and “write like” disciplinary experts ( Meizlish et al. , 2013 ). As a result, student writing affords the opportunities to both foster and examine the interrelationship of scientific reasoning and critical-thinking skills within and across disciplinary contexts. The purpose of this study was to better understand the relationship between students’ critical-thinking skills and scientific reasoning skills as reflected in the genre of undergraduate thesis writing in biology departments at two research universities, the University of Minnesota and Duke University.

In the following subsections, we discuss in greater detail the constructs of scientific reasoning and critical thinking, as well as the assessment of scientific reasoning in students’ thesis writing. In subsequent sections, we discuss our study design, findings, and the implications for enhancing educational practices.

Critical Thinking

The advances in cognitive science in the 21st century have increased our understanding of the mental processes involved in thinking and reasoning, as well as memory, learning, and problem solving. Critical thinking is understood to include both a cognitive dimension and a disposition dimension (e.g., reflective thinking) and is defined as “purposeful, self-regulatory judgment which results in interpretation, analysis, evaluation, and inference, as well as explanation of the evidential, conceptual, methodological, criteriological, or contextual considera­tions upon which that judgment is based” ( Facione, 1990, p. 3 ). Although various other definitions of critical thinking have been proposed, researchers have generally coalesced on this consensus: expert view ( Blattner and Frazier, 2002 ; Condon and Kelly-Riley, 2004 ; Bissell and Lemons, 2006 ; Quitadamo and Kurtz, 2007 ) and the corresponding measures of critical-­thinking skills ( August, 2016 ; Stephenson and Sadler-McKnight, 2016 ).

Both the cognitive skills and dispositional components of critical thinking have been recognized as important to science education ( Quitadamo and Kurtz, 2007 ). Empirical research demonstrates that specific pedagogical practices in science courses are effective in fostering students’ critical-thinking skills. Quitadamo and Kurtz (2007) found that students who engaged in a laboratory writing component in the context of a general education biology course significantly improved their overall critical-thinking skills (and their analytical and inference skills, in particular), whereas students engaged in a traditional quiz-based laboratory did not improve their critical-thinking skills. In related work, Quitadamo et al. (2008) found that a community-based inquiry experience, involving inquiry, writing, research, and analysis, was associated with improved critical thinking in a biology course for nonmajors, compared with traditionally taught sections. In both studies, students who exhibited stronger presemester critical-thinking skills exhibited stronger gains, suggesting that “students who have not been explicitly taught how to think critically may not reach the same potential as peers who have been taught these skills” ( Quitadamo and Kurtz, 2007 , p. 151).

Recently, Stephenson and Sadler-McKnight (2016) found that first-year general chemistry students who engaged in a science writing heuristic laboratory, which is an inquiry-based, writing-to-learn approach to instruction ( Hand and Keys, 1999 ), had significantly greater gains in total critical-thinking scores than students who received traditional laboratory instruction. Each of the four components—inquiry, writing, collaboration, and reflection—have been linked to critical thinking ( Stephenson and Sadler-McKnight, 2016 ). Like the other studies, this work highlights the value of targeting critical-thinking skills and the effectiveness of an inquiry-based, writing-to-learn approach to enhance critical thinking. Across studies, authors advocate adopting critical thinking as the course framework ( Pukkila, 2004 ) and developing explicit examples of how critical thinking relates to the scientific method ( Miri et al. , 2007 ).

In these examples, the important connection between writing and critical thinking is highlighted by the fact that each intervention involves the incorporation of writing into science, technology, engineering, and mathematics education (either alone or in combination with other pedagogical practices). However, critical-thinking skills are not always the primary learning outcome; in some contexts, scientific reasoning is the primary outcome that is assessed.

Scientific Reasoning

Scientific reasoning is a complex process that is broadly defined as “the skills involved in inquiry, experimentation, evidence evaluation, and inference that are done in the service of conceptual change or scientific understanding” ( Zimmerman, 2007 , p. 172). Scientific reasoning is understood to include both conceptual knowledge and the cognitive processes involved with generation of hypotheses (i.e., inductive processes involved in the generation of hypotheses and the deductive processes used in the testing of hypotheses), experimentation strategies, and evidence evaluation strategies. These dimensions are interrelated, in that “experimentation and inference strategies are selected based on prior conceptual knowledge of the domain” ( Zimmerman, 2000 , p. 139). Furthermore, conceptual and procedural knowledge and cognitive process dimensions can be general and domain specific (or discipline specific).

With regard to conceptual knowledge, attention has been focused on the acquisition of core methodological concepts fundamental to scientists’ causal reasoning and metacognitive distancing (or decontextualized thinking), which is the ability to reason independently of prior knowledge or beliefs ( Greenhoot et al. , 2004 ). The latter involves what Kuhn and Dean (2004) refer to as the coordination of theory and evidence, which requires that one question existing theories (i.e., prior knowledge and beliefs), seek contradictory evidence, eliminate alternative explanations, and revise one’s prior beliefs in the face of contradictory evidence. Kuhn and colleagues (2008) further elaborate that scientific thinking requires “a mature understanding of the epistemological foundations of science, recognizing scientific knowledge as constructed by humans rather than simply discovered in the world,” and “the ability to engage in skilled argumentation in the scientific domain, with an appreciation of argumentation as entailing the coordination of theory and evidence” ( Kuhn et al. , 2008 , p. 435). “This approach to scientific reasoning not only highlights the skills of generating and evaluating evidence-based inferences, but also encompasses epistemological appreciation of the functions of evidence and theory” ( Ding et al. , 2016 , p. 616). Evaluating evidence-based inferences involves epistemic cognition, which Moshman (2015) defines as the subset of metacognition that is concerned with justification, truth, and associated forms of reasoning. Epistemic cognition is both general and domain specific (or discipline specific; Moshman, 2015 ).

There is empirical support for the contributions of both prior knowledge and an understanding of the epistemological foundations of science to scientific reasoning. In a study of undergraduate science students, advanced scientific reasoning was most often accompanied by accurate prior knowledge as well as sophisticated epistemological commitments; additionally, for students who had comparable levels of prior knowledge, skillful reasoning was associated with a strong epistemological commitment to the consistency of theory with evidence ( Zeineddin and Abd-El-Khalick, 2010 ). These findings highlight the importance of the need for instructional activities that intentionally help learners develop sophisticated epistemological commitments focused on the nature of knowledge and the role of evidence in supporting knowledge claims ( Zeineddin and Abd-El-Khalick, 2010 ).

Scientific Reasoning in Students’ Thesis Writing

Pedagogical approaches that incorporate writing have also focused on enhancing scientific reasoning. Many rubrics have been developed to assess aspects of scientific reasoning in written artifacts. For example, Timmerman and colleagues (2011) , in the course of describing their own rubric for assessing scientific reasoning, highlight several examples of scientific reasoning assessment criteria ( Haaga, 1993 ; Tariq et al. , 1998 ; Topping et al. , 2000 ; Kelly and Takao, 2002 ; Halonen et al. , 2003 ; Willison and O’Regan, 2007 ).

At both the University of Minnesota and Duke University, we have focused on the genre of the undergraduate honors thesis as the rhetorical context in which to study and improve students’ scientific reasoning and writing. We view the process of writing an undergraduate honors thesis as a form of professional development in the sciences (i.e., a way of engaging students in the practices of a community of discourse). We have found that structured courses designed to scaffold the thesis-­writing process and promote metacognition can improve writing and reasoning skills in biology, chemistry, and economics ( Reynolds and Thompson, 2011 ; Dowd et al. , 2015a , b ). In the context of this prior work, we have defined scientific reasoning in writing as the emergent, underlying construct measured across distinct aspects of students’ written discussion of independent research in their undergraduate theses.

The Biology Thesis Assessment Protocol (BioTAP) was developed at Duke University as a tool for systematically guiding students and faculty through a “draft–feedback–revision” writing process, modeled after professional scientific peer-review processes ( Reynolds et al. , 2009 ). BioTAP includes activities and worksheets that allow students to engage in critical peer review and provides detailed descriptions, presented as rubrics, of the questions (i.e., dimensions, shown in Table 1 ) upon which such review should focus. Nine rubric dimensions focus on communication to the broader scientific community, and four rubric dimensions focus on the accuracy and appropriateness of the research. These rubric dimensions provide criteria by which the thesis is assessed, and therefore allow BioTAP to be used as an assessment tool as well as a teaching resource ( Reynolds et al. , 2009 ). Full details are available at www.science-writing.org/biotap.html .

Theses assessment protocol dimensions

In previous work, we have used BioTAP to quantitatively assess students’ undergraduate honors theses and explore the relationship between thesis-writing courses (or specific interventions within the courses) and the strength of students’ science reasoning in writing across different science disciplines: biology ( Reynolds and Thompson, 2011 ); chemistry ( Dowd et al. , 2015b ); and economics ( Dowd et al. , 2015a ). We have focused exclusively on the nine dimensions related to reasoning and writing (questions 1–9), as the other four dimensions (questions 10–13) require topic-specific expertise and are intended to be used by the student’s thesis supervisor.

Beyond considering individual dimensions, we have investigated whether meaningful constructs underlie students’ thesis scores. We conducted exploratory factor analysis of students’ theses in biology, economics, and chemistry and found one dominant underlying factor in each discipline; we termed the factor “scientific reasoning in writing” ( Dowd et al. , 2015a , b , 2016 ). That is, each of the nine dimensions could be understood as reflecting, in different ways and to different degrees, the construct of scientific reasoning in writing. The findings indicated evidence of both general and discipline-specific components to scientific reasoning in writing that relate to epistemic beliefs and paradigms, in keeping with broader ideas about science reasoning discussed earlier. Specifically, scientific reasoning in writing is more strongly associated with formulating a compelling argument for the significance of the research in the context of current literature in biology, making meaning regarding the implications of the findings in chemistry, and providing an organizational framework for interpreting the thesis in economics. We suggested that instruction, whether occurring in writing studios or in writing courses to facilitate thesis preparation, should attend to both components.

Research Question and Study Design

The genre of thesis writing combines the pedagogies of writing and inquiry found to foster scientific reasoning ( Reynolds et al. , 2012 ) and critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-­McKnight, 2016 ). However, there is no empirical evidence regarding the general or domain-specific interrelationships of scientific reasoning and critical-thinking skills, particularly in the rhetorical context of the undergraduate thesis. The BioTAP studies discussed earlier indicate that the rubric-based assessment produces evidence of scientific reasoning in the undergraduate thesis, but it was not designed to foster or measure critical thinking. The current study was undertaken to address the research question: How are students’ critical-thinking skills related to scientific reasoning as reflected in the genre of undergraduate thesis writing in biology? Determining these interrelationships could guide efforts to enhance students’ scientific reasoning and writing skills through focusing instruction on specific critical-thinking skills as well as disciplinary conventions.

To address this research question, we focused on undergraduate thesis writers in biology courses at two institutions, Duke University and the University of Minnesota, and examined the extent to which students’ scientific reasoning in writing, assessed in the undergraduate thesis using BioTAP, corresponds to students’ critical-thinking skills, assessed using the California Critical Thinking Skills Test (CCTST; August, 2016 ).

Study Sample

The study sample was composed of students enrolled in courses designed to scaffold the thesis-writing process in the Department of Biology at Duke University and the College of Biological Sciences at the University of Minnesota. Both courses complement students’ individual work with research advisors. The course is required for thesis writers at the University of Minnesota and optional for writers at Duke University. Not all students are required to complete a thesis, though it is required for students to graduate with honors; at the University of Minnesota, such students are enrolled in an honors program within the college. In total, 28 students were enrolled in the course at Duke University and 44 students were enrolled in the course at the University of Minnesota. Of those students, two students did not consent to participate in the study; additionally, five students did not validly complete the CCTST (i.e., attempted fewer than 60% of items or completed the test in less than 15 minutes). Thus, our overall rate of valid participation is 90%, with 27 students from Duke University and 38 students from the University of Minnesota. We found no statistically significant differences in thesis assessment between students with valid CCTST scores and invalid CCTST scores. Therefore, we focus on the 65 students who consented to participate and for whom we have complete and valid data in most of this study. Additionally, in asking students for their consent to participate, we allowed them to choose whether to provide or decline access to academic and demographic background data. Of the 65 students who consented to participate, 52 students granted access to such data. Therefore, for additional analyses involving academic and background data, we focus on the 52 students who consented. We note that the 13 students who participated but declined to share additional data performed slightly lower on the CCTST than the 52 others (perhaps suggesting that they differ by other measures, but we cannot determine this with certainty). Among the 52 students, 60% identified as female and 10% identified as being from underrepresented ethnicities.

In both courses, students completed the CCTST online, either in class or on their own, late in the Spring 2016 semester. This is the same assessment that was used in prior studies of critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-McKnight, 2016 ). It is “an objective measure of the core reasoning skills needed for reflective decision making concerning what to believe or what to do” ( Insight Assessment, 2016a ). In the test, students are asked to read and consider information as they answer multiple-choice questions. The questions are intended to be appropriate for all users, so there is no expectation of prior disciplinary knowledge in biology (or any other subject). Although actual test items are protected, sample items are available on the Insight Assessment website ( Insight Assessment, 2016b ). We have included one sample item in the Supplemental Material.

The CCTST is based on a consensus definition of critical thinking, measures cognitive and metacognitive skills associated with critical thinking, and has been evaluated for validity and reliability at the college level ( August, 2016 ; Stephenson and Sadler-McKnight, 2016 ). In addition to providing overall critical-thinking score, the CCTST assesses seven dimensions of critical thinking: analysis, interpretation, inference, evaluation, explanation, induction, and deduction. Scores on each dimension are calculated based on students’ performance on items related to that dimension. Analysis focuses on identifying assumptions, reasons, and claims and examining how they interact to form arguments. Interpretation, related to analysis, focuses on determining the precise meaning and significance of information. Inference focuses on drawing conclusions from reasons and evidence. Evaluation focuses on assessing the credibility of sources of information and claims they make. Explanation, related to evaluation, focuses on describing the evidence, assumptions, or rationale for beliefs and conclusions. Induction focuses on drawing inferences about what is probably true based on evidence. Deduction focuses on drawing conclusions about what must be true when the context completely determines the outcome. These are not independent dimensions; the fact that they are related supports their collective interpretation as critical thinking. Together, the CCTST dimensions provide a basis for evaluating students’ overall strength in using reasoning to form reflective judgments about what to believe or what to do ( August, 2016 ). Each of the seven dimensions and the overall CCTST score are measured on a scale of 0–100, where higher scores indicate superior performance. Scores correspond to superior (86–100), strong (79–85), moderate (70–78), weak (63–69), or not manifested (62 and below) skills.

Scientific Reasoning in Writing

At the end of the semester, students’ final, submitted undergraduate theses were assessed using BioTAP, which consists of nine rubric dimensions that focus on communication to the broader scientific community and four additional dimensions that focus on the exhibition of topic-specific expertise ( Reynolds et al. , 2009 ). These dimensions, framed as questions, are displayed in Table 1 .

Student theses were assessed on questions 1–9 of BioTAP using the same procedures described in previous studies ( Reynolds and Thompson, 2011 ; Dowd et al. , 2015a , b ). In this study, six raters were trained in the valid, reliable use of BioTAP rubrics. Each dimension was rated on a five-point scale: 1 indicates the dimension is missing, incomplete, or below acceptable standards; 3 indicates that the dimension is adequate but not exhibiting mastery; and 5 indicates that the dimension is excellent and exhibits mastery (intermediate ratings of 2 and 4 are appropriate when different parts of the thesis make a single category challenging). After training, two raters independently assessed each thesis and then discussed their independent ratings with one another to form a consensus rating. The consensus score is not an average score, but rather an agreed-upon, discussion-based score. On a five-point scale, raters independently assessed dimensions to be within 1 point of each other 82.4% of the time before discussion and formed consensus ratings 100% of the time after discussion.

In this study, we consider both categorical (mastery/nonmastery, where a score of 5 corresponds to mastery) and numerical treatments of individual BioTAP scores to better relate the manifestation of critical thinking in BioTAP assessment to all of the prior studies. For comprehensive/cumulative measures of BioTAP, we focus on the partial sum of questions 1–5, as these questions relate to higher-order scientific reasoning (whereas questions 6–9 relate to mid- and lower-order writing mechanics [ Reynolds et al. , 2009 ]), and the factor scores (i.e., numerical representations of the extent to which each student exhibits the underlying factor), which are calculated from the factor loadings published by Dowd et al. (2016) . We do not focus on questions 6–9 individually in statistical analyses, because we do not expect critical-thinking skills to relate to mid- and lower-order writing skills.

The final, submitted thesis reflects the student’s writing, the student’s scientific reasoning, the quality of feedback provided to the student by peers and mentors, and the student’s ability to incorporate that feedback into his or her work. Therefore, our assessment is not the same as an assessment of unpolished, unrevised samples of students’ written work. While one might imagine that such an unpolished sample may be more strongly correlated with critical-thinking skills measured by the CCTST, we argue that the complete, submitted thesis, assessed using BioTAP, is ultimately a more appropriate reflection of how students exhibit science reasoning in the scientific community.

Statistical Analyses

We took several steps to analyze the collected data. First, to provide context for subsequent interpretations, we generated descriptive statistics for the CCTST scores of the participants based on the norms for undergraduate CCTST test takers. To determine the strength of relationships among CCTST dimensions (including overall score) and the BioTAP dimensions, partial-sum score (questions 1–5), and factor score, we calculated Pearson’s correlations for each pair of measures. To examine whether falling on one side of the nonmastery/mastery threshold (as opposed to a linear scale of performance) was related to critical thinking, we grouped BioTAP dimensions into categories (mastery/nonmastery) and conducted Student’s t tests to compare the means scores of the two groups on each of the seven dimensions and overall score of the CCTST. Finally, for the strongest relationship that emerged, we included additional academic and background variables as covariates in multiple linear-regression analysis to explore questions about how much observed relationships between critical-thinking skills and science reasoning in writing might be explained by variation in these other factors.

Although BioTAP scores represent discreet, ordinal bins, the five-point scale is intended to capture an underlying continuous construct (from inadequate to exhibiting mastery). It has been argued that five categories is an appropriate cutoff for treating ordinal variables as pseudo-continuous ( Rhemtulla et al. , 2012 )—and therefore using continuous-variable statistical methods (e.g., Pearson’s correlations)—as long as the underlying assumption that ordinal scores are linearly distributed is valid. Although we have no way to statistically test this assumption, we interpret adequate scores to be approximately halfway between inadequate and mastery scores, resulting in a linear scale. In part because this assumption is subject to disagreement, we also consider and interpret a categorical (mastery/nonmastery) treatment of BioTAP variables.

We corrected for multiple comparisons using the Holm-Bonferroni method ( Holm, 1979 ). At the most general level, where we consider the single, comprehensive measures for BioTAP (partial-sum and factor score) and the CCTST (overall score), there is no need to correct for multiple comparisons, because the multiple, individual dimensions are collapsed into single dimensions. When we considered individual CCTST dimensions in relation to comprehensive measures for BioTAP, we accounted for seven comparisons; similarly, when we considered individual dimensions of BioTAP in relation to overall CCTST score, we accounted for five comparisons. When all seven CCTST and five BioTAP dimensions were examined individually and without prior knowledge, we accounted for 35 comparisons; such a rigorous threshold is likely to reject weak and moderate relationships, but it is appropriate if there are no specific pre-existing hypotheses. All p values are presented in tables for complete transparency, and we carefully consider the implications of our interpretation of these data in the Discussion section.

CCTST scores for students in this sample ranged from the 39th to 99th percentile of the general population of undergraduate CCTST test takers (mean percentile = 84.3, median = 85th percentile; Table 2 ); these percentiles reflect overall scores that range from moderate to superior. Scores on individual dimensions and overall scores were sufficiently normal and far enough from the ceiling of the scale to justify subsequent statistical analyses.

Descriptive statistics of CCTST dimensions a

a Scores correspond to superior (86–100), strong (79–85), moderate (70–78), weak (63–69), or not manifested (62 and lower) skills.

The Pearson’s correlations between students’ cumulative scores on BioTAP (the factor score based on loadings published by Dowd et al. , 2016 , and the partial sum of scores on questions 1–5) and students’ overall scores on the CCTST are presented in Table 3 . We found that the partial-sum measure of BioTAP was significantly related to the overall measure of critical thinking ( r = 0.27, p = 0.03), while the BioTAP factor score was marginally related to overall CCTST ( r = 0.24, p = 0.05). When we looked at relationships between comprehensive BioTAP measures and scores for individual dimensions of the CCTST ( Table 3 ), we found significant positive correlations between the both BioTAP partial-sum and factor scores and CCTST inference ( r = 0.45, p < 0.001, and r = 0.41, p < 0.001, respectively). Although some other relationships have p values below 0.05 (e.g., the correlations between BioTAP partial-sum scores and CCTST induction and interpretation scores), they are not significant when we correct for multiple comparisons.

Correlations between dimensions of CCTST and dimensions of BioTAP a

a In each cell, the top number is the correlation, and the bottom, italicized number is the associated p value. Correlations that are statistically significant after correcting for multiple comparisons are shown in bold.

b This is the partial sum of BioTAP scores on questions 1–5.

c This is the factor score calculated from factor loadings published by Dowd et al. (2016) .

When we expanded comparisons to include all 35 potential correlations among individual BioTAP and CCTST dimensions—and, accordingly, corrected for 35 comparisons—we did not find any additional statistically significant relationships. The Pearson’s correlations between students’ scores on each dimension of BioTAP and students’ scores on each dimension of the CCTST range from −0.11 to 0.35 ( Table 3 ); although the relationship between discussion of implications (BioTAP question 5) and inference appears to be relatively large ( r = 0.35), it is not significant ( p = 0.005; the Holm-Bonferroni cutoff is 0.00143). We found no statistically significant relationships between BioTAP questions 6–9 and CCTST dimensions (unpublished data), regardless of whether we correct for multiple comparisons.

The results of Student’s t tests comparing scores on each dimension of the CCTST of students who exhibit mastery with those of students who do not exhibit mastery on each dimension of BioTAP are presented in Table 4 . Focusing first on the overall CCTST scores, we found that the difference between those who exhibit mastery and those who do not in discussing implications of results (BioTAP question 5) is statistically significant ( t = 2.73, p = 0.008, d = 0.71). When we expanded t tests to include all 35 comparisons—and, like above, corrected for 35 comparisons—we found a significant difference in inference scores between students who exhibit mastery on question 5 and students who do not ( t = 3.41, p = 0.0012, d = 0.88), as well as a marginally significant difference in these students’ induction scores ( t = 3.26, p = 0.0018, d = 0.84; the Holm-Bonferroni cutoff is p = 0.00147). Cohen’s d effect sizes, which reveal the strength of the differences for statistically significant relationships, range from 0.71 to 0.88.

The t statistics and effect sizes of differences in ­dimensions of CCTST across dimensions of BioTAP a

a In each cell, the top number is the t statistic for each comparison, and the middle, italicized number is the associated p value. The bottom number is the effect size. Correlations that are statistically significant after correcting for multiple comparisons are shown in bold.

Finally, we more closely examined the strongest relationship that we observed, which was between the CCTST dimension of inference and the BioTAP partial-sum composite score (shown in Table 3 ), using multiple regression analysis ( Table 5 ). Focusing on the 52 students for whom we have background information, we looked at the simple relationship between BioTAP and inference (model 1), a robust background model including multiple covariates that one might expect to explain some part of the variation in BioTAP (model 2), and a combined model including all variables (model 3). As model 3 shows, the covariates explain very little variation in BioTAP scores, and the relationship between inference and BioTAP persists even in the presence of all of the covariates.

Partial sum (questions 1–5) of BioTAP scores ( n = 52)

** p < 0.01.

*** p < 0.001.

The aim of this study was to examine the extent to which the various components of scientific reasoning—manifested in writing in the genre of undergraduate thesis and assessed using BioTAP—draw on general and specific critical-thinking skills (assessed using CCTST) and to consider the implications for educational practices. Although science reasoning involves critical-thinking skills, it also relates to conceptual knowledge and the epistemological foundations of science disciplines ( Kuhn et al. , 2008 ). Moreover, science reasoning in writing , captured in students’ undergraduate theses, reflects habits, conventions, and the incorporation of feedback that may alter evidence of individuals’ critical-thinking skills. Our findings, however, provide empirical evidence that cumulative measures of science reasoning in writing are nonetheless related to students’ overall critical-thinking skills ( Table 3 ). The particularly significant roles of inference skills ( Table 3 ) and the discussion of implications of results (BioTAP question 5; Table 4 ) provide a basis for more specific ideas about how these constructs relate to one another and what educational interventions may have the most success in fostering these skills.

Our results build on previous findings. The genre of thesis writing combines pedagogies of writing and inquiry found to foster scientific reasoning ( Reynolds et al. , 2012 ) and critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-McKnight, 2016 ). Quitadamo and Kurtz (2007) reported that students who engaged in a laboratory writing component in a general education biology course significantly improved their inference and analysis skills, and Quitadamo and colleagues (2008) found that participation in a community-based inquiry biology course (that included a writing component) was associated with significant gains in students’ inference and evaluation skills. The shared focus on inference is noteworthy, because these prior studies actually differ from the current study; the former considered critical-­thinking skills as the primary learning outcome of writing-­focused interventions, whereas the latter focused on emergent links between two learning outcomes (science reasoning in writing and critical thinking). In other words, inference skills are impacted by writing as well as manifested in writing.

Inference focuses on drawing conclusions from argument and evidence. According to the consensus definition of critical thinking, the specific skill of inference includes several processes: querying evidence, conjecturing alternatives, and drawing conclusions. All of these activities are central to the independent research at the core of writing an undergraduate thesis. Indeed, a critical part of what we call “science reasoning in writing” might be characterized as a measure of students’ ability to infer and make meaning of information and findings. Because the cumulative BioTAP measures distill underlying similarities and, to an extent, suppress unique aspects of individual dimensions, we argue that it is appropriate to relate inference to scientific reasoning in writing . Even when we control for other potentially relevant background characteristics, the relationship is strong ( Table 5 ).

In taking the complementary view and focusing on BioTAP, when we compared students who exhibit mastery with those who do not, we found that the specific dimension of “discussing the implications of results” (question 5) differentiates students’ performance on several critical-thinking skills. To achieve mastery on this dimension, students must make connections between their results and other published studies and discuss the future directions of the research; in short, they must demonstrate an understanding of the bigger picture. The specific relationship between question 5 and inference is the strongest observed among all individual comparisons. Altogether, perhaps more than any other BioTAP dimension, this aspect of students’ writing provides a clear view of the role of students’ critical-thinking skills (particularly inference and, marginally, induction) in science reasoning.

While inference and discussion of implications emerge as particularly strongly related dimensions in this work, we note that the strongest contribution to “science reasoning in writing in biology,” as determined through exploratory factor analysis, is “argument for the significance of research” (BioTAP question 2, not question 5; Dowd et al. , 2016 ). Question 2 is not clearly related to critical-thinking skills. These findings are not contradictory, but rather suggest that the epistemological and disciplinary-specific aspects of science reasoning that emerge in writing through BioTAP are not completely aligned with aspects related to critical thinking. In other words, science reasoning in writing is not simply a proxy for those critical-thinking skills that play a role in science reasoning.

In a similar vein, the content-related, epistemological aspects of science reasoning, as well as the conventions associated with writing the undergraduate thesis (including feedback from peers and revision), may explain the lack of significant relationships between some science reasoning dimensions and some critical-thinking skills that might otherwise seem counterintuitive (e.g., BioTAP question 2, which relates to making an argument, and the critical-thinking skill of argument). It is possible that an individual’s critical-thinking skills may explain some variation in a particular BioTAP dimension, but other aspects of science reasoning and practice exert much stronger influence. Although these relationships do not emerge in our analyses, the lack of significant correlation does not mean that there is definitively no correlation. Correcting for multiple comparisons suppresses type 1 error at the expense of exacerbating type 2 error, which, combined with the limited sample size, constrains statistical power and makes weak relationships more difficult to detect. Ultimately, though, the relationships that do emerge highlight places where individuals’ distinct critical-thinking skills emerge most coherently in thesis assessment, which is why we are particularly interested in unpacking those relationships.

We recognize that, because only honors students submit theses at these institutions, this study sample is composed of a selective subset of the larger population of biology majors. Although this is an inherent limitation of focusing on thesis writing, links between our findings and results of other studies (with different populations) suggest that observed relationships may occur more broadly. The goal of improved science reasoning and critical thinking is shared among all biology majors, particularly those engaged in capstone research experiences. So while the implications of this work most directly apply to honors thesis writers, we provisionally suggest that all students could benefit from further study of them.

There are several important implications of this study for science education practices. Students’ inference skills relate to the understanding and effective application of scientific content. The fact that we find no statistically significant relationships between BioTAP questions 6–9 and CCTST dimensions suggests that such mid- to lower-order elements of BioTAP ( Reynolds et al. , 2009 ), which tend to be more structural in nature, do not focus on aspects of the finished thesis that draw strongly on critical thinking. In keeping with prior analyses ( Reynolds and Thompson, 2011 ; Dowd et al. , 2016 ), these findings further reinforce the notion that disciplinary instructors, who are most capable of teaching and assessing scientific reasoning and perhaps least interested in the more mechanical aspects of writing, may nonetheless be best suited to effectively model and assess students’ writing.

The goal of the thesis writing course at both Duke University and the University of Minnesota is not merely to improve thesis scores but to move students’ writing into the category of mastery across BioTAP dimensions. Recognizing that students with differing critical-thinking skills (particularly inference) are more or less likely to achieve mastery in the undergraduate thesis (particularly in discussing implications [question 5]) is important for developing and testing targeted pedagogical interventions to improve learning outcomes for all students.

The competencies characterized by the Vision and Change in Undergraduate Biology Education Initiative provide a general framework for recognizing that science reasoning and critical-thinking skills play key roles in major learning outcomes of science education. Our findings highlight places where science reasoning–related competencies (like “understanding the process of science”) connect to critical-thinking skills and places where critical thinking–related competencies might be manifested in scientific products (such as the ability to discuss implications in scientific writing). We encourage broader efforts to build empirical connections between competencies and pedagogical practices to further improve science education.

One specific implication of this work for science education is to focus on providing opportunities for students to develop their critical-thinking skills (particularly inference). Of course, as this correlational study is not designed to test causality, we do not claim that enhancing students’ inference skills will improve science reasoning in writing. However, as prior work shows that science writing activities influence students’ inference skills ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ), there is reason to test such a hypothesis. Nevertheless, the focus must extend beyond inference as an isolated skill; rather, it is important to relate inference to the foundations of the scientific method ( Miri et al. , 2007 ) in terms of the epistemological appreciation of the functions and coordination of evidence ( Kuhn and Dean, 2004 ; Zeineddin and Abd-El-Khalick, 2010 ; Ding et al. , 2016 ) and disciplinary paradigms of truth and justification ( Moshman, 2015 ).

Although this study is limited to the domain of biology at two institutions with a relatively small number of students, the findings represent a foundational step in the direction of achieving success with more integrated learning outcomes. Hopefully, it will spur greater interest in empirically grounding discussions of the constructs of scientific reasoning and critical-thinking skills.

This study contributes to the efforts to improve science education, for both majors and nonmajors, through an empirically driven analysis of the relationships between scientific reasoning reflected in the genre of thesis writing and critical-thinking skills. This work is rooted in the usefulness of BioTAP as a method 1) to facilitate communication and learning and 2) to assess disciplinary-specific and general dimensions of science reasoning. The findings support the important role of the critical-thinking skill of inference in scientific reasoning in writing, while also highlighting ways in which other aspects of science reasoning (epistemological considerations, writing conventions, etc.) are not significantly related to critical thinking. Future research into the impact of interventions focused on specific critical-thinking skills (i.e., inference) for improved science reasoning in writing will build on this work and its implications for science education.

Supplementary Material

Acknowledgments.

We acknowledge the contributions of Kelaine Haas and Alexander Motten to the implementation and collection of data. We also thank Mine Çetinkaya-­Rundel for her insights regarding our statistical analyses. This research was funded by National Science Foundation award DUE-1525602.

  • American Association for the Advancement of Science. (2011). Vision and change in undergraduate biology education: A call to action . Washington, DC: Retrieved September 26, 2017, from https://visionandchange.org/files/2013/11/aaas-VISchange-web1113.pdf . [ Google Scholar ]
  • August D. (2016). California Critical Thinking Skills Test user manual and resource guide . San Jose: Insight Assessment/California Academic Press. [ Google Scholar ]
  • Beyer C. H., Taylor E., Gillmore G. M. (2013). Inside the undergraduate teaching experience: The University of Washington’s growth in faculty teaching study . Albany, NY: SUNY Press. [ Google Scholar ]
  • Bissell A. N., Lemons P. P. (2006). A new method for assessing critical thinking in the classroom . BioScience , ( 1 ), 66–72. https://doi.org/10.1641/0006-3568(2006)056[0066:ANMFAC]2.0.CO;2 . [ Google Scholar ]
  • Blattner N. H., Frazier C. L. (2002). Developing a performance-based assessment of students’ critical thinking skills . Assessing Writing , ( 1 ), 47–64. [ Google Scholar ]
  • Clase K. L., Gundlach E., Pelaez N. J. (2010). Calibrated peer review for computer-assisted learning of biological research competencies . Biochemistry and Molecular Biology Education , ( 5 ), 290–295. [ PubMed ] [ Google Scholar ]
  • Condon W., Kelly-Riley D. (2004). Assessing and teaching what we value: The relationship between college-level writing and critical thinking abilities . Assessing Writing , ( 1 ), 56–75. https://doi.org/10.1016/j.asw.2004.01.003 . [ Google Scholar ]
  • Ding L., Wei X., Liu X. (2016). Variations in university students’ scientific reasoning skills across majors, years, and types of institutions . Research in Science Education , ( 5 ), 613–632. https://doi.org/10.1007/s11165-015-9473-y . [ Google Scholar ]
  • Dowd J. E., Connolly M. P., Thompson R. J., Jr., Reynolds J. A. (2015a). Improved reasoning in undergraduate writing through structured workshops . Journal of Economic Education , ( 1 ), 14–27. https://doi.org/10.1080/00220485.2014.978924 . [ Google Scholar ]
  • Dowd J. E., Roy C. P., Thompson R. J., Jr., Reynolds J. A. (2015b). “On course” for supporting expanded participation and improving scientific reasoning in undergraduate thesis writing . Journal of Chemical Education , ( 1 ), 39–45. https://doi.org/10.1021/ed500298r . [ Google Scholar ]
  • Dowd J. E., Thompson R. J., Jr., Reynolds J. A. (2016). Quantitative genre analysis of undergraduate theses: Uncovering different ways of writing and thinking in science disciplines . WAC Journal , , 36–51. [ Google Scholar ]
  • Facione P. A. (1990). Critical thinking: a statement of expert consensus for purposes of educational assessment and instruction. Research findings and recommendations . Newark, DE: American Philosophical Association; Retrieved September 26, 2017, from https://philpapers.org/archive/FACCTA.pdf . [ Google Scholar ]
  • Gerdeman R. D., Russell A. A., Worden K. J., Gerdeman R. D., Russell A. A., Worden K. J. (2007). Web-based student writing and reviewing in a large biology lecture course . Journal of College Science Teaching , ( 5 ), 46–52. [ Google Scholar ]
  • Greenhoot A. F., Semb G., Colombo J., Schreiber T. (2004). Prior beliefs and methodological concepts in scientific reasoning . Applied Cognitive Psychology , ( 2 ), 203–221. https://doi.org/10.1002/acp.959 . [ Google Scholar ]
  • Haaga D. A. F. (1993). Peer review of term papers in graduate psychology courses . Teaching of Psychology , ( 1 ), 28–32. https://doi.org/10.1207/s15328023top2001_5 . [ Google Scholar ]
  • Halonen J. S., Bosack T., Clay S., McCarthy M., Dunn D. S., Hill G. W., Whitlock K. (2003). A rubric for learning, teaching, and assessing scientific inquiry in psychology . Teaching of Psychology , ( 3 ), 196–208. https://doi.org/10.1207/S15328023TOP3003_01 . [ Google Scholar ]
  • Hand B., Keys C. W. (1999). Inquiry investigation . Science Teacher , ( 4 ), 27–29. [ Google Scholar ]
  • Holm S. (1979). A simple sequentially rejective multiple test procedure . Scandinavian Journal of Statistics , ( 2 ), 65–70. [ Google Scholar ]
  • Holyoak K. J., Morrison R. G. (2005). The Cambridge handbook of thinking and reasoning . New York: Cambridge University Press. [ Google Scholar ]
  • Insight Assessment. (2016a). California Critical Thinking Skills Test (CCTST) Retrieved September 26, 2017, from www.insightassessment.com/Products/Products-Summary/Critical-Thinking-Skills-Tests/California-Critical-Thinking-Skills-Test-CCTST .
  • Insight Assessment. (2016b). Sample thinking skills questions. Retrieved September 26, 2017, from www.insightassessment.com/Resources/Teaching-Training-and-Learning-Tools/node_1487 .
  • Kelly G. J., Takao A. (2002). Epistemic levels in argument: An analysis of university oceanography students’ use of evidence in writing . Science Education , ( 3 ), 314–342. https://doi.org/10.1002/sce.10024 . [ Google Scholar ]
  • Kuhn D., Dean D., Jr. (2004). Connecting scientific reasoning and causal inference . Journal of Cognition and Development , ( 2 ), 261–288. https://doi.org/10.1207/s15327647jcd0502_5 . [ Google Scholar ]
  • Kuhn D., Iordanou K., Pease M., Wirkala C. (2008). Beyond control of variables: What needs to develop to achieve skilled scientific thinking? . Cognitive Development , ( 4 ), 435–451. https://doi.org/10.1016/j.cogdev.2008.09.006 . [ Google Scholar ]
  • Lawson A. E. (2010). Basic inferences of scientific reasoning, argumentation, and discovery . Science Education , ( 2 ), 336–364. https://doi.org/­10.1002/sce.20357 . [ Google Scholar ]
  • Meizlish D., LaVaque-Manty D., Silver N., Kaplan M. (2013). Think like/write like: Metacognitive strategies to foster students’ development as disciplinary thinkers and writers . In Thompson R. J. (Ed.), Changing the conversation about higher education (pp. 53–73). Lanham, MD: Rowman & Littlefield. [ Google Scholar ]
  • Miri B., David B.-C., Uri Z. (2007). Purposely teaching for the promotion of higher-order thinking skills: A case of critical thinking . Research in Science Education , ( 4 ), 353–369. https://doi.org/10.1007/s11165-006-9029-2 . [ Google Scholar ]
  • Moshman D. (2015). Epistemic cognition and development: The psychology of justification and truth . New York: Psychology Press. [ Google Scholar ]
  • National Research Council. (2000). How people learn: Brain, mind, experience, and school . Expanded ed. Washington, DC: National Academies Press. [ Google Scholar ]
  • Pukkila P. J. (2004). Introducing student inquiry in large introductory genetics classes . Genetics , ( 1 ), 11–18. https://doi.org/10.1534/genetics.166.1.11 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Quitadamo I. J., Faiola C. L., Johnson J. E., Kurtz M. J. (2008). Community-based inquiry improves critical thinking in general education biology . CBE—Life Sciences Education , ( 3 ), 327–337. https://doi.org/10.1187/cbe.07-11-0097 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Quitadamo I. J., Kurtz M. J. (2007). Learning to improve: Using writing to increase critical thinking performance in general education biology . CBE—Life Sciences Education , ( 2 ), 140–154. https://doi.org/10.1187/cbe.06-11-0203 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Reynolds J. A., Smith R., Moskovitz C., Sayle A. (2009). BioTAP: A systematic approach to teaching scientific writing and evaluating undergraduate theses . BioScience , ( 10 ), 896–903. https://doi.org/10.1525/bio.2009.59.10.11 . [ Google Scholar ]
  • Reynolds J. A., Thaiss C., Katkin W., Thompson R. J. (2012). Writing-to-learn in undergraduate science education: A community-based, conceptually driven approach . CBE—Life Sciences Education , ( 1 ), 17–25. https://doi.org/10.1187/cbe.11-08-0064 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Reynolds J. A., Thompson R. J. (2011). Want to improve undergraduate thesis writing? Engage students and their faculty readers in scientific peer review . CBE—Life Sciences Education , ( 2 ), 209–215. https://doi.org/­10.1187/cbe.10-10-0127 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Rhemtulla M., Brosseau-Liard P. E., Savalei V. (2012). When can categorical variables be treated as continuous? A comparison of robust continuous and categorical SEM estimation methods under suboptimal conditions . Psychological Methods , ( 3 ), 354–373. https://doi.org/­10.1037/a0029315 . [ PubMed ] [ Google Scholar ]
  • Stephenson N. S., Sadler-McKnight N. P. (2016). Developing critical thinking skills using the science writing heuristic in the chemistry laboratory . Chemistry Education Research and Practice , ( 1 ), 72–79. https://doi.org/­10.1039/C5RP00102A . [ Google Scholar ]
  • Tariq V. N., Stefani L. A. J., Butcher A. C., Heylings D. J. A. (1998). Developing a new approach to the assessment of project work . Assessment and Evaluation in Higher Education , ( 3 ), 221–240. https://doi.org/­10.1080/0260293980230301 . [ Google Scholar ]
  • Timmerman B. E. C., Strickland D. C., Johnson R. L., Payne J. R. (2011). Development of a “universal” rubric for assessing undergraduates’ scientific reasoning skills using scientific writing . Assessment and Evaluation in Higher Education , ( 5 ), 509–547. https://doi.org/10.1080/­02602930903540991 . [ Google Scholar ]
  • Topping K. J., Smith E. F., Swanson I., Elliot A. (2000). Formative peer assessment of academic writing between postgraduate students . Assessment and Evaluation in Higher Education , ( 2 ), 149–169. https://doi.org/10.1080/713611428 . [ Google Scholar ]
  • Willison J., O’Regan K. (2007). Commonly known, commonly not known, totally unknown: A framework for students becoming researchers . Higher Education Research and Development , ( 4 ), 393–409. https://doi.org/10.1080/07294360701658609 . [ Google Scholar ]
  • Woodin T., Carter V. C., Fletcher L. (2010). Vision and Change in Biology Undergraduate Education: A Call for Action—Initial responses . CBE—Life Sciences Education , ( 2 ), 71–73. https://doi.org/10.1187/cbe.10-03-0044 . [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Zeineddin A., Abd-El-Khalick F. (2010). Scientific reasoning and epistemological commitments: Coordination of theory and evidence among college science students . Journal of Research in Science Teaching , ( 9 ), 1064–1093. https://doi.org/10.1002/tea.20368 . [ Google Scholar ]
  • Zimmerman C. (2000). The development of scientific reasoning skills . Developmental Review , ( 1 ), 99–149. https://doi.org/10.1006/drev.1999.0497 . [ Google Scholar ]
  • Zimmerman C. (2007). The development of scientific thinking skills in elementary and middle school . Developmental Review , ( 2 ), 172–223. https://doi.org/10.1016/j.dr.2006.12.001 . [ Google Scholar ]

Collaborative Learning and Critical Thinking

  • Reference work entry
  • Cite this reference work entry

discuss the relationship between critical thinking and problem solving

  • Anu A. Gokhale 2  

2447 Accesses

28 Citations

Cooperative learning ; Creative thinking ; Problem-solving

The term “collaborative learning” refers to an instruction method in which students at various performance levels work together in small groups toward a common goal. Collaborative learning is a relationship among learners that fosters positive interdependence, individual accountability, and interpersonal skills. “Critical thinking” involves asking appropriate questions, gathering and creatively sorting through relevant information, relating new information to existing knowledge, reexamining beliefs, reasoning logically, and drawing reliable and trustworthy conclusions.

Theoretical Background

The advent of revolutionary information and communication technologies has effected changes in the organizational infrastructure and altered the characteristics of the workplace putting an increased emphasis on teamwork and processes that require individuals to pool their resources and integrate specializations. The...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

American Philosophical Association. (1990). Critical thinking: A statement of expert consensus for purposes of educational assessment and instruction. ERIC document ED (pp. 315–423).

Google Scholar  

Cooper, J., Prescott, S., Cook, L., Smith, L., Mueck, R., & Cuseo, J. (1990). Cooperative learning and college instruction: Effective use of student learning teams . Long Beach: California State University Foundation.

Gokhale, A. A. (1995). Collaborative learning enhances critical thinking. Journal of Technology Education, 7 (1), 22–30.

Johnson, D. W., & Johnson, R. T. (1989). Cooperation and competition: Theory and research . Edina: Interaction Book Company.

Kollar, I., Fischer, F., & Hesse, F. (2006). Collaboration scripts – A conceptual analysis. Educational Psychology Review, 18 (2), 159–185.

Slavin, R. E. (1995). Cooperative learning: Theory, research, and practice (2nd ed.). Boston: Allyn & Bacon.

Download references

Author information

Authors and affiliations.

Department of Technology, Illinois State University, Campus Box 5100, Normal, IL, 61790-5110, USA

Dr. Anu A. Gokhale

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Anu A. Gokhale .

Editor information

Editors and affiliations.

Faculty of Economics and Behavioral Sciences, Department of Education, University of Freiburg, 79085, Freiburg, Germany

Norbert M. Seel

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer Science+Business Media, LLC

About this entry

Cite this entry.

Gokhale, A.A. (2012). Collaborative Learning and Critical Thinking. In: Seel, N.M. (eds) Encyclopedia of the Sciences of Learning. Springer, Boston, MA. https://doi.org/10.1007/978-1-4419-1428-6_910

Download citation

DOI : https://doi.org/10.1007/978-1-4419-1428-6_910

Publisher Name : Springer, Boston, MA

Print ISBN : 978-1-4419-1427-9

Online ISBN : 978-1-4419-1428-6

eBook Packages : Humanities, Social Sciences and Law

Share this entry

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

Change Password

Your password must have 8 characters or more and contain 3 of the following:.

  • a lower case character, 
  • an upper case character, 
  • a special character 

Password Changed Successfully

Your password has been changed

  • Sign in / Register

Request Username

Can't sign in? Forgot your username?

Enter your email address below and we will send you your username

If the address matches an existing account you will receive an email with instructions to retrieve your username

Understanding the Complex Relationship between Critical Thinking and Science Reasoning among Undergraduate Thesis Writers

  • Jason E. Dowd
  • Robert J. Thompson
  • Leslie A. Schiff
  • Julie A. Reynolds

*Address correspondence to: Jason E. Dowd ( E-mail Address: [email protected] ).

Department of Biology, Duke University, Durham, NC 27708

Search for more papers by this author

Department of Psychology and Neuroscience, Duke University, Durham, NC 27708

Department of Microbiology and Immunology, University of Minnesota, Minneapolis, MN 55455

Developing critical-thinking and scientific reasoning skills are core learning objectives of science education, but little empirical evidence exists regarding the interrelationships between these constructs. Writing effectively fosters students’ development of these constructs, and it offers a unique window into studying how they relate. In this study of undergraduate thesis writing in biology at two universities, we examine how scientific reasoning exhibited in writing (assessed using the Biology Thesis Assessment Protocol) relates to general and specific critical-thinking skills (assessed using the California Critical Thinking Skills Test), and we consider implications for instruction. We find that scientific reasoning in writing is strongly related to inference , while other aspects of science reasoning that emerge in writing (epistemological considerations, writing conventions, etc.) are not significantly related to critical-thinking skills. Science reasoning in writing is not merely a proxy for critical thinking. In linking features of students’ writing to their critical-thinking skills, this study 1) provides a bridge to prior work suggesting that engagement in science writing enhances critical thinking and 2) serves as a foundational step for subsequently determining whether instruction focused explicitly on developing critical-thinking skills (particularly inference ) can actually improve students’ scientific reasoning in their writing.

INTRODUCTION

Critical-thinking and scientific reasoning skills are core learning objectives of science education for all students, regardless of whether or not they intend to pursue a career in science or engineering. Consistent with the view of learning as construction of understanding and meaning ( National Research Council, 2000 ), the pedagogical practice of writing has been found to be effective not only in fostering the development of students’ conceptual and procedural knowledge ( Gerdeman et al. , 2007 ) and communication skills ( Clase et al. , 2010 ), but also scientific reasoning ( Reynolds et al. , 2012 ) and critical-thinking skills ( Quitadamo and Kurtz, 2007 ).

Critical thinking and scientific reasoning are similar but different constructs that include various types of higher-order cognitive processes, metacognitive strategies, and dispositions involved in making meaning of information. Critical thinking is generally understood as the broader construct ( Holyoak and Morrison, 2005 ), comprising an array of cognitive processes and dispostions that are drawn upon differentially in everyday life and across domains of inquiry such as the natural sciences, social sciences, and humanities. Scientific reasoning, then, may be interpreted as the subset of critical-thinking skills (cognitive and metacognitive processes and dispositions) that 1) are involved in making meaning of information in scientific domains and 2) support the epistemological commitment to scientific methodology and paradigm(s).

Although there has been an enduring focus in higher education on promoting critical thinking and reasoning as general or “transferable” skills, research evidence provides increasing support for the view that reasoning and critical thinking are also situational or domain specific ( Beyer et al. , 2013 ). Some researchers, such as Lawson (2010) , present frameworks in which science reasoning is characterized explicitly in terms of critical-thinking skills. There are, however, limited coherent frameworks and empirical evidence regarding either the general or domain-specific interrelationships of scientific reasoning, as it is most broadly defined, and critical-thinking skills.

The Vision and Change in Undergraduate Biology Education Initiative provides a framework for thinking about these constructs and their interrelationship in the context of the core competencies and disciplinary practice they describe ( American Association for the Advancement of Science, 2011 ). These learning objectives aim for undergraduates to “understand the process of science, the interdisciplinary nature of the new biology and how science is closely integrated within society; be competent in communication and collaboration; have quantitative competency and a basic ability to interpret data; and have some experience with modeling, simulation and computational and systems level approaches as well as with using large databases” ( Woodin et al. , 2010 , pp. 71–72). This framework makes clear that science reasoning and critical-thinking skills play key roles in major learning outcomes; for example, “understanding the process of science” requires students to engage in (and be metacognitive about) scientific reasoning, and having the “ability to interpret data” requires critical-thinking skills. To help students better achieve these core competencies, we must better understand the interrelationships of their composite parts. Thus, the next step is to determine which specific critical-thinking skills are drawn upon when students engage in science reasoning in general and with regard to the particular scientific domain being studied. Such a determination could be applied to improve science education for both majors and nonmajors through pedagogical approaches that foster critical-thinking skills that are most relevant to science reasoning.

Writing affords one of the most effective means for making thinking visible ( Reynolds et al. , 2012 ) and learning how to “think like” and “write like” disciplinary experts ( Meizlish et al. , 2013 ). As a result, student writing affords the opportunities to both foster and examine the interrelationship of scientific reasoning and critical-thinking skills within and across disciplinary contexts. The purpose of this study was to better understand the relationship between students’ critical-thinking skills and scientific reasoning skills as reflected in the genre of undergraduate thesis writing in biology departments at two research universities, the University of Minnesota and Duke University.

In the following subsections, we discuss in greater detail the constructs of scientific reasoning and critical thinking, as well as the assessment of scientific reasoning in students’ thesis writing. In subsequent sections, we discuss our study design, findings, and the implications for enhancing educational practices.

Critical Thinking

The advances in cognitive science in the 21st century have increased our understanding of the mental processes involved in thinking and reasoning, as well as memory, learning, and problem solving. Critical thinking is understood to include both a cognitive dimension and a disposition dimension (e.g., reflective thinking) and is defined as “purposeful, self-regulatory judgment which results in interpretation, analysis, evaluation, and inference, as well as explanation of the evidential, conceptual, methodological, criteriological, or contextual considera­tions upon which that judgment is based” ( Facione, 1990, p. 3 ). Although various other definitions of critical thinking have been proposed, researchers have generally coalesced on this consensus: expert view ( Blattner and Frazier, 2002 ; Condon and Kelly-Riley, 2004 ; Bissell and Lemons, 2006 ; Quitadamo and Kurtz, 2007 ) and the corresponding measures of critical-­thinking skills ( August, 2016 ; Stephenson and Sadler-McKnight, 2016 ).

Both the cognitive skills and dispositional components of critical thinking have been recognized as important to science education ( Quitadamo and Kurtz, 2007 ). Empirical research demonstrates that specific pedagogical practices in science courses are effective in fostering students’ critical-thinking skills. Quitadamo and Kurtz (2007) found that students who engaged in a laboratory writing component in the context of a general education biology course significantly improved their overall critical-thinking skills (and their analytical and inference skills, in particular), whereas students engaged in a traditional quiz-based laboratory did not improve their critical-thinking skills. In related work, Quitadamo et al. (2008) found that a community-based inquiry experience, involving inquiry, writing, research, and analysis, was associated with improved critical thinking in a biology course for nonmajors, compared with traditionally taught sections. In both studies, students who exhibited stronger presemester critical-thinking skills exhibited stronger gains, suggesting that “students who have not been explicitly taught how to think critically may not reach the same potential as peers who have been taught these skills” ( Quitadamo and Kurtz, 2007 , p. 151).

Recently, Stephenson and Sadler-McKnight (2016) found that first-year general chemistry students who engaged in a science writing heuristic laboratory, which is an inquiry-based, writing-to-learn approach to instruction ( Hand and Keys, 1999 ), had significantly greater gains in total critical-thinking scores than students who received traditional laboratory instruction. Each of the four components—inquiry, writing, collaboration, and reflection—have been linked to critical thinking ( Stephenson and Sadler-McKnight, 2016 ). Like the other studies, this work highlights the value of targeting critical-thinking skills and the effectiveness of an inquiry-based, writing-to-learn approach to enhance critical thinking. Across studies, authors advocate adopting critical thinking as the course framework ( Pukkila, 2004 ) and developing explicit examples of how critical thinking relates to the scientific method ( Miri et al. , 2007 ).

In these examples, the important connection between writing and critical thinking is highlighted by the fact that each intervention involves the incorporation of writing into science, technology, engineering, and mathematics education (either alone or in combination with other pedagogical practices). However, critical-thinking skills are not always the primary learning outcome; in some contexts, scientific reasoning is the primary outcome that is assessed.

Scientific Reasoning

Scientific reasoning is a complex process that is broadly defined as “the skills involved in inquiry, experimentation, evidence evaluation, and inference that are done in the service of conceptual change or scientific understanding” ( Zimmerman, 2007 , p. 172). Scientific reasoning is understood to include both conceptual knowledge and the cognitive processes involved with generation of hypotheses (i.e., inductive processes involved in the generation of hypotheses and the deductive processes used in the testing of hypotheses), experimentation strategies, and evidence evaluation strategies. These dimensions are interrelated, in that “experimentation and inference strategies are selected based on prior conceptual knowledge of the domain” ( Zimmerman, 2000 , p. 139). Furthermore, conceptual and procedural knowledge and cognitive process dimensions can be general and domain specific (or discipline specific).

With regard to conceptual knowledge, attention has been focused on the acquisition of core methodological concepts fundamental to scientists’ causal reasoning and metacognitive distancing (or decontextualized thinking), which is the ability to reason independently of prior knowledge or beliefs ( Greenhoot et al. , 2004 ). The latter involves what Kuhn and Dean (2004) refer to as the coordination of theory and evidence, which requires that one question existing theories (i.e., prior knowledge and beliefs), seek contradictory evidence, eliminate alternative explanations, and revise one’s prior beliefs in the face of contradictory evidence. Kuhn and colleagues (2008) further elaborate that scientific thinking requires “a mature understanding of the epistemological foundations of science, recognizing scientific knowledge as constructed by humans rather than simply discovered in the world,” and “the ability to engage in skilled argumentation in the scientific domain, with an appreciation of argumentation as entailing the coordination of theory and evidence” ( Kuhn et al. , 2008 , p. 435). “This approach to scientific reasoning not only highlights the skills of generating and evaluating evidence-based inferences, but also encompasses epistemological appreciation of the functions of evidence and theory” ( Ding et al. , 2016 , p. 616). Evaluating evidence-based inferences involves epistemic cognition, which Moshman (2015) defines as the subset of metacognition that is concerned with justification, truth, and associated forms of reasoning. Epistemic cognition is both general and domain specific (or discipline specific; Moshman, 2015 ).

There is empirical support for the contributions of both prior knowledge and an understanding of the epistemological foundations of science to scientific reasoning. In a study of undergraduate science students, advanced scientific reasoning was most often accompanied by accurate prior knowledge as well as sophisticated epistemological commitments; additionally, for students who had comparable levels of prior knowledge, skillful reasoning was associated with a strong epistemological commitment to the consistency of theory with evidence ( Zeineddin and Abd-El-Khalick, 2010 ). These findings highlight the importance of the need for instructional activities that intentionally help learners develop sophisticated epistemological commitments focused on the nature of knowledge and the role of evidence in supporting knowledge claims ( Zeineddin and Abd-El-Khalick, 2010 ).

Scientific Reasoning in Students’ Thesis Writing

Pedagogical approaches that incorporate writing have also focused on enhancing scientific reasoning. Many rubrics have been developed to assess aspects of scientific reasoning in written artifacts. For example, Timmerman and colleagues (2011) , in the course of describing their own rubric for assessing scientific reasoning, highlight several examples of scientific reasoning assessment criteria ( Haaga, 1993 ; Tariq et al. , 1998 ; Topping et al. , 2000 ; Kelly and Takao, 2002 ; Halonen et al. , 2003 ; Willison and O’Regan, 2007 ).

At both the University of Minnesota and Duke University, we have focused on the genre of the undergraduate honors thesis as the rhetorical context in which to study and improve students’ scientific reasoning and writing. We view the process of writing an undergraduate honors thesis as a form of professional development in the sciences (i.e., a way of engaging students in the practices of a community of discourse). We have found that structured courses designed to scaffold the thesis-­writing process and promote metacognition can improve writing and reasoning skills in biology, chemistry, and economics ( Reynolds and Thompson, 2011 ; Dowd et al. , 2015a , b ). In the context of this prior work, we have defined scientific reasoning in writing as the emergent, underlying construct measured across distinct aspects of students’ written discussion of independent research in their undergraduate theses.

The Biology Thesis Assessment Protocol (BioTAP) was developed at Duke University as a tool for systematically guiding students and faculty through a “draft–feedback–revision” writing process, modeled after professional scientific peer-review processes ( Reynolds et al. , 2009 ). BioTAP includes activities and worksheets that allow students to engage in critical peer review and provides detailed descriptions, presented as rubrics, of the questions (i.e., dimensions, shown in Table 1 ) upon which such review should focus. Nine rubric dimensions focus on communication to the broader scientific community, and four rubric dimensions focus on the accuracy and appropriateness of the research. These rubric dimensions provide criteria by which the thesis is assessed, and therefore allow BioTAP to be used as an assessment tool as well as a teaching resource ( Reynolds et al. , 2009 ). Full details are available at www.science-writing.org/biotap.html .

In previous work, we have used BioTAP to quantitatively assess students’ undergraduate honors theses and explore the relationship between thesis-writing courses (or specific interventions within the courses) and the strength of students’ science reasoning in writing across different science disciplines: biology ( Reynolds and Thompson, 2011 ); chemistry ( Dowd et al. , 2015b ); and economics ( Dowd et al. , 2015a ). We have focused exclusively on the nine dimensions related to reasoning and writing (questions 1–9), as the other four dimensions (questions 10–13) require topic-specific expertise and are intended to be used by the student’s thesis supervisor.

Beyond considering individual dimensions, we have investigated whether meaningful constructs underlie students’ thesis scores. We conducted exploratory factor analysis of students’ theses in biology, economics, and chemistry and found one dominant underlying factor in each discipline; we termed the factor “scientific reasoning in writing” ( Dowd et al. , 2015a , b , 2016 ). That is, each of the nine dimensions could be understood as reflecting, in different ways and to different degrees, the construct of scientific reasoning in writing. The findings indicated evidence of both general and discipline-specific components to scientific reasoning in writing that relate to epistemic beliefs and paradigms, in keeping with broader ideas about science reasoning discussed earlier. Specifically, scientific reasoning in writing is more strongly associated with formulating a compelling argument for the significance of the research in the context of current literature in biology, making meaning regarding the implications of the findings in chemistry, and providing an organizational framework for interpreting the thesis in economics. We suggested that instruction, whether occurring in writing studios or in writing courses to facilitate thesis preparation, should attend to both components.

Research Question and Study Design

The genre of thesis writing combines the pedagogies of writing and inquiry found to foster scientific reasoning ( Reynolds et al. , 2012 ) and critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-­McKnight, 2016 ). However, there is no empirical evidence regarding the general or domain-specific interrelationships of scientific reasoning and critical-thinking skills, particularly in the rhetorical context of the undergraduate thesis. The BioTAP studies discussed earlier indicate that the rubric-based assessment produces evidence of scientific reasoning in the undergraduate thesis, but it was not designed to foster or measure critical thinking. The current study was undertaken to address the research question: How are students’ critical-thinking skills related to scientific reasoning as reflected in the genre of undergraduate thesis writing in biology? Determining these interrelationships could guide efforts to enhance students’ scientific reasoning and writing skills through focusing instruction on specific critical-thinking skills as well as disciplinary conventions.

To address this research question, we focused on undergraduate thesis writers in biology courses at two institutions, Duke University and the University of Minnesota, and examined the extent to which students’ scientific reasoning in writing, assessed in the undergraduate thesis using BioTAP, corresponds to students’ critical-thinking skills, assessed using the California Critical Thinking Skills Test (CCTST; August, 2016 ).

Study Sample

The study sample was composed of students enrolled in courses designed to scaffold the thesis-writing process in the Department of Biology at Duke University and the College of Biological Sciences at the University of Minnesota. Both courses complement students’ individual work with research advisors. The course is required for thesis writers at the University of Minnesota and optional for writers at Duke University. Not all students are required to complete a thesis, though it is required for students to graduate with honors; at the University of Minnesota, such students are enrolled in an honors program within the college. In total, 28 students were enrolled in the course at Duke University and 44 students were enrolled in the course at the University of Minnesota. Of those students, two students did not consent to participate in the study; additionally, five students did not validly complete the CCTST (i.e., attempted fewer than 60% of items or completed the test in less than 15 minutes). Thus, our overall rate of valid participation is 90%, with 27 students from Duke University and 38 students from the University of Minnesota. We found no statistically significant differences in thesis assessment between students with valid CCTST scores and invalid CCTST scores. Therefore, we focus on the 65 students who consented to participate and for whom we have complete and valid data in most of this study. Additionally, in asking students for their consent to participate, we allowed them to choose whether to provide or decline access to academic and demographic background data. Of the 65 students who consented to participate, 52 students granted access to such data. Therefore, for additional analyses involving academic and background data, we focus on the 52 students who consented. We note that the 13 students who participated but declined to share additional data performed slightly lower on the CCTST than the 52 others (perhaps suggesting that they differ by other measures, but we cannot determine this with certainty). Among the 52 students, 60% identified as female and 10% identified as being from underrepresented ethnicities.

In both courses, students completed the CCTST online, either in class or on their own, late in the Spring 2016 semester. This is the same assessment that was used in prior studies of critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-McKnight, 2016 ). It is “an objective measure of the core reasoning skills needed for reflective decision making concerning what to believe or what to do” ( Insight Assessment, 2016a ). In the test, students are asked to read and consider information as they answer multiple-choice questions. The questions are intended to be appropriate for all users, so there is no expectation of prior disciplinary knowledge in biology (or any other subject). Although actual test items are protected, sample items are available on the Insight Assessment website ( Insight Assessment, 2016b ). We have included one sample item in the Supplemental Material.

The CCTST is based on a consensus definition of critical thinking, measures cognitive and metacognitive skills associated with critical thinking, and has been evaluated for validity and reliability at the college level ( August, 2016 ; Stephenson and Sadler-McKnight, 2016 ). In addition to providing overall critical-thinking score, the CCTST assesses seven dimensions of critical thinking: analysis, interpretation, inference, evaluation, explanation, induction, and deduction. Scores on each dimension are calculated based on students’ performance on items related to that dimension. Analysis focuses on identifying assumptions, reasons, and claims and examining how they interact to form arguments. Interpretation, related to analysis, focuses on determining the precise meaning and significance of information. Inference focuses on drawing conclusions from reasons and evidence. Evaluation focuses on assessing the credibility of sources of information and claims they make. Explanation, related to evaluation, focuses on describing the evidence, assumptions, or rationale for beliefs and conclusions. Induction focuses on drawing inferences about what is probably true based on evidence. Deduction focuses on drawing conclusions about what must be true when the context completely determines the outcome. These are not independent dimensions; the fact that they are related supports their collective interpretation as critical thinking. Together, the CCTST dimensions provide a basis for evaluating students’ overall strength in using reasoning to form reflective judgments about what to believe or what to do ( August, 2016 ). Each of the seven dimensions and the overall CCTST score are measured on a scale of 0–100, where higher scores indicate superior performance. Scores correspond to superior (86–100), strong (79–85), moderate (70–78), weak (63–69), or not manifested (62 and below) skills.

Scientific Reasoning in Writing

At the end of the semester, students’ final, submitted undergraduate theses were assessed using BioTAP, which consists of nine rubric dimensions that focus on communication to the broader scientific community and four additional dimensions that focus on the exhibition of topic-specific expertise ( Reynolds et al. , 2009 ). These dimensions, framed as questions, are displayed in Table 1 .

Student theses were assessed on questions 1–9 of BioTAP using the same procedures described in previous studies ( Reynolds and Thompson, 2011 ; Dowd et al. , 2015a , b ). In this study, six raters were trained in the valid, reliable use of BioTAP rubrics. Each dimension was rated on a five-point scale: 1 indicates the dimension is missing, incomplete, or below acceptable standards; 3 indicates that the dimension is adequate but not exhibiting mastery; and 5 indicates that the dimension is excellent and exhibits mastery (intermediate ratings of 2 and 4 are appropriate when different parts of the thesis make a single category challenging). After training, two raters independently assessed each thesis and then discussed their independent ratings with one another to form a consensus rating. The consensus score is not an average score, but rather an agreed-upon, discussion-based score. On a five-point scale, raters independently assessed dimensions to be within 1 point of each other 82.4% of the time before discussion and formed consensus ratings 100% of the time after discussion.

In this study, we consider both categorical (mastery/nonmastery, where a score of 5 corresponds to mastery) and numerical treatments of individual BioTAP scores to better relate the manifestation of critical thinking in BioTAP assessment to all of the prior studies. For comprehensive/cumulative measures of BioTAP, we focus on the partial sum of questions 1–5, as these questions relate to higher-order scientific reasoning (whereas questions 6–9 relate to mid- and lower-order writing mechanics [ Reynolds et al. , 2009 ]), and the factor scores (i.e., numerical representations of the extent to which each student exhibits the underlying factor), which are calculated from the factor loadings published by Dowd et al. (2016) . We do not focus on questions 6–9 individually in statistical analyses, because we do not expect critical-thinking skills to relate to mid- and lower-order writing skills.

The final, submitted thesis reflects the student’s writing, the student’s scientific reasoning, the quality of feedback provided to the student by peers and mentors, and the student’s ability to incorporate that feedback into his or her work. Therefore, our assessment is not the same as an assessment of unpolished, unrevised samples of students’ written work. While one might imagine that such an unpolished sample may be more strongly correlated with critical-thinking skills measured by the CCTST, we argue that the complete, submitted thesis, assessed using BioTAP, is ultimately a more appropriate reflection of how students exhibit science reasoning in the scientific community.

Statistical Analyses

We took several steps to analyze the collected data. First, to provide context for subsequent interpretations, we generated descriptive statistics for the CCTST scores of the participants based on the norms for undergraduate CCTST test takers. To determine the strength of relationships among CCTST dimensions (including overall score) and the BioTAP dimensions, partial-sum score (questions 1–5), and factor score, we calculated Pearson’s correlations for each pair of measures. To examine whether falling on one side of the nonmastery/mastery threshold (as opposed to a linear scale of performance) was related to critical thinking, we grouped BioTAP dimensions into categories (mastery/nonmastery) and conducted Student’s t tests to compare the means scores of the two groups on each of the seven dimensions and overall score of the CCTST. Finally, for the strongest relationship that emerged, we included additional academic and background variables as covariates in multiple linear-regression analysis to explore questions about how much observed relationships between critical-thinking skills and science reasoning in writing might be explained by variation in these other factors.

Although BioTAP scores represent discreet, ordinal bins, the five-point scale is intended to capture an underlying continuous construct (from inadequate to exhibiting mastery). It has been argued that five categories is an appropriate cutoff for treating ordinal variables as pseudo-continuous ( Rhemtulla et al. , 2012 )—and therefore using continuous-variable statistical methods (e.g., Pearson’s correlations)—as long as the underlying assumption that ordinal scores are linearly distributed is valid. Although we have no way to statistically test this assumption, we interpret adequate scores to be approximately halfway between inadequate and mastery scores, resulting in a linear scale. In part because this assumption is subject to disagreement, we also consider and interpret a categorical (mastery/nonmastery) treatment of BioTAP variables.

We corrected for multiple comparisons using the Holm-Bonferroni method ( Holm, 1979 ). At the most general level, where we consider the single, comprehensive measures for BioTAP (partial-sum and factor score) and the CCTST (overall score), there is no need to correct for multiple comparisons, because the multiple, individual dimensions are collapsed into single dimensions. When we considered individual CCTST dimensions in relation to comprehensive measures for BioTAP, we accounted for seven comparisons; similarly, when we considered individual dimensions of BioTAP in relation to overall CCTST score, we accounted for five comparisons. When all seven CCTST and five BioTAP dimensions were examined individually and without prior knowledge, we accounted for 35 comparisons; such a rigorous threshold is likely to reject weak and moderate relationships, but it is appropriate if there are no specific pre-existing hypotheses. All p values are presented in tables for complete transparency, and we carefully consider the implications of our interpretation of these data in the Discussion section.

CCTST scores for students in this sample ranged from the 39th to 99th percentile of the general population of undergraduate CCTST test takers (mean percentile = 84.3, median = 85th percentile; Table 2 ); these percentiles reflect overall scores that range from moderate to superior. Scores on individual dimensions and overall scores were sufficiently normal and far enough from the ceiling of the scale to justify subsequent statistical analyses.

a Scores correspond to superior (86–100), strong (79–85), moderate (70–78), weak (63–69), or not manifested (62 and lower) skills.

The Pearson’s correlations between students’ cumulative scores on BioTAP (the factor score based on loadings published by Dowd et al. , 2016 , and the partial sum of scores on questions 1–5) and students’ overall scores on the CCTST are presented in Table 3 . We found that the partial-sum measure of BioTAP was significantly related to the overall measure of critical thinking ( r = 0.27, p = 0.03), while the BioTAP factor score was marginally related to overall CCTST ( r = 0.24, p = 0.05). When we looked at relationships between comprehensive BioTAP measures and scores for individual dimensions of the CCTST ( Table 3 ), we found significant positive correlations between the both BioTAP partial-sum and factor scores and CCTST inference ( r = 0.45, p < 0.001, and r = 0.41, p < 0.001, respectively). Although some other relationships have p values below 0.05 (e.g., the correlations between BioTAP partial-sum scores and CCTST induction and interpretation scores), they are not significant when we correct for multiple comparisons.

a In each cell, the top number is the correlation, and the bottom, italicized number is the associated p value. Correlations that are statistically significant after correcting for multiple comparisons are shown in bold.

b This is the partial sum of BioTAP scores on questions 1–5.

c This is the factor score calculated from factor loadings published by Dowd et al. (2016) .

When we expanded comparisons to include all 35 potential correlations among individual BioTAP and CCTST dimensions—and, accordingly, corrected for 35 comparisons—we did not find any additional statistically significant relationships. The Pearson’s correlations between students’ scores on each dimension of BioTAP and students’ scores on each dimension of the CCTST range from −0.11 to 0.35 ( Table 3 ); although the relationship between discussion of implications (BioTAP question 5) and inference appears to be relatively large ( r = 0.35), it is not significant ( p = 0.005; the Holm-Bonferroni cutoff is 0.00143). We found no statistically significant relationships between BioTAP questions 6–9 and CCTST dimensions (unpublished data), regardless of whether we correct for multiple comparisons.

The results of Student’s t tests comparing scores on each dimension of the CCTST of students who exhibit mastery with those of students who do not exhibit mastery on each dimension of BioTAP are presented in Table 4 . Focusing first on the overall CCTST scores, we found that the difference between those who exhibit mastery and those who do not in discussing implications of results (BioTAP question 5) is statistically significant ( t = 2.73, p = 0.008, d = 0.71). When we expanded t tests to include all 35 comparisons—and, like above, corrected for 35 comparisons—we found a significant difference in inference scores between students who exhibit mastery on question 5 and students who do not ( t = 3.41, p = 0.0012, d = 0.88), as well as a marginally significant difference in these students’ induction scores ( t = 3.26, p = 0.0018, d = 0.84; the Holm-Bonferroni cutoff is p = 0.00147). Cohen’s d effect sizes, which reveal the strength of the differences for statistically significant relationships, range from 0.71 to 0.88.

a In each cell, the top number is the t statistic for each comparison, and the middle, italicized number is the associated p value. The bottom number is the effect size. Correlations that are statistically significant after correcting for multiple comparisons are shown in bold.

Finally, we more closely examined the strongest relationship that we observed, which was between the CCTST dimension of inference and the BioTAP partial-sum composite score (shown in Table 3 ), using multiple regression analysis ( Table 5 ). Focusing on the 52 students for whom we have background information, we looked at the simple relationship between BioTAP and inference (model 1), a robust background model including multiple covariates that one might expect to explain some part of the variation in BioTAP (model 2), and a combined model including all variables (model 3). As model 3 shows, the covariates explain very little variation in BioTAP scores, and the relationship between inference and BioTAP persists even in the presence of all of the covariates.

** p < 0.01.

*** p < 0.001.

The aim of this study was to examine the extent to which the various components of scientific reasoning—manifested in writing in the genre of undergraduate thesis and assessed using BioTAP—draw on general and specific critical-thinking skills (assessed using CCTST) and to consider the implications for educational practices. Although science reasoning involves critical-thinking skills, it also relates to conceptual knowledge and the epistemological foundations of science disciplines ( Kuhn et al. , 2008 ). Moreover, science reasoning in writing , captured in students’ undergraduate theses, reflects habits, conventions, and the incorporation of feedback that may alter evidence of individuals’ critical-thinking skills. Our findings, however, provide empirical evidence that cumulative measures of science reasoning in writing are nonetheless related to students’ overall critical-thinking skills ( Table 3 ). The particularly significant roles of inference skills ( Table 3 ) and the discussion of implications of results (BioTAP question 5; Table 4 ) provide a basis for more specific ideas about how these constructs relate to one another and what educational interventions may have the most success in fostering these skills.

Our results build on previous findings. The genre of thesis writing combines pedagogies of writing and inquiry found to foster scientific reasoning ( Reynolds et al. , 2012 ) and critical thinking ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ; Stephenson and Sadler-McKnight, 2016 ). Quitadamo and Kurtz (2007) reported that students who engaged in a laboratory writing component in a general education biology course significantly improved their inference and analysis skills, and Quitadamo and colleagues (2008) found that participation in a community-based inquiry biology course (that included a writing component) was associated with significant gains in students’ inference and evaluation skills. The shared focus on inference is noteworthy, because these prior studies actually differ from the current study; the former considered critical-­thinking skills as the primary learning outcome of writing-­focused interventions, whereas the latter focused on emergent links between two learning outcomes (science reasoning in writing and critical thinking). In other words, inference skills are impacted by writing as well as manifested in writing.

Inference focuses on drawing conclusions from argument and evidence. According to the consensus definition of critical thinking, the specific skill of inference includes several processes: querying evidence, conjecturing alternatives, and drawing conclusions. All of these activities are central to the independent research at the core of writing an undergraduate thesis. Indeed, a critical part of what we call “science reasoning in writing” might be characterized as a measure of students’ ability to infer and make meaning of information and findings. Because the cumulative BioTAP measures distill underlying similarities and, to an extent, suppress unique aspects of individual dimensions, we argue that it is appropriate to relate inference to scientific reasoning in writing . Even when we control for other potentially relevant background characteristics, the relationship is strong ( Table 5 ).

In taking the complementary view and focusing on BioTAP, when we compared students who exhibit mastery with those who do not, we found that the specific dimension of “discussing the implications of results” (question 5) differentiates students’ performance on several critical-thinking skills. To achieve mastery on this dimension, students must make connections between their results and other published studies and discuss the future directions of the research; in short, they must demonstrate an understanding of the bigger picture. The specific relationship between question 5 and inference is the strongest observed among all individual comparisons. Altogether, perhaps more than any other BioTAP dimension, this aspect of students’ writing provides a clear view of the role of students’ critical-thinking skills (particularly inference and, marginally, induction) in science reasoning.

While inference and discussion of implications emerge as particularly strongly related dimensions in this work, we note that the strongest contribution to “science reasoning in writing in biology,” as determined through exploratory factor analysis, is “argument for the significance of research” (BioTAP question 2, not question 5; Dowd et al. , 2016 ). Question 2 is not clearly related to critical-thinking skills. These findings are not contradictory, but rather suggest that the epistemological and disciplinary-specific aspects of science reasoning that emerge in writing through BioTAP are not completely aligned with aspects related to critical thinking. In other words, science reasoning in writing is not simply a proxy for those critical-thinking skills that play a role in science reasoning.

In a similar vein, the content-related, epistemological aspects of science reasoning, as well as the conventions associated with writing the undergraduate thesis (including feedback from peers and revision), may explain the lack of significant relationships between some science reasoning dimensions and some critical-thinking skills that might otherwise seem counterintuitive (e.g., BioTAP question 2, which relates to making an argument, and the critical-thinking skill of argument). It is possible that an individual’s critical-thinking skills may explain some variation in a particular BioTAP dimension, but other aspects of science reasoning and practice exert much stronger influence. Although these relationships do not emerge in our analyses, the lack of significant correlation does not mean that there is definitively no correlation. Correcting for multiple comparisons suppresses type 1 error at the expense of exacerbating type 2 error, which, combined with the limited sample size, constrains statistical power and makes weak relationships more difficult to detect. Ultimately, though, the relationships that do emerge highlight places where individuals’ distinct critical-thinking skills emerge most coherently in thesis assessment, which is why we are particularly interested in unpacking those relationships.

We recognize that, because only honors students submit theses at these institutions, this study sample is composed of a selective subset of the larger population of biology majors. Although this is an inherent limitation of focusing on thesis writing, links between our findings and results of other studies (with different populations) suggest that observed relationships may occur more broadly. The goal of improved science reasoning and critical thinking is shared among all biology majors, particularly those engaged in capstone research experiences. So while the implications of this work most directly apply to honors thesis writers, we provisionally suggest that all students could benefit from further study of them.

There are several important implications of this study for science education practices. Students’ inference skills relate to the understanding and effective application of scientific content. The fact that we find no statistically significant relationships between BioTAP questions 6–9 and CCTST dimensions suggests that such mid- to lower-order elements of BioTAP ( Reynolds et al. , 2009 ), which tend to be more structural in nature, do not focus on aspects of the finished thesis that draw strongly on critical thinking. In keeping with prior analyses ( Reynolds and Thompson, 2011 ; Dowd et al. , 2016 ), these findings further reinforce the notion that disciplinary instructors, who are most capable of teaching and assessing scientific reasoning and perhaps least interested in the more mechanical aspects of writing, may nonetheless be best suited to effectively model and assess students’ writing.

The goal of the thesis writing course at both Duke University and the University of Minnesota is not merely to improve thesis scores but to move students’ writing into the category of mastery across BioTAP dimensions. Recognizing that students with differing critical-thinking skills (particularly inference) are more or less likely to achieve mastery in the undergraduate thesis (particularly in discussing implications [question 5]) is important for developing and testing targeted pedagogical interventions to improve learning outcomes for all students.

The competencies characterized by the Vision and Change in Undergraduate Biology Education Initiative provide a general framework for recognizing that science reasoning and critical-thinking skills play key roles in major learning outcomes of science education. Our findings highlight places where science reasoning–related competencies (like “understanding the process of science”) connect to critical-thinking skills and places where critical thinking–related competencies might be manifested in scientific products (such as the ability to discuss implications in scientific writing). We encourage broader efforts to build empirical connections between competencies and pedagogical practices to further improve science education.

One specific implication of this work for science education is to focus on providing opportunities for students to develop their critical-thinking skills (particularly inference). Of course, as this correlational study is not designed to test causality, we do not claim that enhancing students’ inference skills will improve science reasoning in writing. However, as prior work shows that science writing activities influence students’ inference skills ( Quitadamo and Kurtz, 2007 ; Quitadamo et al. , 2008 ), there is reason to test such a hypothesis. Nevertheless, the focus must extend beyond inference as an isolated skill; rather, it is important to relate inference to the foundations of the scientific method ( Miri et al. , 2007 ) in terms of the epistemological appreciation of the functions and coordination of evidence ( Kuhn and Dean, 2004 ; Zeineddin and Abd-El-Khalick, 2010 ; Ding et al. , 2016 ) and disciplinary paradigms of truth and justification ( Moshman, 2015 ).

Although this study is limited to the domain of biology at two institutions with a relatively small number of students, the findings represent a foundational step in the direction of achieving success with more integrated learning outcomes. Hopefully, it will spur greater interest in empirically grounding discussions of the constructs of scientific reasoning and critical-thinking skills.

This study contributes to the efforts to improve science education, for both majors and nonmajors, through an empirically driven analysis of the relationships between scientific reasoning reflected in the genre of thesis writing and critical-thinking skills. This work is rooted in the usefulness of BioTAP as a method 1) to facilitate communication and learning and 2) to assess disciplinary-specific and general dimensions of science reasoning. The findings support the important role of the critical-thinking skill of inference in scientific reasoning in writing, while also highlighting ways in which other aspects of science reasoning (epistemological considerations, writing conventions, etc.) are not significantly related to critical thinking. Future research into the impact of interventions focused on specific critical-thinking skills (i.e., inference) for improved science reasoning in writing will build on this work and its implications for science education.

ACKNOWLEDGMENTS

We acknowledge the contributions of Kelaine Haas and Alexander Motten to the implementation and collection of data. We also thank Mine Çetinkaya-­Rundel for her insights regarding our statistical analyses. This research was funded by National Science Foundation award DUE-1525602.

  • American Association for the Advancement of Science . ( 2011 ). Vision and change in undergraduate biology education: A call to action . Washington, DC Retrieved September 26, 2017, from https://visionandchange.org/files/2013/11/aaas-VISchange-web1113.pdf . Google Scholar
  • August, D. ( 2016 ). California Critical Thinking Skills Test user manual and resource guide . San Jose: Insight Assessment/California Academic Press. Google Scholar
  • Beyer, C. H., Taylor, E., & Gillmore, G. M. ( 2013 ). Inside the undergraduate teaching experience: The University of Washington’s growth in faculty teaching study . Albany, NY: SUNY Press. Google Scholar
  • Bissell, A. N., & Lemons, P. P. ( 2006 ). A new method for assessing critical thinking in the classroom . BioScience , 56 (1), 66–72. https://doi.org/10.1641/0006-3568(2006)056[0066:ANMFAC]2.0.CO;2 . Google Scholar
  • Blattner, N. H., & Frazier, C. L. ( 2002 ). Developing a performance-based assessment of students’ critical thinking skills . Assessing Writing , 8 (1), 47–64. Google Scholar
  • Clase, K. L., Gundlach, E., & Pelaez, N. J. ( 2010 ). Calibrated peer review for computer-assisted learning of biological research competencies . Biochemistry and Molecular Biology Education , 38 (5), 290–295. Medline ,  Google Scholar
  • Condon, W., & Kelly-Riley, D. ( 2004 ). Assessing and teaching what we value: The relationship between college-level writing and critical thinking abilities . Assessing Writing , 9 (1), 56–75. https://doi.org/10.1016/j.asw.2004.01.003 . Google Scholar
  • Ding, L., Wei, X., & Liu, X. ( 2016 ). Variations in university students’ scientific reasoning skills across majors, years, and types of institutions . Research in Science Education , 46 (5), 613–632. https://doi.org/10.1007/s11165-015-9473-y . Google Scholar
  • Dowd, J. E., Connolly, M. P., Thompson, R. J.Jr., & Reynolds, J. A. ( 2015a ). Improved reasoning in undergraduate writing through structured workshops . Journal of Economic Education , 46 (1), 14–27. https://doi.org/10.1080/00220485.2014.978924 . Google Scholar
  • Dowd, J. E., Roy, C. P., Thompson, R. J.Jr., & Reynolds, J. A. ( 2015b ). “On course” for supporting expanded participation and improving scientific reasoning in undergraduate thesis writing . Journal of Chemical Education , 92 (1), 39–45. https://doi.org/10.1021/ed500298r . Google Scholar
  • Dowd, J. E., Thompson, R. J.Jr., & Reynolds, J. A. ( 2016 ). Quantitative genre analysis of undergraduate theses: Uncovering different ways of writing and thinking in science disciplines . WAC Journal , 27 , 36–51. Google Scholar
  • Facione, P. A. ( 1990 ). Critical thinking: a statement of expert consensus for purposes of educational assessment and instruction. Research findings and recommendations . Newark, DE: American Philosophical Association. Retrieved September 26, 2017, from https://philpapers.org/archive/FACCTA.pdf . Google Scholar
  • Gerdeman, R. D., Russell, A. A., Worden, K. J., Gerdeman, R. D., Russell, A. A., & Worden, K. J. ( 2007 ). Web-based student writing and reviewing in a large biology lecture course . Journal of College Science Teaching , 36 (5), 46–52. Google Scholar
  • Greenhoot, A. F., Semb, G., Colombo, J., & Schreiber, T. ( 2004 ). Prior beliefs and methodological concepts in scientific reasoning . Applied Cognitive Psychology , 18 (2), 203–221. https://doi.org/10.1002/acp.959 . Google Scholar
  • Haaga, D. A. F. ( 1993 ). Peer review of term papers in graduate psychology courses . Teaching of Psychology , 20 (1), 28–32. https://doi.org/10.1207/s15328023top2001_5 . Google Scholar
  • Halonen, J. S., Bosack, T., Clay, S., McCarthy, M., Dunn, D. S., Hill, G. W., … Whitlock, K. ( 2003 ). A rubric for learning, teaching, and assessing scientific inquiry in psychology . Teaching of Psychology , 30 (3), 196–208. https://doi.org/10.1207/S15328023TOP3003_01 . Google Scholar
  • Hand, B., & Keys, C. W. ( 1999 ). Inquiry investigation . Science Teacher , 66 (4), 27–29. Google Scholar
  • Holm, S. ( 1979 ). A simple sequentially rejective multiple test procedure . Scandinavian Journal of Statistics , 6 (2), 65–70. Google Scholar
  • Holyoak, K. J., & Morrison, R. G. ( 2005 ). The Cambridge handbook of thinking and reasoning . New York: Cambridge University Press. Google Scholar
  • Insight Assessment . ( 2016a ). California Critical Thinking Skills Test (CCTST) Retrieved September 26, 2017, from www.insightassessment.com/Products/Products-Summary/Critical-Thinking-Skills-Tests/California-Critical-Thinking-Skills-Test-CCTST . Google Scholar
  • Insight Assessment . ( 2016b ). Sample thinking skills questions. Retrieved September 26, 2017, from www.insightassessment.com/Resources/Teaching-Training-and-Learning-Tools/node_1487 . Google Scholar
  • Kelly, G. J., & Takao, A. ( 2002 ). Epistemic levels in argument: An analysis of university oceanography students’ use of evidence in writing . Science Education , 86 (3), 314–342. https://doi.org/10.1002/sce.10024 . Google Scholar
  • Kuhn, D., & Dean, D.Jr. ( 2004 ). Connecting scientific reasoning and causal inference . Journal of Cognition and Development , 5 (2), 261–288. https://doi.org/10.1207/s15327647jcd0502_5 . Google Scholar
  • Kuhn, D., Iordanou, K., Pease, M., & Wirkala, C. ( 2008 ). Beyond control of variables: What needs to develop to achieve skilled scientific thinking? . Cognitive Development , 23 (4), 435–451. https://doi.org/10.1016/j.cogdev.2008.09.006 . Google Scholar
  • Lawson, A. E. ( 2010 ). Basic inferences of scientific reasoning, argumentation, and discovery . Science Education , 94 (2), 336–364. https://doi.org/­10.1002/sce.20357 . Google Scholar
  • Meizlish, D., LaVaque-Manty, D., Silver, N., & Kaplan, M. ( 2013 ). Think like/write like: Metacognitive strategies to foster students’ development as disciplinary thinkers and writers . In Thompson, R. J. (Ed.), Changing the conversation about higher education (pp. 53–73). Lanham, MD: Rowman & Littlefield. Google Scholar
  • Miri, B., David, B.-C., & Uri, Z. ( 2007 ). Purposely teaching for the promotion of higher-order thinking skills: A case of critical thinking . Research in Science Education , 37 (4), 353–369. https://doi.org/10.1007/s11165-006-9029-2 . Google Scholar
  • Moshman, D. ( 2015 ). Epistemic cognition and development: The psychology of justification and truth . New York: Psychology Press. Google Scholar
  • National Research Council . ( 2000 ). How people learn: Brain, mind, experience, and school . Expanded ed.. Washington, DC: National Academies Press. Google Scholar
  • Pukkila, P. J. ( 2004 ). Introducing student inquiry in large introductory genetics classes . Genetics , 166 (1), 11–18. https://doi.org/10.1534/genetics.166.1.11 . Medline ,  Google Scholar
  • Quitadamo, I. J., Faiola, C. L., Johnson, J. E., & Kurtz, M. J. ( 2008 ). Community-based inquiry improves critical thinking in general education biology . CBE—Life Sciences Education , 7 (3), 327–337. https://doi.org/10.1187/cbe.07-11-0097 . Link ,  Google Scholar
  • Quitadamo, I. J., & Kurtz, M. J. ( 2007 ). Learning to improve: Using writing to increase critical thinking performance in general education biology . CBE—Life Sciences Education , 6 (2), 140–154. https://doi.org/10.1187/cbe.06-11-0203 . Link ,  Google Scholar
  • Reynolds, J. A., Smith, R., Moskovitz, C., & Sayle, A. ( 2009 ). BioTAP: A systematic approach to teaching scientific writing and evaluating undergraduate theses . BioScience , 59 (10), 896–903. https://doi.org/10.1525/bio.2009.59.10.11 . Google Scholar
  • Reynolds, J. A., Thaiss, C., Katkin, W., & Thompson, R. J. ( 2012 ). Writing-to-learn in undergraduate science education: A community-based, conceptually driven approach . CBE—Life Sciences Education , 11 (1), 17–25. https://doi.org/10.1187/cbe.11-08-0064 . Link ,  Google Scholar
  • Reynolds, J. A., & Thompson, R. J. ( 2011 ). Want to improve undergraduate thesis writing? Engage students and their faculty readers in scientific peer review . CBE—Life Sciences Education , 10 (2), 209–215. https://doi.org/­10.1187/cbe.10-10-0127 . Link ,  Google Scholar
  • Rhemtulla, M., Brosseau-Liard, P. E., & Savalei, V. ( 2012 ). When can categorical variables be treated as continuous? A comparison of robust continuous and categorical SEM estimation methods under suboptimal conditions . Psychological Methods , 17 (3), 354–373. https://doi.org/­10.1037/a0029315 . Medline ,  Google Scholar
  • Stephenson, N. S., & Sadler-McKnight, N. P. ( 2016 ). Developing critical thinking skills using the science writing heuristic in the chemistry laboratory . Chemistry Education Research and Practice , 17 (1), 72–79. https://doi.org/­10.1039/C5RP00102A . Google Scholar
  • Tariq, V. N., Stefani, L. A. J., Butcher, A. C., & Heylings, D. J. A. ( 1998 ). Developing a new approach to the assessment of project work . Assessment and Evaluation in Higher Education , 23 (3), 221–240. https://doi.org/­10.1080/0260293980230301 . Google Scholar
  • Timmerman, B. E. C., Strickland, D. C., Johnson, R. L., & Payne, J. R. ( 2011 ). Development of a “universal” rubric for assessing undergraduates’ scientific reasoning skills using scientific writing . Assessment and Evaluation in Higher Education , 36 (5), 509–547. https://doi.org/10.1080/­02602930903540991 . Google Scholar
  • Topping, K. J., Smith, E. F., Swanson, I., & Elliot, A. ( 2000 ). Formative peer assessment of academic writing between postgraduate students . Assessment and Evaluation in Higher Education , 25 (2), 149–169. https://doi.org/10.1080/713611428 . Google Scholar
  • Willison, J., & O’Regan, K. ( 2007 ). Commonly known, commonly not known, totally unknown: A framework for students becoming researchers . Higher Education Research and Development , 26 (4), 393–409. https://doi.org/10.1080/07294360701658609 . Google Scholar
  • Woodin, T., Carter, V. C., & Fletcher, L. ( 2010 ). Vision and Change in Biology Undergraduate Education: A Call for Action—Initial responses . CBE—Life Sciences Education , 9 (2), 71–73. https://doi.org/10.1187/cbe.10-03-0044 . Link ,  Google Scholar
  • Zeineddin, A., & Abd-El-Khalick, F. ( 2010 ). Scientific reasoning and epistemological commitments: Coordination of theory and evidence among college science students . Journal of Research in Science Teaching , 47 (9), 1064–1093. https://doi.org/10.1002/tea.20368 . Google Scholar
  • Zimmerman, C. ( 2000 ). The development of scientific reasoning skills . Developmental Review , 20 (1), 99–149. https://doi.org/10.1006/drev.1999.0497 . Google Scholar
  • Zimmerman, C. ( 2007 ). The development of scientific thinking skills in elementary and middle school . Developmental Review , 27 (2), 172–223. https://doi.org/10.1016/j.dr.2006.12.001 . Google Scholar
  • Redesign a biology introduction course to promote life consciousness 14 May 2024 | Cogent Education, Vol. 11, No. 1
  • Gender, Equity, and Science Writing: Examining Differences in Undergraduate Life Science Majors’ Attitudes toward Writing Lab Reports 6 March 2024 | Education Sciences, Vol. 14, No. 3
  • Designing a framework to improve critical reflection writing in teacher education using action research 24 February 2022 | Educational Action Research, Vol. 32, No. 1
  • Scientific Thinking and Critical Thinking in Science Education  5 September 2023 | Science & Education, Vol. 11
  • Students Need More than Content Knowledge To Counter Vaccine Hesitancy Journal of Microbiology & Biology Education, Vol. 24, No. 2
  • Critical thinking during science investigations: what do practicing teachers value and observe? 16 March 2023 | Teachers and Teaching, Vol. 29, No. 6
  • Effect of Web-Based Collaborative Learning Method with Scratch on Critical Thinking Skills of 5th Grade Students 30 March 2023 | Participatory Educational Research, Vol. 10, No. 2
  • Are We on the Way to Successfully Educating Future Citizens?—A Spotlight on Critical Thinking Skills and Beliefs about the Nature of Science among Pre-Service Biology Teachers in Germany 22 March 2023 | Behavioral Sciences, Vol. 13, No. 3
  • A Systematic Review on Inquiry-Based Writing Instruction in Tertiary Settings 30 November 2022 | Written Communication, Vol. 40, No. 1
  • An empirical analysis of the relationship between nature of science and critical thinking through science definitions and thinking skills 8 December 2022 | SN Social Sciences, Vol. 2, No. 12
  • TEACHING OF CRITICAL THINKING SKILLS BY SCIENCE TEACHERS IN JAPANESE PRIMARY SCHOOLS 25 October 2022 | Journal of Baltic Science Education, Vol. 21, No. 5
  • A Team-Based Activity to Support Knowledge Transfer and Experimental Design Skills of Undergraduate Science Students 4 May 2022 | Journal of Microbiology & Biology Education, Vol. 21
  • Curriculum Design of College Students’ English Critical Ability in the Internet Age Wireless Communications and Mobile Computing, Vol. 2022
  • Exploring the structure of students’ scientific higher order thinking in science education Thinking Skills and Creativity, Vol. 43
  • The Asia-Pacific Education Researcher, Vol. 31, No. 4
  • Conspiratorial Beliefs and Cognitive Styles: An Integrated Look on Analytic Thinking, Critical Thinking, and Scientific Reasoning in Relation to (Dis)trust in Conspiracy Theories 12 October 2021 | Frontiers in Psychology, Vol. 12
  • Professional Knowledge and Self-Efficacy Expectations of Pre-Service Teachers Regarding Scientific Reasoning and Diagnostics 11 October 2021 | Education Sciences, Vol. 11, No. 10
  • Developing textbook based on scientific approach, critical thinking, and science process skills Journal of Physics: Conference Series, Vol. 1839, No. 1
  • Using Models of Cognitive Development to Design College Learning Experiences
  • Thinking Skills and Creativity, Vol. 42
  • Assessing students’ prior knowledge on critical thinking skills in the biology classroom: Has it already been good?
  • Critical Thinking Level among Medical Sciences Students in Iran Education Research International, Vol. 2020
  • Teaching during a pandemic: Using high‐impact writing assignments to balance rigor, engagement, flexibility, and workload 12 October 2020 | Ecology and Evolution, Vol. 10, No. 22
  • Mini-Review - Teaching Writing in the Undergraduate Neuroscience Curriculum: Its Importance and Best Practices Neuroscience Letters, Vol. 737
  • Developing critical thinking skills assessment for pre-service elementary school teacher about the basic concept of science: validity and reliability Journal of Physics: Conference Series, Vol. 1567, No. 2
  • Challenging endocrinology students with a critical-thinking workbook Advances in Physiology Education, Vol. 44, No. 1
  • Jason E. Dowd ,
  • Robert J. Thompson ,
  • Leslie Schiff ,
  • Kelaine Haas ,
  • Christine Hohmann ,
  • Chris Roy ,
  • Warren Meck ,
  • John Bruno , and
  • Rebecca Price, Monitoring Editor
  • Kari L. Nelson ,
  • Claudia M. Rauter , and
  • Christine E. Cutucache
  • Elisabeth Schussler, Monitoring Editor

Submitted: 17 March 2017 Revised: 19 October 2017 Accepted: 20 October 2017

© 2018 J. E. Dowd et al. CBE—Life Sciences Education © 2018 The American Society for Cell Biology. This article is distributed by The American Society for Cell Biology under license from the author(s). It is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).

Critical thinking definition

discuss the relationship between critical thinking and problem solving

Critical thinking, as described by Oxford Languages, is the objective analysis and evaluation of an issue in order to form a judgement.

Active and skillful approach, evaluation, assessment, synthesis, and/or evaluation of information obtained from, or made by, observation, knowledge, reflection, acumen or conversation, as a guide to belief and action, requires the critical thinking process, which is why it's often used in education and academics.

Some even may view it as a backbone of modern thought.

However, it's a skill, and skills must be trained and encouraged to be used at its full potential.

People turn up to various approaches in improving their critical thinking, like:

  • Developing technical and problem-solving skills
  • Engaging in more active listening
  • Actively questioning their assumptions and beliefs
  • Seeking out more diversity of thought
  • Opening up their curiosity in an intellectual way etc.

Is critical thinking useful in writing?

Critical thinking can help in planning your paper and making it more concise, but it's not obvious at first. We carefully pinpointed some the questions you should ask yourself when boosting critical thinking in writing:

  • What information should be included?
  • Which information resources should the author look to?
  • What degree of technical knowledge should the report assume its audience has?
  • What is the most effective way to show information?
  • How should the report be organized?
  • How should it be designed?
  • What tone and level of language difficulty should the document have?

Usage of critical thinking comes down not only to the outline of your paper, it also begs the question: How can we use critical thinking solving problems in our writing's topic?

Let's say, you have a Powerpoint on how critical thinking can reduce poverty in the United States. You'll primarily have to define critical thinking for the viewers, as well as use a lot of critical thinking questions and synonyms to get them to be familiar with your methods and start the thinking process behind it.

Are there any services that can help me use more critical thinking?

We understand that it's difficult to learn how to use critical thinking more effectively in just one article, but our service is here to help.

We are a team specializing in writing essays and other assignments for college students and all other types of customers who need a helping hand in its making. We cover a great range of topics, offer perfect quality work, always deliver on time and aim to leave our customers completely satisfied with what they ordered.

The ordering process is fully online, and it goes as follows:

  • Select the topic and the deadline of your essay.
  • Provide us with any details, requirements, statements that should be emphasized or particular parts of the essay writing process you struggle with.
  • Leave the email address, where your completed order will be sent to.
  • Select your prefered payment type, sit back and relax!

With lots of experience on the market, professionally degreed essay writers , online 24/7 customer support and incredibly low prices, you won't find a service offering a better deal than ours.

discuss the relationship between critical thinking and problem solving

Explained: Importance of critical thinking, problem-solving skills in curriculum

F uture careers are no longer about domain expertise or technical skills. Rather, critical thinking and problem-solving skills in employees are on the wish list of every big organization today. Even curriculums and pedagogies across the globe and within India are now requiring skilled workers who are able to think critically and are analytical.

The reason for this shift in perspective is very simple.

These skills provide a staunch foundation for comprehensive learning that extends beyond books or the four walls of the classroom. In a nutshell, critical thinking and problem-solving skills are a part of '21st Century Skills' that can help unlock valuable learning for life.

Over the years, the education system has been moving away from the system of rote and other conventional teaching and learning parameters.

They are aligning their curriculums to the changing scenario which is becoming more tech-driven and demands a fusion of critical skills, life skills, values, and domain expertise. There's no set formula for success.

Rather, there's a defined need for humans to be more creative, innovative, adaptive, agile, risk-taking, and have a problem-solving mindset.

In today's scenario, critical thinking and problem-solving skills have become more important because they open the human mind to multiple possibilities, solutions, and a mindset that is interdisciplinary in nature.

Therefore, many schools and educational institutions are deploying AI and immersive learning experiences via gaming, and AR-VR technologies to give a more realistic and hands-on learning experience to their students that hone these abilities and help them overcome any doubt or fear.

ADVANTAGES OF CRITICAL THINKING AND PROBLEM-SOLVING IN CURRICULUM

Ability to relate to the real world:  Instead of theoretical knowledge, critical thinking, and problem-solving skills encourage students to look at their immediate and extended environment through a spirit of questioning, curiosity, and learning. When the curriculum presents students with real-world problems, the learning is immense.

Confidence, agility & collaboration : Critical thinking and problem-solving skills boost self-belief and confidence as students examine, re-examine, and sometimes fail or succeed while attempting to do something.

They are able to understand where they may have gone wrong, attempt new approaches, ask their peers for feedback and even seek their opinion, work together as a team, and learn to face any challenge by responding to it.

Willingness to try new things: When problem-solving skills and critical thinking are encouraged by teachers, they set a robust foundation for young learners to experiment, think out of the box, and be more innovative and creative besides looking for new ways to upskill.

It's important to understand that merely introducing these skills into the curriculum is not enough. Schools and educational institutions must have upskilling workshops and conduct special training for teachers so as to ensure that they are skilled and familiarized with new teaching and learning techniques and new-age concepts that can be used in the classrooms via assignments and projects.

Critical thinking and problem-solving skills are two of the most sought-after skills. Hence, schools should emphasise the upskilling of students as a part of the academic curriculum.

The article is authored by Dr Tassos Anastasiades, Principal- IB, Genesis Global School, Noida. 

Watch Live TV in English

Watch Live TV in Hindi

Explained: Importance of critical thinking, problem-solving skills in curriculum

COMMENTS

  1. Critical Thinking versus Problem Solving

    The first step to enhancing your critical thinking and problem solving skills is to think about them, become aware of them, then you can actively practice to improve them. Critical thinking and problem-solving are two important "soft" or essential skills hiring managers are looking for. According to a Linkedin survey, 57% of business ...

  2. Critical Thinking vs. Problem-Solving: What's the Difference?

    Critical thinking and problem-solving can both help you resolve challenges, but the two practices have distinct purposes and strategies. Here are some differences between the two skills: Critical thinking This is a mode of thinking, compared to problem-solving, which is a set of solution-oriented strategies. Since critical thinking strengthens ...

  3. Critical Thinking and Decision-Making

    Simply put, critical thinking is the act of deliberately analyzing information so that you can make better judgements and decisions. It involves using things like logic, reasoning, and creativity, to draw conclusions and generally understand things better. This may sound like a pretty broad definition, and that's because critical thinking is a ...

  4. 5.3: Using Critical Thinking Skills- Decision Making and Problem Solving

    Using Critical Thinking Skills in Problem Solving. Think of problem solving as a process with four Ps: Define the problem, generate possibilities,. create a plan, and perform your plan.. Step 1: Define the problem. To define a problem effectively, understand what a problem is—a mismatch between what you want and what you have.

  5. Critical Thinking and Problem-Solving

    Critical thinking involves asking questions, defining a problem, examining evidence, analyzing assumptions and biases, avoiding emotional reasoning, avoiding oversimplification, considering other interpretations, and tolerating ambiguity. Dealing with ambiguity is also seen by Strohm & Baukus (1995) as an essential part of critical thinking ...

  6. Critical Thinking

    Critical Thinking. Critical thinking is a widely accepted educational goal. Its definition is contested, but the competing definitions can be understood as differing conceptions of the same basic concept: careful thinking directed to a goal. Conceptions differ with respect to the scope of such thinking, the type of goal, the criteria and norms ...

  7. Critical Thinking and Problem-Solving Skills

    Critical thinking helps people solve problems systematically using facts and data. It involves going through a process of gathering information, analyzing, evaluating, and synthesizing the information to help solve problems in a timely manner. Logic plays a predominant role in critical thinking, because the goal of critical thinking is to solve ...

  8. What Are Critical Thinking Skills and Why Are They Important?

    Problem-solving: Problem-solving is perhaps the most important skill that critical thinkers can possess. The ability to solve issues and bounce back from conflict is what helps you succeed, be a leader, and effect change. ... Critical thinking, in part, is the cognitive process of reading the situation: the words coming out of their mouth ...

  9. Cognition and Instruction/Problem Solving, Critical Thinking and

    Table 1. Summarizes the difference between well-defined and ill-defined problems. Differences in Solving Ill-defined and Well-defined Problems [edit | edit source]. In earlier times, researchers assumed both types of problems were solved in similar ways , more contemporary research highlights some distinct differences between processes behind finding a solution.

  10. Critical Thinking, Decision Making, and Problem Solving

    The "set" interfered with their problem solving. Subsequently Luchins divided another group into two subgroups. The first sub-group worked through the problem in the usual way, using the three jar solution. The second sub-group were told to think more carefully about how to solve the problems.

  11. Decision-making and Problem-solving

    Appreciate the Complexities Involved in Decision-Making & Problem Solving. Develop evidence to support views. Analyze situations carefully. Discuss subjects in an organized way. Predict the consequences of actions. Weigh alternatives. Generate and organize ideas. Form and apply concepts. Design systematic plans of action. A 5-Step Problem ...

  12. Critical thinking

    Critical thinking is the analysis of available facts, evidence, observations, and arguments in order to form a judgement by the application of rational, skeptical, and unbiased analyses and evaluation. The application of critical thinking includes self-directed, self-disciplined, self-monitored, and self-corrective habits of the mind, thus a critical thinker is a person who practices the ...

  13. Understanding the Complex Relationship between Critical Thinking and

    The advances in cognitive science in the 21st century have increased our understanding of the mental processes involved in thinking and reasoning, as well as memory, learning, and problem solving. Critical thinking is understood to include both a cognitive dimension and a disposition dimension (e.g., reflective thinking) and is defined as ...

  14. Relationship between Problem Solving and Critical Thinking (adapted

    The study was carried out to assess the relationship between students' critical thinking abilities and their success in writing an argumentative essay. The descriptive-correlation design was used ...

  15. PDF The complex relationship between students' critical thinking and ...

    The complex relationship between students' critical thinking and epistemological beliefs in the context of problem solving Heidi Hyytinena, Katariina Holmab, Auli Tooma, Richard J ... The study utilized a multi-method approach to explore the connection between critical thinking and epistemological beliefs in a specific problem-solving ...

  16. Understanding the Complex Relationship between Critical Thinking and

    understand the relationship between students' critical-think- ... the University of Minnesota and Duke University. In the following subsections, we discuss in greater detail the constructs of scientific reasoning and critical thinking, as well as ... problem solving. Critical thinking is understood to include both a cognitive dimension and a ...

  17. PDF Problem Solving and Critical Thinking

    Problem Solving and Critical Thinking Everyone experiences problems from time to time. Some of our problems are big and complicated, while others may be more easily solved. There is no shortage of challenges and issues that can arise on the job. ... Discuss the difference between praise, criticism, and feedback and ask participants for examples ...

  18. Collaborative Learning and Critical Thinking

    Critical thinking and collaboration are intricately linked to realize gains in productivity. Let us individually examine the constructs of collaboration and critical thinking and next discuss the relationship between them. The training to work effectively in teams should begin at an early age in school and continued through college.

  19. PDF CHAPTER 1 What Is Critical Thinking, Clinical Reasoning, and Clinical

    another: critical thinking, clinical reasoning, clinical judgment, decision-making, problem-solving, and nursing process. 4. Identify four principles of the scientifi c method that are evident in CT. 5. Compare and contrast the terms problem-focused thinking and outcome-focused thinking. 6. Clarify the term critical thinking indicator (CTI). 7.

  20. The complex relationship between students' critical thinking and

    The study utilized a multi-method approach to explore the connection between critical thinking and epistemological beliefs in a specific problem-solving situation.

  21. Critical Thinking: A Simple Guide and Why It's Important

    ☑ Problem-Solving Mastery. Visualize critical thinking as the Sherlock Holmes of your career journey. It facilitates swift problem resolution akin to a detective unraveling a mystery. By methodically analyzing situations and deconstructing complexities, critical thinkers emerge as adept problem solvers, rendering them invaluable assets in the ...

  22. Understanding the Complex Relationship between Critical Thinking and

    The advances in cognitive science in the 21st century have increased our understanding of the mental processes involved in thinking and reasoning, as well as memory, learning, and problem solving. Critical thinking is understood to include both a cognitive dimension and a disposition dimension (e.g., reflective thinking) and is defined as ...

  23. Boost Remote Work Problem-Solving with Critical Thinking

    Here's how you can enhance problem-solving in remote work situations with critical thinking. Powered by AI and the LinkedIn community. 1. Define Issues. 2. Gather Data. Be the first to add your ...

  24. Using Critical Thinking in Essays and other Assignments

    Critical thinking, as described by Oxford Languages, is the objective analysis and evaluation of an issue in order to form a judgement. Active and skillful approach, evaluation, assessment, synthesis, and/or evaluation of information obtained from, or made by, observation, knowledge, reflection, acumen or conversation, as a guide to belief and action, requires the critical thinking process ...

  25. Explained: Importance of critical thinking, problem-solving skills in

    In a nutshell, critical thinking and problem-solving skills are a part of '21st Century Skills' that can help unlock valuable learning for life. Over the years, the education system has been ...

  26. Chapter 15: Critical Thinking in Nursing Practice Flashcards

    a. In basic critical thinking, the learner trusts that experts have the right answers for every problem; thinking is concrete and based on a set of rules or principles. b. In complex critical thinking, learners begin to separate themselves from experts and analyze and examine choices more independently. c.