Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Inquiry-based training improves teaching effectiveness of biology teaching assistants

Abstract

Graduate teaching assistants (GTAs) are used extensively as undergraduate science lab instructors at universities, yet they often have having minimal instructional training and little is known about effective training methods. This blind randomized control trial study assessed the impact of two training regimens on GTA teaching effectiveness. GTAs teaching undergraduate biology labs (n = 52) completed five hours of training in either inquiry-based learning pedagogy or general instructional “best practices”. GTA teaching effectiveness was evaluated using: (1) a nine-factor student evaluation of educational quality; (2) a six-factor questionnaire for student learning; and (3) course grades. Ratings from both GTAs and undergraduates indicated that indicated that the inquiry-based learning pedagogy training has a positive effect on GTA teaching effectiveness.

Introduction

Graduate teaching assistant (GTA)-run introductory science courses are the norm at higher education institutions in North America, Australia and New Zealand, and are becoming more prevalent elsewhere [1]. GTAs are part-time employees (often research students) hired to lead lab sessions, grade papers, and provide assistance to course instructors, and account for many of the contact hours undergraduates have with the department. GTAs have a powerful influence on undergraduate student learning, but not without cost [2,3]. Graduate students gain income and teaching experience, and science departments gain a pool of inexpensive laborers versed in discipline-specific content, but because GTA teaching can be inconsistent, and students find it difficult to learn from inexperienced teachers, ensuring that undergraduates receive high-quality teaching from GTAs is one of the main teaching issues facing science departments [47].

One solution includes general teaching training for GTAs, but the effectiveness of this training is mixed [24,8,9]. The training content is not standardized, as it is for schoolteachers, and attendance is not compulsory. A 1997 survey of 153 biology departments in the United States indicated that 49% of GTA programs required no formal training whatsoever [10]. For those programs that do require formalized GTA training, the form that training takes can vary widely, from university-wide mass orientation workshops to subject-based instruction from instructors, supervisors or peers [4]. Content might be as important as format – a 2004 study cited the lack of adequate preparation for facilitating open inquiry labs as a major difficulty both for undergraduate students and the GTAs themselves [5].

Well-constructed science lab activities are powerful learning tools; through guided inquiry, undergraduates gain first-person experience of scientific principles and phenomena learned in lectures, and learn to employ experimental methods to solve discrete problems [11]. As such, training that exposes GTAs to the inquiry process may improve teaching effectiveness. Common misperceptions about undergraduate labs include: (a) that the purpose of lab activity is to recapitulate content learned through lectures or from readings; or (b) that the purpose is to familiarize students with cutting-edge experimental techniques. Such objections ignore what is unique about the lab experience – the opportunity to practice applying the scientific method. Many introductory labs deliberately use older, highly predictable experimental setups to teach students basic scientific inquiry skills.

Inquiry-based learning is a pedagogical approach rooted in constructivist learning theories that advocates teaching cognitive skills through open-ended exploration by the learner ([12,13]). Inquiry activities commonly include designing protocols, developing procedures and proposing unified explanations for experimental or anecdotal phenomena. Structured-inquiry activities, which are commonly used as undergraduate lab exercises, typically ask students to propose specific explanations, rooted in relevant course content, for data gathered from experimental trials [14,15]. Teaching “facts”, “knowledge” or “content” is only the secondary objective of such activities: the main goal is to develop the reasoning skills that are required to plan, execute and interpret scientific experiments.

Successful inquiry-oriented initiatives such as POGIL (Process Oriented Guided Inquiry Learning) emphasize the coordination of content and methodological skills during science learning activities [1618], but implementing large-scale reorganization of teaching materials and curricular content calls for substantial investment in time and money, and may be beyond the capability of many course instructors. Moreover, reorienting major introductory course lecture content toward inquiry learning is a considerable effort, and requires departmental consensus to implement effectively. The current study investigated whether it is possible to increase learning effectiveness in an inquiry-based activity without taking on large-scale curricular redesign.

Previous studies have suggested that training GTAs on the use of inquiry-based methods will improve GTA teaching effectiveness in undergraduate labs [19,20]. Although few empirical studies have identified which kinds of GTA training content actually improve general lab GTA teaching effectiveness, and none have specifically focused on inquiry-based methods, some authors have used correlational data or made general recommendations regarding GTA teaching [7,13,21]. Inquiry-based methods may be of particular value to GTAs teaching introductory labs, since labs are designed to teach scientific enquiry as well as course content [2224].

Here, our aim was to test whether grounding GTA training in inquiry-based learning theory would measurably improve teaching effectiveness in undergraduate biology labs, as assessed by undergraduate and GTA responses to online questionnaires, as well as standardized student grades. We predict that explaining inquiry-based methods would improve lab GTA teaching performance, although it might also be reasonable to argue that the limited time allotted might be better spent if allocated to practical teaching activities rather than theoretical ones because GTA training is extremely brief (typically less than five hours per year). In order to test this hypothesis, we conducted a semi-randomized control trial to compare two training regimens for GTAs of introductory biology labs: a control regimen that is taught various “best practices” associated with lab teaching; and an experimental regimen that explained inquiry-oriented pedagogy to GTAs.

Methods

i) Participants

This study took place during the first semester of the academic year at a large research-intensive North American university with about 25,000 students. Of a total number of 126 GTAs, 54 GTAs volunteered to take part in the study and met the study inclusion criteria. Each of these GTAs was employed as a GTA in a single lab section of an introductory biology course in the Fall 2012 semester, and was a graduate student registered in an MSc or PhD programme in the Department of Biology during the same semester. An online pre-assignment questionnaire recorded experience and program data for each GTA two weeks before the onset of the training. On the basis of the data collected in this questionnaire, participants were divided into two groups based on academic program (Masters or PhD). GTAs were allocated to either a “best practice” (Control) or inquiry-based learning pedagogy (Inquiry) training group. This allocation was semi-random; the proportion of GTAs in either academic program was held constant between training groups to balance the experimental design and avoid a confounding factor of academic program, but was otherwise random. Two participants from the control group discontinued the study after the first training session; 52 GTAs finished the training sessions (24 control and 28 inquiry GTAs) and all of these 52 participants completed an online survey conducted at the end of the semester.

GTAs were blind to their training group assignment and were not given specific details about the experimental design. Details about their lab section were collected during the training group workshops and from the online survey. This information allowed for pre-training responses to be matched with later survey responses as well as matching undergraduate students to their lab section GTA while protecting participant confidentiality.

Undergraduate students were recruited to rate their GTAs through advertisement posters, email invitations, and social media announcements. Undergraduates were told that GTAs were part of a training experiment, and were part of more than one GTA training group, but were not told: (a) what the training groups were, or (b) which training group their GTA had been allocated to. A total of 602 undergraduates provided responses, with 352 (58.5%) completing the entire online survey. Approximately 1,250 undergraduate students were enrolled in undergraduate biology during this semester. All GTA lab groups were represented by at least one undergraduate.

At the end of the academic semester, all 16 course professors were contacted by email and notified of the nature of the training experiment (i.e. that GTA training was being manipulated in a controlled experiment – although the training groups, training regimens and the GTAs involved were not revealed), then asked to provide average course grades. Average course marks for 49 of the 52 sections taught by GTAs taking part in the experiment were collected.

ii) Training Workshops

The general format of the training for both groups was the same and the same instructor taught all sessions. Each training group attended a pair of workshops; each workshop lasted 2.5 hours for a total of 5 hours of training. The two workshops were held one week apart. The first workshop was oriented toward teaching and the second was oriented toward grading. All workshops involved a mix of PowerPoint® slides, group discussions, peer-led problem-solving sessions and question-and-answer sessions. Both training groups included the same core topics. The first session included information on: lecturing, moderating group discussions, facilitating lab activities, diagnosing problems in student understanding, giving effective prelab talks, dealing with student questions, and encouraging/motivating students in the lab. The second session included material related to marking: grading lab reports, providing useful feedback, rubric design, and speed grading tips.

Because our aim was to investigate whether inquiry pedagogy training improved GTA teaching effectiveness, the training groups mainly differed in terms of whether inquiry-based learning was explained as a theory of learning to be used in the science lab. Content for both GTA training groups was drawn from previous recommendations [9,25,26] and was common between the two training groups.

The areas of teaching focus for the control (“best practices”) training group seminars were: (1) to identify effective and efficient strategies for GTAs to teach content knowledge to undergraduates during lab activities; and (2) to ensure that lab activity assessments were marked fairly. In general, the material taught to this group was designed to reflect the status quo GTA training curriculum at North American universities, and so “best practices” identified in GTA training materials or educational studies were provided to GTAs [2628]. The training group focused on specific teaching practices of the GTAs themselves rather than explaining how to teach scientific enquiry; because many of the “best practices” for GTAs related to the teaching of content knowledge (i.e. “tips for teaching undergraduate students to use a light microscope”) rather than the scientific method, this training group did not explicitly teach GTAs how to teach higher-order cognitive skills. A full list of the activities that took place in the two control training group seminars (teaching and assessment) is given in Tables 1 and 2.

ItemTime AllocatedActivityDesired Learning OutcomeTeaching Method
130 minGTA General Orientation- understanding the role of GTAsInstructor lecture (Powerpoint)
215 minBrainstorming Activity: Laboratory Learning Outcomes- identification of determinants of lab learning successStudent-led small groups, answers shared with class and discussed
345 minLab teaching best practices- understanding of evidence-based methods for teaching effectivelyInstructor lecture (Powerpoint)
- understanding specific teaching skills, learning styles and learning outcomes
430 minMini-microteaching Activity - Sample Lab Scenarios (light microscope)- understanding of how to apply best practice teaching techniques in lab activity situationsMicroteaching, large group activity
515 minTroubleshooting inquiry-learning activities- how to handle unexpected problems in lab activitiesInstructor lecture (Powerpoint)
615 minQuestions- clear up unresolved questionsStudent-led question and answer

Table 1. Teaching Seminar Activity Chart for Control GTA Training Group.

CSV
Download CSV
ItemTime AllocatedActivityDesired Learning OutcomeTeaching Method
130 minHistory and Purpose of Assessment- understanding of role of assessment in the learning processLecture (Powerpoint)
215 minBrainstorming Activity: Laboratory Assessment Methods- GTA identification of purpose of assessment in labs; criteria determining successful assessment methods for labsStudent-led small groups, answers shared with class and discussed
330 minGrading Best Practices- understanding of evidence-based methods for grading effectivelyInstructor Lecture (Powerpoint)
- understanding specific grading requirements for in-lab and lab report activities
445 minSmall Group Rubric Design- how rubrics are interpreted Student-led small groups, answers shared with class and discussed
- understanding rubric inclusion criteria
- understanding how feedback is different in-lab versus on lab reports
515 minHow to Grade Equitably and Quickly - how to maintain grading consistencyInstructor Lecture (Powerpoint)
- how to grade lab feedback quickly
615 minQuestions- clear up unresolved questionsStudent-led question and answer

Table 2. Assessment Seminar Activity Chart for Control GTA Training Group.

CSV
Download CSV

For the inquiry pedagogy training group, teaching and grading were presented from a constructivist perspective: learning objectives were explained using the revised Bloom’s taxonomy, scaffolding and inquiry facilitation were explained as teaching methods that help students learn methodological skills, and giving feedback was oriented toward encouraging further inquiry [13,20,21,2933]. The areas of teaching focus for the inquiry pedagogy training group seminars were: (1) to teach GTAs how inquiry-based practices teach students to reason independently and apply the scientific method; (2) to teach GTAs how to facilitate structured inquiry and open-ended learning as lab activities; and (3) to teach GTAs how to assess inquiry. Instead of teaching GTAs to teach undergraduates how to use a light microscope, GTAs were told instead how to facilitate and evaluate student-centered inquiry with respect to using a light microscope to ask scientific questions. A full list of the activities that took place in the two inquiry-based learning pedagogy training group seminars (teaching and assessment) is given in Tables 3 and 4.

ItemTime AllocatedActivityDesired Learning OutcomeTeaching Method
130 minGTA General Orientation- understanding the role of GTAsInstructor lecture (Powerpoint)
215 minBrainstorming Activity: Laboratory Learning Outcomes- identification of determinants of lab learning successStudent-led small groups, answers shared with class and discussed
345 minIntroduction to Inquiry-based Learning- understanding of the Revised Bloom’s Taxonomy of learningInstructor lecture (Powerpoint)
- how scientific enquiry is taught using inquiry
- understanding of facilitation as a teaching method
430 minFacilitation Situation Activity - Sample Lab Scenarios- understanding of how to apply facilitation techniques in teaching situationsLarge group activity
515 minTroubleshooting inquiry-learning activities- how to handle unexpected problems in inquiry-based lab activitiesInstructor lecture (Powerpoint)
615 minQuestions- clear up unresolved questionsStudent-led question and answer session

Table 3. Teaching Seminar Activity Chart for Inquiry-based Learning Pedagogy GTA Training Group.

CSV
Download CSV
ItemTime AllocatedActivityDesired Learning OutcomeTeaching Method
130 minHistory and Purpose of Assessment- understanding the role of assessment in the learning processInstructor lecture (Powerpoint)
215 minBrainstorming Activity: Laboratory assessment Methods- identification of: (1) purpose of assessment in labs; (2) criteria determining successful assessment methods for labsStudent-led small groups, answers shared with class and discussed
330 minHow to Assess Scientific Inquiry- understanding of the distinction between lower- and higher-order cognitive skillsInstructor lecture (Powerpoint)
- understanding how to use assessment tolls to give feedback on cognitive skill development
415 minAssessment Rubric Examples- how rubrics are interpretedStudent-led question and answer
545 minSmall Group Rubric Design- how rubrics are interpretedStudent-led small groups, answers shared with class and discussed
- understanding rubric inclusion criteria
- understanding how feedback can assess different cognitive skills
615 minQuestions- clear up unresolved questionsStudent-led question and answer

Table 4. Assessment Seminar Activity Chart for Inquiry-based Learning Pedagogy GTA Training Group.

CSV
Download CSV

The common two-seminar format meant that both training groups used the same peer-led learning activities. For example, GTAs of both training groups paired up to spend time practicing answering common student questions – but these took different forms based on the learning goals of the respective training group. For example, GTAs in the control training group practiced answering questions about lab protocol by answering common questions about the relevant section of the lab manual. GTAs in the inquiry pedagogy training group practiced answering questions about lab protocol by learning to design questions to help students figure out the next steps in the protocol for themselves. Extensive efforts were made to ensure the consistency and reliability of the workshops for both training groups. GTAs in both training groups were instructed not to discuss the content of their GTA training sessions with one another or with undergraduates.

iii) Measures

We assessed teaching effectiveness for the GTAs in the two training groups using three measures: (a) a 32-item, nine-factor student evaluation of educational quality (SEEQ) [3436]; (b) a 6-item cognitive learning evaluation (CLE) questionnaire; and (c) standardized mean student grade.

a) Student Evaluation of Educational Quality.

The Student Evaluation of Educational Quality (SEEQ) inventory is a nine-factor, 32-item validated survey instrument that has been widely used to evaluate undergraduate instruction at North American universities. The original SEEQ inventory [34] was written to appraise course instructors, but for this study we modified it to evaluate GTAs by simple modifications of the text (see Table 5). Nine identifiable dimensions of teaching quality can be evaluated using SEEQ: (1) learning/academic value; (2) instructor enthusiasm; (3) organization; (4) group interaction; (5) individual rapport; (6) breadth of coverage; (7) examination and grading; (8) assignments; and (9) overall instructional ability.

SEEQ FactorItemsDescriptionSample Item
(1) Learning/ Academic Value4Refers to the feeling of achievement and academic success that students obtain from participation in a course. High instructor ratings indicate that the instructor is successfully imparting useful information to students, and helping them feel that what they have learned is worthwhile and challenging.“Through my GTA, I have learned and understood the subject materials in this course”
(2) Instructor Enthusiasm4Refers to the instructor’s ability to created attentiveness and interest in the educational material on behalf of the student. High instructor ratings indicate that the instructor is creating engagement through dynamic presentation, and relating course material in a way that evokes interest.“My GTA was energized and dynamic in conducting the course”
(3) Organization4Refers to the structure and transparency of the instructor’s explanation of subject matter. High instructor ratings indicate that the instructor is relating information clearly and precisely, in a way that is easy for students to understand.“GTA explanations were clear”
(4) Group Interaction4Refers to the ability to foster academically useful social interactions within the classroom. High instructor ratings indicate that the instructor is encouraging group work in a positive way, and is motivating students to share knowledge effectively.“Students were invited to express their own ideas and/or question the GTA”
(5) Individual Rapport4Refers to the capacity to engage personally with individual learners and provide academically significant help and encouragement. High instructor ratings indicate that the instructor is able to relate to students on a personal level and provide meaningful guidance.“My GTA had a genuine interest in individual students”
(6) Breadth of Coverage4Refers to the ability to explain and compare alternative ideas, theories and techniques in a way that highlights essential features. High instructor ratings indicate that the instructor is able to relate knowledge to students effectively through contrasting specific ideas.“My GTA contrasted the implications of various theories”
(7) Examination and Grading3Refers to the ability to provide fair and useful evaluative feedback. High instructor ratings indicate that the instructor equitably assesses student work and provides meaningful correction to students.“My GTA’s feedback on examinations/graded materials was valuable”
(8) Assignments2Refers to the ability to create or use assignments to relate material to students. High instructor ratings indicate that the instructor is capable of designing or implementing assessments in such a way that new subject matter is taught or that errors are corrected.“Readings/texts/references suggested by my GTA were valuable”
(9) Overall Instructional Ability2A general evaluation of teaching effectiveness.“Overall, my GTA was a good teacher”

Table 5. Student Evaluation of Educational Quality Factor Descriptions and Example Items.

Note: sample items shown refer to the undergraduate survey; GTA survey contained the same items, but replaced “my GTA” with “I” (e.g., “Overall, I am a good teacher”). Items were rated on a five-point Likert scale (strongly disagree, disagree, neutral, agree, strongly agree) with a “NA/Don’t Know” null response included.
CSV
Download CSV

We used the first eight factors from Marsh’s original SEEQ inventory to create two SEEQ instruments (available from the original papers [28,30]). We changed Factor 9 from its original (1982) version. Marsh’s 1982 version of the SEEQ inventory did not use “overall instructional ability” but rather “course difficulty” as the ninth SEEQ factor [34,36]. GTAs do not have control over course difficulty because they do not design the course. As such, we opted to use a newer version of SEEQ developed by the University of Saskatchewan, which includes “overall instructional ability” as factor 9. After these modifications, two SEEQ instruments were created: one was used by GTAs for self-evaluation, and the other by undergraduates to rate their GTA (see Table 5 for examples)

SEEQ responses were given using a five-point Likert scale and scored as: “strongly disagree” = 1; “disagree” = 2; “neutral”= 3; “agree”= 4; and “strongly agree”= 5. When participants selected “not applicable”, then that question was given a null value and excluded from further analyses.

b) Cognitive Learning Evaluation.

The cognitive learning evaluation (CLE) instrument is a short six-factor questionnaire evaluating individual learning outcomes derived from Bloom’s revised taxonomy of learning [37]. The six factors assessed were: (1) knowledge; (2) comprehension; (3) problem-solving; (4) conceptual-analytic; (5) planning; and (6) evaluation (Table 6). Factors occur in a hierarchical order of increasing learning depth and difficulty of instruction that assumed that higher factors represent more in-depth learning. Each CLE factor was assessed by a single item. Responses were collected using a five-point Likert scale and scored as: “strongly disagree”= 1; “disagree” = 2; “neutral”= 3; “agree”= 4; and “strongly agree”= 5. When participants selected “not applicable”, then that question was given a null value and excluded from further analyses.

CLE FactorAbstract ExampleConcrete Example Item
(1) KnowledgeRecalling informationLearning exact molecular weights from the periodic table“My GTA helped me learn knowledge skills.”
(2) ComprehensionComparing or contrasting two ideasLearning to restate a word problem using equations“My GTA helped me learn comprehension skills.”
(3) Application/Problem-solvingApplying knowledge to find a solution to a specific questionLearning to select the appropriate statistical test for an analysis“My GTA helped me learn problem-solving skills.”
(4) Conceptual AnalysisDetermining causes and identifying relationshipsLearning to troubleshoot a lab protocol“My GTA helped me learn conceptual-analytic skills.”
(5) PlanningCreating a strategy by using ideas in a new wayLearning to design a new lab protocol using first principles“My GTA helped me learn planning skills.”
(6) EvaluationUsing critical reasoning to make specific judgments about ideasLearning to identify the most relevant theoretical approach to design a set of experiments“My GTA helped me learn evaluation skills.”

Table 6. Cognitive Learning Evaluation Factor Descriptions and Example Items.

Note: sample items shown refer to undergraduate survey; GTA survey contained the same items, but replaced “my GTA” with “I” and “me” with “undergraduate students” (e.g., “I helped undergraduate students learn problem-solving skills”). Items were rated on a five-point Likert scale (strongly disagree, disagree, neutral, agree, strongly agree) with a “NA/Don’t Know” null response included.
CSV
Download CSV

c) Grades.

Undergraduates self-reported their final grades for the course. Final grades reflected contributions from both lab and lecture components, although GTAs graded only lab assignments. Grading frameworks differed amongst course professors, with the lab component accounting for 25-50% of the final grade. Final exams were the largest single other contribution to final grades, ranging 30-75% of the final grade and usually included multiple-choice, objectively graded questions (i.e., by Scantron). The objective nature of the final course grades should mitigate the likelihood that GTA grading standards differed between training groups (influenced by a Hawthorne effect), or that common examinations and syllabi were not used by course professors.

Reported final grades fell on the 12-point GPA (grade point average) scale used by the university where this study took place. Using this scale, student work is graded by percentile score (0-100%), which is linked to a corresponding letter and GPA grade (e.g., 90-100% = A+ or 12.0; GPA 73-76.9% = B or 8.0 GPA; 60-62.9% = C- or 4.0 GPA; 53-56.9% = D or 2.0 GPA). Global course mean GPA values were collected from course professors, and grades were standardized by subtracting course mean GPA scores (calculated by averaging undergraduate self-reported grades per course) from individual GTA lab group GPA scores. The resulting standardized grades ranged from 3.7 to -3.1, and represented how much better (i.e., positive values) or worse (i.e., negative values) a given GTA lab group mean grade was than the average grade for any group in that course (in terms of GPA). Standardized grades were used instead of raw grades to account for the variation between courses, professors and lab material.

iv) Procedures

GTAs completed the pre-assignment questionnaires during the first week of a 13-week university semester. Training sessions were held on the third and fifth weeks of that same semester.

The SEEQ, CLE, and other demographic measures were collected from GTAs and undergraduates using a 116-item online survey hosted by Qualtrics Inc. The surveys for GTAs and undergraduate students were nearly identical, with only minor modifications relevant to the their roles. (e.g., references to “me (the GTA)” for the GTA survey were “my GTA” in the undergraduate survey). GTAs completed the surveys near the end of the semester. Approximately 7 weeks elapsed between the first GTA training session and the completion of the survey. Undergraduates were surveyed during a 45-day period beginning on the last day of exams for that semester. Data from course professors were collected during this same 45-day period.

v) Analyses

A pretest logistic regression of program distribution indicated that there were no significant differences in the ratio of Masters to PhD students between training groups (R2= .0003, χ21,1 = 0.02, p = .895). A similar pretest one-way ANOVA showed no significant difference in GTA experience (in years of teaching) between the two training groups (F(1,50) = 0.01, p= .933). Shapiro-Wilks tests indicated that all of the dependent variables were normally distributed.

For the SEEQ and CLE questionnaires, we ran multivariate analyses because the responses were composed of multiple non-independent subscales. Responses for these instruments were analyzed using a 2x2 between-subjects MANCOVA, and follow-up ANCOVA tests were performed where the multivariate response was significantly predicted by training group (similar to the analyses suggested in other studies [38,39]. For follow-up tests, Type I error was controlled for using a Bonferroni-corrected alpha value (e.g., the SEEQ inventory had 9 factors, so we used α = .0056). We ran separate MANCOVAs for GTA self-evaluation (n = 52 for both SEEQ and CLE) and for undergraduate evaluation (n = 47 for SEEQ, n = 50 for CLE). Two independent variables were included in our MANCOVAs: training group (control or inquiry) and GTA academic program (Masters or PhD), while previous GTA experience (in years of teaching) was included as a covariate. The dependent variables were all of the nine SEEQ and six CLE factors. GTA self-responses are analyzed as raw ratings. Undergraduate responses were pooled for each GTA lab section and analyzed as a mean rating for each GTA. We assessed significance of training group as a multivariate effect using the Pillai’s Trace statistic, since Pillai’s Trace is the multivariate measure which is both: (1) most robust to multivariate non-normality; and (2) conservative when the sample size is small (<100) [40].

We ran each MANCOVA without the covariate (i.e., as a MANOVA) to see if the significance levels of any multivariate or univariate effects depended on its inclusion. These analyses indicated that this was not the case, and as such, we therefore present only the MANCOVA results here.

To analyze grade data, we used a two-factor ANCOVA (n = 49) to examine standardized grade scores, again with training group (control or inquiry) and GTA academic program (Masters or PhD) as independent variables and GTA experience (in years of teaching) as a covariate. Again, the analysis was re-run without the covariate (i.e., as an ANOVA) and the overall findings were consistent with the ANCOVA results. As such, only the ANCOVA results are reported here.

All statistical analyses were performed using SPSS 20. Effect sizes are reported using partial η2 for multivariate tests and r for univariate tests. Partial η2 indicates the estimated amount of the total variation in the response due to variation in an individual predictor [41].

vi) Research Ethics

This study was examined and approved by the Carleton University Research Ethics Board (Project Number 13-0563). Written informed consent was obtained from all human subjects – both GTA participants and undergraduate survey respondents – involved in this study. Subjects were specifically informed that: (1) they were under no obligation to continue their participation in the research project; (2) that there were no known risks associated with participation; and (3) no identifying information would be collected, responses would be kept anonymous, and results would be reported only in aggregate.

Results

a) Student Evaluation of Educational Quality Results

There was not a significant effect of the covariate - GTA experience (in years) - on the undergraduate responses to the SEEQ questionnaire (see Table 3 for a list of SEEQ Factors) (Pillai’s trace = 0.18, F(9,34) = 0.81, p = .61, partial η2 = 0.18, observed power = .33). There was a significant multivariate main effect for training group, (Pillai’s trace = 0.51, F(9,34) = 3.93, p = .002, partial η2 = 0.51, observed power = .98; Table 7). However, there was neither a significant effect of GTA academic program (Pillai’s trace = 0.16, F(9,34) = 0.70, p = .71, partial η2 = 0.16, observed power = .28) nor a significant interaction between training group and GTA academic program (Pillai’s trace = 0.29, F(9,34) = 1.57, p = .16, partial η2 = 0.29, observed power = .63).

Effect / InteractionPillai's TraceFdfppartial η2power
GTA Experience (in years teaching)0.180.819, 34.6090.18.33
Training Group0.513.939, 34.002*0.51.98
GTA Academic Program0.160.709, 34.7050.16.28
Training Group × GTA Academic Program0.291.579, 34.1630.29.63

Table 7. MANCOVA test results for undergraduate Student Evaluation of Educational Quality evaluation ratings (n = 47, α = .05).

*p<.05, two-tailed.
CSV
Download CSV

For the undergraduate responses, there were significant univariate main effects for training group (using a Bonferroni-corrected α = .0056) on six of the nine SEEQ factors: learning (SEEQ Factor 1), F(1,42) = 26.47, p < .0005, r = .62; enthusiasm (SEEQ Factor 2), F(1,42) = 10.30, p < .003, r = .44; organization (SEEQ Factor 3), F(1,42) = 12.09, p < .001, r = .47; rapport (SEEQ Factor 5), F(1,42) = 14.47, p < .0005, r = .51; assignments (SEEQ Factor 7), F(1,42) = 25.79, p < .0005, r = .62; and overall instructional ability (SEEQ Factor 9), F(1,42) = 18.64, p < .0005, r = .55 (Figure 1, Table 8). A follow-up MANOVA revealed no difference in the significance level of any multivariate or univariate effects.

thumbnail
Figure 1. Mean undergraduate (A) and graduate teaching assistant (B) evaluations of GTA teaching effectiveness by Student Evaluation of Educational Quality (SEEQ) inventory factor for the “best practices” (control) and inquiry-based learning pedagogy (inquiry) training groups (with standard error bars).

Paired columns with stars are significantly different (Bonferroni-corrected significance levels: *p<0.0083, **p<0.001).

https://doi.org/10.1371/journal.pone.0078540.g001

SEEQ FactorFdfprTraining GroupMeanSE
(1) Learning 26.471, 42<.001**.62control2.830.13
inquiry3.410.12
(2) Instructor Enthusiasm 10.301, 42.003*.44control2.860.13
inquiry3.000.11
(3) Organization 12.091, 42.001*.47control2.920.12
inquiry3.040.11
(4) Group Interaction7.451, 42.009.39control2.890.14
inquiry3.570.12
(5) Individual Rapport 14.471, 42<.001**.51control3.060.15
inquiry2.910.10
(6) Breadth of Coverage 6.811, 42.013.37control2.760.13
inquiry3.280.13
(7) Examination 25.791, 42<.001**.62control3.060.16
inquiry3.100.13
(8) Assignments 5.861, 42.020.35control2.900.19
inquiry2.960.19
(9) Overall Instructional Ability18.641, 42<.001**.55control3.840.14
inquiry4.520.14

Table 8. ANCOVA test results for undergraduate Student Evaluation of Educational Quality evaluation ratings (n = 47, α = .0056).

* p<.0056, two-tailed. ** p <.0001, two-tailed.
CSV
Download CSV

There was not a significant effect of GTA experience (in years) on the GTA responses to the SEEQ questions (Pillai’s trace = 0.22, F(9,39) = 1.22, p = .31, partial η2 = 0.22, observed power = .51). There was a significant multivariate main effect for training group, (Pillai’s trace = 0.38, F(9,39) = 1.68, p = .017, partial η2 = 0.38, observed power = .89; Table 9). However, there was neither a significant main effect of GTA academic program (Pillai’s trace = 0.15, F(9,39) = 0.74, p = .67, partial η2 = 0.15, observed power = .31) nor a significant interaction between training group and GTA academic program (Pillai’s trace = 0.22, F(9,39) = 1.19, p = .33, partial η2 = 0.22, observed power = .50).

Effect / InteractionPillai's TraceFdfppartial η2power
GTA Experience (in years teaching)0.221.229, 39.3120.22.51
Training Group0.382.659, 39.017*0.38.89
GTA Academic Program0.150.749, 39.6700.15.31
Training Group × GTA Academic Program0.221.199, 39.3280.22.50

Table 9. MANCOVA test results for GTA Student Evaluation of Educational Quality ratings (n = 52, α = .05).

*p<.05, two-tailed.
CSV
Download CSV

For the GTA responses, there were significant univariate main effects for training group (using a Bonferroni-corrected α = .0056) for three of the nine SEEQ factors: learning (SEEQ Factor 1), F(1,47) = 11.86, p < .001, r = .45; group interaction (SEEQ Factor 4), F(1,47) = 11.86, p < .001, r = .48; and overall instructional ability (SEEQ Factor 9), F(1,47) = 15.34, p < .0005, r = .50 (Figure 1, Table 10). A follow-up MANOVA revealed no difference in the significance level of any multivariate or univariate effects.

SEEQ FactorFdfprTraining GroupMeanSE
(1) Learning 11.861, 47.001*.45control2.420.14
inquiry3.470.11
(2) Instructor Enthusiasm 1.981, 47.17.20control2.610.21
inquiry3.440.13
(3) Organization 1.201, 47.280.16control2.750.19
inquiry3.610.12
(4) Group Interaction14.041, 47<.001**.48control3.090.14
inquiry3.600.12
(5) Individual Rapport 0.431, 47.52.10control2.940.18
inquiry3.830.10
(6) Breadth of Coverage 7.191, 47.01.36control2.920.16
inquiry3.520.13
(7) Examination 0.121, 47.73.05control2.540.16
inquiry3.520.10
(8) Assignments 0.011, 47.91.01control3.210.17
inquiry3.940.17
(9) Overall Instructional Ability15.341, 47<.001**.50control2.630.19
inquiry3.590.09

Table 10. ANCOVA test results for GTA Student Evaluation of Educational Quality ratings (n = 52, α = .0056).

* p< .0056, two-tailed. ** p < .0001, two-tailed.
CSV
Download CSV

b) Cognitive Learning Evaluation Results

There was not a significant effect of GTA experience (in years) on the undergraduate responses to the CLE questions (see Table 4 for a list of CLE Factors) (Pillai’s trace = 0.08, F(6,40) = 0.60, p = .73, partial η2 = 0.08, observed power = .21). There was a significant multivariate main effect for training group, (Pillai’s trace = 0.41, F(6,40) = 4.71, p < .001, partial η2 = 0.41, observed power = .98; Table 11). However, there was not a significant effect for GTA academic program (Pillai’s trace = 0.03, F(6,40) = 0.17, p = .98, partial η2 = 0.03, observed power = .09) or a significant interaction between training group and program (Pillai’s trace = 0.03, F(6,40) = 0.22, p = .97, partial η2 = 0.03, observed power = .10).

Effect / InteractionPillai's TraceFdfppartial η2power
GTA Experience (in years teaching)0.080.606, 40.7320.08.21
Training Group0.414.716, 40.001*0.41.98
GTA Academic Program0.030.176, 40.9840.03.09
Training Group × GTA Academic Program0.030.226, 40.9680.03.10

Table 11. MANCOVA test results for undergraduate Cognitive Learning Evaluation ratings (n = 50, α =.05).

*p<.05, two-tailed.
CSV
Download CSV

For the undergraduate responses, there was a significant univariate main effect for training group (using a Bonferroni-corrected α = .0083) for four of the six CLE factors, including: comprehension skills (CLE Factor 2), F(1,45) = 9.68, p < .003, r = .42; problem-solving/application skills (CLE Factor 3), F(1,45) = 9.76, p < .003, r = .42; synthesis skills (CLE Factor 5), F(1,45) = 11.22, p < .001, r = .48; and evaluation skills (CLE Factor 6), F(1,45) = 26.12, p < .0005, r = .61 (Figure 2, Table 12). A follow-up MANOVA revealed no difference in the significance level of any multivariate or univariate effects.

thumbnail
Figure 2. Mean undergraduate (A) and graduate teaching assistant (B) evaluations of GTA teaching effectiveness by Cognitive Learning Evaluation (CLE) questionnaire factor for the “best practices” (control) and inquiry-based learning pedagogy (inquiry) training groups (with standard error bars).

Paired columns with stars are significantly different (Bonferroni-corrected significance levels: *p<0.0083, **p<0.001).

https://doi.org/10.1371/journal.pone.0078540.g002

CLE FactorFdfprTraining GroupMeanSE
(1) Knowledge5.371, 45.025.33control3.040.16
inquiry3.540.11
(2) Comprehension9.681, 45<.001**.42control2.880.18
inquiry3.610.12
(3) Application / Problem-solving9.761, 45<.001**.42control2.770.14
inquiry3.470.13
(4) Analysis7.681, 45.008*.38control2.950.17
inquiry3.520.10
(5) Synthesis 13.821, 45<.001**.48control2.850.23
inquiry3.820.12
(6) Evaluation 26.121, 45<.001**.61control2.630.17
inquiry3.720.12

Table 12. ANCOVA test results for undergraduate Cognitive Learning Evaluation ratings (n = 50, α =.0083).

* p < .0083, two-tailed. ** p < .0001, two-tailed.
CSV
Download CSV

There was not a significant effect of GTA experience (in years), on the GTA responses to the CLE questions (Pillai’s trace = 0.14, F(6,42) = 1.14, p = .36, partial η2 = 0.14, observed power = .40; Table 13). There was not a significant multivariate main effect for training group, (Pillai’s trace = 0.19, F(6,42) = 1.68, p = .15, partial η2 = 0.19, observed power = .57) or GTA academic program (Pillai’s trace = 0.10, F(6,42) = 0.79, p = .58, partial η2 = 0.10, observed power = .28). In addition, there was not a significant interaction between training group and program (Pillai’s trace = 0.09, F(6,42) = 0.69, p = .66, partial η2 = 0.09, observed power = .24). Given these results, univariate tests were not performed (Figure 2, Table 14). A follow-up MANOVA revealed no difference in the significance level of any multivariate or univariate effects.

Effect / InteractionPillai's TraceFdfppartial η2power
GTA Experience (in years teaching)0.141.146, 42.3590.14.40
Training Group0.191.686, 42.1490.19.57
GTA Academic Program0.100.796, 42.5830.10.28
Training Group × GTA Academic Program0.090.696, 42.6600.09.24

Table 13. MANCOVA test results for GTA Cognitive Learning Evaluation ratings (n = 52, α = .05).

*p<.05, two-tailed.
CSV
Download CSV
CLE FactorFdfprTraining GroupMeanSE
(1) Knowledge0.751, 51.39.12control3.570.14
inquiry3.490.10
(2) Comprehension0.151, 51.71.05control3.270.78
inquiry3.400.42
(3) Application / Problem-solving1.071, 51.31.14control3.200.14
inquiry3.420.12
(4) Analysis5.981, 51.02.32control3.130.14
inquiry3.540.10
(5) Synthesis 4.411, 51.04.28control2.950.22
inquiry3.460.11
(6) Evaluation 6.451, 51.01.34control2.760.27
inquiry3.520.11

Table 14. ANCOVA test results for undergraduate Cognitive Learning Evaluation ratings (n = 52, α =.0083).

*p < .0083, two-tailed. ** p < .0001, two-tailed.
CSV
Download CSV

c) Grades

There was a significant univariate main effect of training group on undergraduate grades in their biology courses, F(1,44) = 26.31, p < .022, r = .34, but not for program, F(1,44) = 1.66, p = .56, r = .19 ,or an interaction between training group and GTA academic program, F(1,44) = 0.53, p =.47, r = .11 (Figure 3, Table 15).

thumbnail
Figure 3. Mean standardized grade differences (in 12-point CGPA) for the “best practices” (control) and inquiry-based learning pedagogy (inquiry) training groups (with standard error bars).

https://doi.org/10.1371/journal.pone.0078540.g003

Fdfpr
GTA Experience (in years teaching)0.721,44.4070.11
Training Group0.901,44.3470.45
GTA Academic Program12.291,44.001*0.12
Training Group × GTA Academic Program0.051,44.8180.03

Table 15. ANCOVA test results for standardized student grades (n = 49).

*p<.05, two-tailed.
CSV
Download CSV

Discussion

These results suggest that offering inquiry-oriented pedagogy GTA training can yield gains in academic performance in undergraduate science labs and improve GTA teaching (as evaluated by both undergraduates and GTAs), particularly with respect to scientific reasoning. Inquiry pedagogy training group GTAs outperformed control training group GTAs, both with respect to SEEQ/CLE questionnaires and to self-reported undergraduate grades, and there were no items on either survey in which control training group GTAs received significantly higher ratings than inquiry pedagogy training group GTAs (this was true for both undergraduate evaluations and GTA self-evaluations).

From the student- and self-evaluation data, GTAs who completed the inquiry pedagogy training received higher ratings than those in the control training group in two main areas: enthusiasm and interpersonal skills, and organization and academic value. This first result is fairly straightforward – the inquiry pedagogy training group was rated significantly higher than control training group GTAs for enthusiasm (SEEQ factor 2) and individual rapport (SEEQ factor 5), both of which are important skills in an interactive role such as a lab facilitator. The second result was that inquiry pedagogy training group GTAs received higher student ratings for organization (SEEQ factor 3) and assignments (SEEQ factor 7), both areas of the lab in which the GTA has only partial control, suggesting improvement in general teaching ability and planning skill. More specifically, undergraduates rated GTAs who completed the inquiry pedagogy training higher than those who completed the control training regimen for lower-order cognitive skills (e.g., SEEQ factor 1, CLE factors 2 and 3), higher-order cognitive skills (CLE factors 5 and 6) and overall measures of teaching ability (SEEQ factor 9). Taken together, these results indicate that GTAs who completed the inquiry pedagogy training were rated as better organized, provided better feedback on assignments, and were better overall teachers of both higher- and lower-order skills than the control training group GTAs.

We note that the differences between mean Likert scores for the two treatments are small in some cases. Likert scores, which have a limited range given that only five responses are possible (i.e. a score of 1-5), required consistently lower or higher values to register as significantly differing from the value of the other GTA training group. Since our analysis used a fairly conservative Bonferroni correction to decrease the likelihood of Type I errors, small but statistically significant differences between training groups for a specific SEEQ or CLE factor likely indicate improvements that are real but small.

Our results indicated that self-reported grade data confirmed the advantage of inquiry pedagogy training, with the lab groups of GTAs who completed the inquiry-based learning pedagogy training reporting significantly higher standardized grades than GTAs who completed the control training regimen. These results suggest that undergraduate students in the lab sections of experimental group GTAs had higher overall course grades than did students in the lab sections of control training group GTAs. Because course grades include both a lab and a lecture component, higher overall grades may not be due entirely to influence of lab GTA, although it may be that effective lab teaching (from the GTA) may have additional in-lecture benefits that we did not quantify. Both training groups outperformed the average grade of the course, perhaps because some GTAs in the course did not have any training whatsoever. The finding that the grades of both GTA groups were higher than the course mean indicates that there might be some benefit to the control “best practice” training. The best practice training focused on practical skills and did use elements deemed successful in other studies (5). This type of GTA training might improve the academic value of lab activities for undergraduate students, but that GTAs, who are not expert teachers, may not be able to identify how or why.

We conclude that the recorded differences in (undergraduate- and GTA-rated) teaching ability between inquiry pedagogy training group GTAs and control training group GTAs are due to the fact that inquiry pedagogy GTA training improves GTAs’ ability to facilitate guided-inquiry activities related to scientific reasoning. While many methods can be used to teach undergraduates content, structured-inquiry activities (i.e. labs) are specifically designed to ask students to use higher-order cognitive skills as they employ the scientific method to solve problems. GTAs are evidently better able to teach undergraduates to “reason like scientists” after inquiry-based learning pedagogy training. We note that students rated inquiry pedagogy training group GTAs as better teachers of both the particular SEEQ factors related to inquiry facilitation (including academic value, enthusiasm, group interaction and rapport) as well as higher-order cognitive skills (i.e. higher CLE factors). We suggest that inquiry pedagogy training, which encourages GTAs to use active feedback as the lab activity is performed, has enabled inquiry GTAs to better teach using guided inquiry.

There are several limitations related to this study. Firstly, our sample size was small (52 GTAs), and the trial had only two training groups. Second, although there was a highly significant predictive relationship between training group and mean standardized grade for GTA lab section, lab grades were only part of the total grade; it is therefore unclear to what degree this treatment might predict improvement in student grades. Third, we note that we did not directly assess GTA teaching skills, and instead used student and GTA evaluations—along with self-reported student grades—to estimate actual GTA teaching ability. We also assume that the students who participated in our survey were representative of the larger undergraduate population. Further study, on a larger scale, of inquiry-based learning pedagogy training for science lab GTAs is our main recommendation for further studies. Although we believe our conclusions are supported by the data collected, a large, hierarchically-designed intervention experiment, involving a greater number of replicate training groups, would provide a more powerful test of our hypothesis. Further study of the reason of the differences between undergraduate and GTA questionnaire responses with respect to individual SEEQ and CLE factors might also be valuable; it is unclear why undergraduates seem better able to identify differences between training groups than the GTAs themselves.

Conclusion

In conclusion, our data indicate that providing GTAs with a theoretical understanding of the guided-inquiry methods that underpin lab activities increases the quality of GTA teaching in introductory science labs, especially in areas related to scientific reasoning. Across all measures, inquiry-based learning pedagogical training for GTAs seems to improve teaching, even when training is limited to only five hours. We believe that our data indicate that there is value in including information on inquiry-oriented teaching methodologies in future GTA training regimens, especially where those GTAs teach inquiry-oriented activities such as teaching labs.

Acknowledgments

The authors would like to thank the course professors, lab coordinators, graduate teaching assistants and undergraduate survey respondents that made this research possible.

Author Contributions

Conceived and designed the experiments: PWH MRE. Performed the experiments: PWH. Analyzed the data: PWH MRE. Contributed reagents/materials/analysis tools: PWH. Wrote the manuscript: PWH MRE.

References

  1. 1. Park C (2004) The graduate teaching assistant (GTA): lessons from North American experience. Teach Higher Educ 9: 349–361. Available: http://www.tandfonline.com/doi/abs/10.1080/1356251042000216660. Accessed 11 February 2013. doi:https://doi.org/10.1080/1356251042000216660.
  2. 2. Meyers S, Prieto L (2000) Training in the Teaching of Psychology: what is done and examining the differences. Teach Psychol 27: 258–261. Available: http://www.tandfonline.com/doi/full/10.1207/S15328023TOP2704_03. Accessed 2013 April 8. doi:https://doi.org/10.1207/S15328023TOP2704_03.
  3. 3. Prieto LR, Altmaier EM (1994) The relationship of prior training and previous teaching experience to self-efficacy among graduate teaching assistants. Research in Higher Education 35: 481–497. Available: http://link.springer.com/10.1007/BF02496384. Accessed 2013 April 2. doi:https://doi.org/10.1007/BF02496384.
  4. 4. Shannon D, Twale D, Moore M (1998) TA Teaching Effectiveness: the impact of training and teaching experience. J Higher Educ 69: 440–466. Available: http://www.jstor.org/stable/10.2307/2649274. Accessed 2013 March 3. doi:https://doi.org/10.2307/2649274.
  5. 5. Luft JA, Kurdziel JP, Roehrig GH, Turner J (2004) Growing a garden without water: Graduate teaching assistants in introductory science laboratories at a doctoral/research university. J Res Sci Teach 41: 211–233. Available: http://doi.wiley.com/10.1002/tea.20004. Accessed 2013 February 15. doi:https://doi.org/10.1002/tea.20004.
  6. 6. Rodriques RAB, Bond-Robinson J (2006) Comparing Faculty and Student Perspectives of Graduate Teaching Assistants’ Teaching. J Chem Educ 83: 305. Available: http://pubs.acs.org/doi/abs/10.1021/ed083p305. Accessed 2013 March 6. doi:https://doi.org/10.1021/ed083p305.
  7. 7. Kendall KD, Schussler EE (2012) Does instructor type matter? Undergraduate student perception of graduate teaching assistants and professors. CBE Life. Sci Educ 11: 187–199. Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3366904&tool=pmcentrez&rendertype=abstract. Accessed 2013 March 7.
  8. 8. Koenig KM, Endorf RJ, Braun GA (2007) Effectiveness of different tutorial recitation teaching methods and its implications for TA training. Phys Rev St Per 3: 10104. Available: http://link.aps.org/doi/10.1103/PhysRevSTPER.3.010104. Accessed 2013 February 27.
  9. 9. Hardré PL, Burris AO (2010) What contributes to teaching assistant development: differential responses to key design features. Instr Sci 40: 93–118. Available: http://www.springerlink.com/index/10.1007/s11251-010-9163-0. Accessed 2013 February 15.
  10. 10. Rushin J, Saix J-D, Lumsden A, Streubel D, Summers G et al. (1997) Graduate Teaching Assistant Training: A Basis for Improvement of College Biology Teaching & Faculty Development? Am Biol Teach 59: 86–90. Available: http://www.jstor.org/stable/10.2307/4450255. Accessed 2013 February 15. doi:https://doi.org/10.2307/4450255.
  11. 11. Rissing S, Cogan J (2009) Can an Inquiry Approach Improve College Student Learning in a Teaching Laboratory? CBE Life. Sci Educ 8: 55–61. Available: http://w.lifescied.org/content/8/1/55.short. Accessed 2013 February 16.
  12. 12. Edelson DC, Gordin DN, Pea RD (1999) Addressing the challenges of inquiry-based learning through technology and curriculum design. J Learn Sci 8: 391–450. doi:https://doi.org/10.1080/10508406.1999.9672075.
  13. 13. Windschitl M (2003) Inquiry projects in science teacher education: What can investigative experiences reveal about teacher thinking and eventual classroom practice? Sci Educ 87: 112–143. Available: http://doi.wiley.com/10.1002/sce.10044. Accessed 2013 February 10. doi:https://doi.org/10.1002/sce.10044.
  14. 14. Banchi H, Bell R (2008) The many levels of inquiry. Sci Child 46: 26–29.
  15. 15. Henige K (2011) Undergraduate student attitudes and perceptions toward low- and high-level inquiry exercise physiology teaching laboratory experiences. Adv Physiol Educ 35: 197–205. Available: http://www.ncbi.nlm.nih.gov/pubmed/21652505. Accessed 2013 March 4. doi:https://doi.org/10.1152/advan.00086.2010. PubMed: 21652505.
  16. 16. Farrell JJ, Moog RS, Spencer JN (1999) A Guided-Inquiry General Chemistry Course. J Chem Educ 76: 570. Available: http://pubs.acs.org/doi/abs/10.1021/ed076p570. Accessed 2013 April 2. doi:https://doi.org/10.1021/ed076p570.
  17. 17. Schoffstall AM, Gaddis BA (2007) Incorporating Guided-Inquiry Learning into the Organic Chemistry Laboratory. J Chem Educ 84: 848. Available: http://pubs.acs.org/doi/abs/10.1021/ed084p848. Accessed 2013 April 2. doi:https://doi.org/10.1021/ed084p848.
  18. 18. Iler H, Justice D (2012) Discovering 13C NMR, 1H NMR, and IR Spectroscopy in the General Chemistry Laboratory through a Sequence of Guided-Inquiry Exercises. J Chem Educ 89: 1178–1182. Available: http://pubs.acs.org/doi/abs/10.1021/ed2005664. Accessed 2013 March 7. doi:https://doi.org/10.1021/ed2005664.
  19. 19. Bohrer K, Stegenga BD, Ferrier A (2007) Training and Mentoring TAs in Inquiry-Based Methods. Tested Studies for Laboratory Teaching. Proceedings of the 28th Workshop/Conference of the Association for Biology Laboratory Education 28. pp. 335–346.
  20. 20. Roehrig GH, Turner JA, Kurdziel JP, Luft JA (2003) Graduate Teaching Assistants and Inquiry-Based Instruction: Implications for Graduate Teaching Assistant Training. J Chem Educ 80: 1206. Available: http://pubs.acs.org/doi/abs/10.1021/ed080p1206. Accessed 2013 April 2. doi:https://doi.org/10.1021/ed080p1206.
  21. 21. Apedoe X, Walker S, Reeves T (2006) Integrating inquiry-based learning into undergraduate geology. J Geosci Educ 54: 414–421. Available: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.117.1751&rep=rep1&type=pdf. Accessed 2013 March 4.
  22. 22. Weaver GC, Russell CB, Wink DJ (2008) Inquiry-based and research-based laboratory pedagogies in undergraduate science. Nat Chem Biol 4: 577–580. Available: http://www.ncbi.nlm.nih.gov/pubmed/18800041. Accessed 2013 March 4. doi:https://doi.org/10.1038/nchembio1008-577. PubMed: 18800041.
  23. 23. Pickering M (1980) Are lab courses a waste of time? Chronlicle of Higher Education 45: 80.
  24. 24. Vale RD, DeRisi J, Phillips R (2012) Graduate education. Interdisciplinary graduate training in teaching labs. Science 338: 1542–1543. Available: http://authors.library.caltech.edu/36473/. Accessed 2013 February 27. doi:https://doi.org/10.1126/science.1216570. PubMed: 23258877.
  25. 25. Marbach-Ad G, Schaefer KL, Kumi BC, Friedman la Thompson KV et al. (2012) Development and Evaluation of a Prep Course for Chemistry Graduate Teaching Assistants at a Research University. J Chem Educ 89: 865–872. Available: http://pubs.acs.org/doi/abs/10.1021/ed200563b. Accessed 2013 April 2. doi:https://doi.org/10.1021/ed200563b.
  26. 26. Lowman J, Mathie VA (1993) What Should Graduate Teaching Assistants know about Teaching? Teach Psychol 20: 84–88. doi:https://doi.org/10.1207/s15328023top2002_4.
  27. 27. Wood WB (2009) Innovations in teaching undergraduate biology and why we need them. Annu Rev Cell Dev Biol 25: 93–112. Available: http://www.ncbi.nlm.nih.gov/pubmed/19575638. Accessed 2013 March 4. doi:https://doi.org/10.1146/annurev.cellbio.24.110707.175306. PubMed: 19575638.
  28. 28. Vale RD, DeRisi J, Phillips R, Mullins RD, Waterman C et al. (2012) Graduate education. Interdisciplinary graduate training in teaching labs. Science 338: 1542–1543. Available: http://www.ncbi.nlm.nih.gov/pubmed/23258877. Accessed 2013 February 14. doi:https://doi.org/10.1126/science.1216570. PubMed: 23258877.
  29. 29. Luckie DB, Maleszewski JJ, Loznak SD, Krha M (2004) Infusion of collaborative inquiry throughout a biology curriculum increases student learning: a four-year study of "Teams and Streams" Adv Physiol Educ 28: 199–209. Available: . http://www.ncbi.nlm.nih.gov/pubmed/15545349.. doi:https://doi.org/10.1152/advan.00025.2004. PubMed: 15545349.
  30. 30. French D, Russell C (2002) Do Graduate Teaching Assistants Benefit from Teaching Inquiry-Based Laboratories? BioScience 52: 1036. Available: http://www.jstor.org/stable/1314349.
  31. 31. Hanauer DI, Jacobs-Sera D, Pedulla ML, Cresawn SG, Hendrix RW et al. (2006) Inquiry learning. Teaching scientific inquiry. Science 314: 1880–1881. doi:https://doi.org/10.1126/science.1136796. PubMed: 17185586.
  32. 32. Goldey ES, Abercrombie CL, Ivy TM, Kusher DI, Moeller JF, Rayner DA, Smith CF, Spivey NW (2012) Biological inquiry: a new course and assessment plan in response to the call to transform undergraduate biology. CBE Life. Sci Educ 11: 353–363. Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3516791&tool=pmcentrez&rendertype=abstract. Accessed 2013 March 4.
  33. 33. Wilke RR, Straits WJ (2005) Practical Advice for Teaching Inquiry-Based Science Process Skills in the Biological Sciences. Am Biol Teach 67: 534–540. Available: http://www.bioone.org/doi/abs/10.1662/0002-7685(2005)067[0534:PAFTIS]2.0.CO;2. Accessed 2013 April 2. doi:https://doi.org/10.1662/0002-7685(2005)067[0534:PAFTIS]2.0.CO;2.
  34. 34. Marsh HW (1982) SEEQ: a reliable, valid and useful instrument for collecting students’ evaluations of university teaching. Br J Educ Psychol 52: 77–95. doi:https://doi.org/10.1111/j.2044-8279.1982.tb02505.x.
  35. 35. Coffey M, Gibbs G (2001) The Evaluation of the Student Evaluation of Educational Quality Questionnaire (SEEQ) in UK Higher Education. Assess Eval HIGH Educ 26: 37–41. Available: http://www.tandfonline.com/doi/abs/10.1080/02602930020022318. Accessed 2013 February 16.
  36. 36. Marsh H, Bailey M (1993) Multidimensional Students’ Evaluations of Teaching Effectiveness: A Profile Analysis. J Higher Educ 64: 1–18. Available: http://www.jstor.org/stable/10.2307/2959975. Accessed 2013 February 16. doi:https://doi.org/10.2307/2959975.
  37. 37. Krathwohl DR (2002) A Revision of Bloom’s Taxonomy: An Overview. Theory into Practice 41: 212–218. Available: http://www.tandfonline.com/doi/abs/10.1207/s15430421tip4104_2. Accessed 2013 April 3. doi:https://doi.org/10.1207/s15430421tip4104_2.
  38. 38. Camodeca M, Goossens FA (2005) Children’s opinions on effective. Educ Res 47: 93–105. doi:https://doi.org/10.1080/0013188042000337587.
  39. 39. Meyers LS, Gamst GC, Guarino AJ (2005) Applied Multivariate Research: Design and Interpretation 6th Edition. London: Sage Publications. 1104 pp.
  40. 40. Olson CL (1976) On choosing a test statistic in multivariate analysis of variance. Psychol Bull 83: 579–586. doi:https://doi.org/10.1037/0033-2909.83.4.579.
  41. 41. Levine TR, Hullett CR (2002) Eta squared, partial eta squared, and misreporting of effect size in communication research. Hum Commun Res 28: 612–625. doi:https://doi.org/10.1111/j.1468-2958.2002.tb00828.x.