Wednesday, 9 January 2013

The ILA (results)


Results
The responses to the three questionnaires from Kuhlthau’s SLIM toolkit were varied and occasionally unexpected. A sample of 10 students was taken from the original 24. Four students were unable to complete all 3 questionnaires due to absence. The sample was chosen on the basis of all three questionnaires being completed fully. Every second student was chosen on an alphabetical basis.  All completed questionnaires were, however checked for consistency of themes and content. To preserve anonymity, students are referred to as A, B, C etc.

Figure 1: Question 1 Number of facts reported by each student for each questionnaire

The students had received their task sheets and completed one lesson on the topic before answering the first questionnaire. During this initial lesson they had completed an online survey which was developed to help anyone determine their “ecological footprint”. This exercise had given them some idea of the topic and task requirements before they completed Questionnaire 1.

Their responses were qualitatively graded as facts, explanations or conclusions as explained in the SLIM toolkit. This question required students to report what they know about their topic with the expectation being that the number of facts would increase with subsequent questionnaires, as the students started researching their topic. Of the 10 sample students examined, this trend was noticed to a certain extent especially with students A, E, F, G and I. See Figure 1 below.

Figure 2: The proportion of facts, explanations and conclusions for the entire cohort

When taking all the students results into account across all three questionnaires we see that “Facts” represent the greatest proportion of the three types of responses.

A reason for the greater number of facts, rather than explanations and conclusions being reported in Questionnaires 1 and 2 could be that students were not encouraged to write their replies in full sentences (as recommended in the SLIM toolkit), and they did not receive sufficient time to complete the questionnaire. During Questionnaire 3 many of those statements were written in response to Questions 6. I have corrected my results taking those statements in account later in this report.
The changes noted between the questionnaires for the number of facts, explanations and conclusions reported for the sample of 10 students can be seen in the following graph.

Figure 3: The number of facts, explanations and conclusions for the student sample across all three questionnaires

 An increase in the amount of facts was reported in Questionnaire 2 followed by a slight decrease for Questionnaire 3.  The student answers for Questionnaire 1 commonly centred on a generalised conception of their ecological footprint.  The same question was asked in Questionnaires 2 and 3, with answers becoming progressively more specific to the students’ chosen research questions. My observation was that although the actual number of facts decreased from Questionnaire 2 to 3; the quality, detail and specificity of each student’s answers improved with each subsequent questionnaire. This is not taken into account in the graphical representations but can be observed in the following table.

 Table 1: The increase in quality and depth of statements in response to Question 1

For some students the number of facts supplied in Questionnaire 3 decreased.  This may be explained by students feeling reluctant to repeat information supplied in previous questionnaires.  In many cases it was also noted that students understood Question 1 and Question 6 of Questionnaire 3 to be referring to the same thing. Many of the statements that should have appeared in Question 1 of the last questionnaire actually appeared in Question 6; skewing the results and making it appear that fewer facts were reported in Questionnaire 3 than actually were.  The following table illustrates excerpts from student statements showing how answers to Question 6 would have been better placed as answers for Question 1. The reason for this was they were statements about the topic rather than statements reflecting the information literacy experience gained whilst undertaking an inquiry task.

Table 2: Examples of Question 6 Questionnaire 3 responses which would have been better suited under Question 1

Table 2 shows us that the results for fact, explanations and conclusions would have changed significantly had Question 6’s statements been taken into account where appropriate.    As explained under “Recommendations”; when repeating an inquiry unit of this type or administering the “SLIM” questionnaires again I would be very careful to explain the distinction between comments on knowledge of the topic vs comments on information literacy gained during the inquiry task.
If Figure 3 is corrected using Question 6’s replies where appropriate, the following graph is obtained.

Figure 4: Answers to Question 1 corrected using answers to Question 6 of Questionnaire 3 where appropriate

Here we see a steady yet significant increase in the number of facts reported during the course of the inquiry task. The number of explanations decreases fairly dramatically in Questionnaire 2 and then increases again in the last questionnaire, almost reaching the initial questionnaire’s level. Bloom’s revised taxonomy describes conclusions as a higher order thinking task comparable to evaluating. As an inquiry task progresses you would expect more of these higher order statements. The number of conclusions drops to zero in Questionnaire 2 and then doubles for the last questionnaire. A reason for the decrease and then increase in explanations and conclusions would most likely be that they students were in the “exploration phase” of Kuhlthau’s ISP.  This is described as the most difficult stage of the ISP with students becoming frustrated and discouraged. It is at this stage that students most require assistance from their instructional team of teachers and librarian. (Kuhlthau 2007) The noticeable increase in explanations and conclusions in Questionnaire 3 demonstrates that these students were able to benefit significantly from intervention strategies provided by the instructional team. These interventions consisted of; advice on search strategies when using the internet, how to determine whether a site is reliable or biased, how to reference and keep notes for a bibliography, using Word for a bibliography, how to do a “Survey Monkey” to get primary data and how to use proforma sentence structure devise to verbally analyse graphical data. During conversations with students it was also possible for me to assist with misunderstandings of words; for instance, the meanings and implications of “primary and secondary data”. Many of these interventions consisted of a one-on-one scaffolding as suggested by Vygotsky in his “Zone of Proximal Development" (Kuhlthau, 2007)



Figure 5: Vygotsky's Zone of Proximal Development.   Photo credit:

The students’ self reported “perceived” knowledge (Question  3) uses a qualitative scoring rubric of 0= Nothing   1= Not much  2= Quite a bit and  3= A great deal.  The question is “How much do you know about this topic?” When results are compared across the three questionnaires we see that in the first questionnaire only 2 students answered “nothing” and the majority of students answered “not much”.  No student felt confident enough at this stage to answer “A great deal”. This is to be expected when students have recently been introduced to a new task. 


 Figure 6: Perceived knowledge Questionnaire 1

When the same question was asked a few weeks later no students answered that they did not know anything, but at this stage of the task no student was confident enough in their knowledge to answer “A great deal” either.


Figure 7: Perceived knowledge Questionnaire 2

 By the end of the task when students had given their presentations, significantly more of them were confident enough in their knowledge to answer “A great deal” and only two answered “Not much”, with no student answering “Nothing”.



Figure 8: Perceived knowledge Questionnaire 3

Using the 10 student sample and making a comparison between answers to questions 1, 2 and 3 will demonstrate the correlation between perceived knowledge, interest and illustrated knowledge.  Illustrated knowledge was corrected proportionally so that all three parameters could be compared using a scale from 0 to 3 where 0=nothing and 3=a great deal. Therefore, “Perceived knowledge” as measured by Question 3: “How much do you know about this topic?”; “Interest” as measured by Question 2: “How interested are you in this topic?” and “Illustrated knowledge” as measured by counting the number of factual, explanatory and conclusive statements made for Question 1: “Write down what you know about your topic” were compared.  A strong correlation between these three parameters for the majority of students is demonstrated, especially for Questionnaire 3.  The steady increase in the length of the bars in the graph below demonstrates that the amount measured from questionnaire to questionnaire increases and is noticeably greater for questionnaire 3. This illustrates that these students, on the whole, are reasonably aware of their own strengths and weaknesses and this correlation lends credibility to their self reported insights.


Figure 9: Correlation between perceived knowledge, interest and illustrated knowledge

The same information was then manipulated to show each student’s  progression. It can be seen that most students experienced a marked improvement between Questionnaires 1 and 3. The graph also shows that many students either stayed the same from Questionnaires 1 to 2 or may even have declined slightly. This may seem counterintuitive, with the reader expecting the results to show a steady improvement in all categories as the task progresses. However the decline in interest can be explained by Kuhlthau’s stages of the ISP [provide a link here], as the second questionnaire would have been administered when most students were experiencing the “exploration” stage of their task. Common emotions felt during this stage would be confusion, frustration and doubt; this shows up as a decline in interest in the topic. Only students E and F experienced an improvement in interest in their topic from Questionnaires 1 to 2.


Figure 9: Correlation between perceived knowledge, interest and illustrated knowledge

Question 4 served the purpose of finding out what students find easy when they do research. I categorised the information literacy standards achieved by the students as:
·           Develop research question/s
·           Access information
·           Determine accuracy
·           Organisation skills
·           Understanding and applying information
·           Communication skills
·           Improvement strategies
·           Acknowledging sources
·           Technology skills


The categories  correlate to those of the Standards for the 21st century learner (American Association of School Librarians).

Most students were able to develop appropriate questions and search strategies leading them to source information on the topics that interested them the most.  After the second questionnaire students were much more comfortable locating relevant information. This can be attributed intervention strategies and training from their teacher and librarian to assist students with the development of focus questions, expert search strategies and evaluating the credibility of their sources.
The following excerpts are of typical student responses for tasks they found easy when doing research;
·         “Finding valuable information from reliable websites was quite easy because Public transit websites have valuable information and so did the state government.”  Student A
·         “I find it easy to get the information in my head and understand it. I find it easier to get information on the internet and newspapers. I find it easy when I get direct quotes from a person to use for evidence”. Student C
·         “Thinking back to my research project, I found finding credible book resources easy through using the school library and the BCC library, I also found finding internet resources easy as well as organising the research. Working out if information is credible.” Student E.
·         “I found it easy to pick out the information that was relevant and what was not which helped me a great deal when writing my AVD [annotated visual display]. I also found it easy to find the correct information when I researched more specific terms” Student I




Figure 11: Tasks deemed as "easy to do" when conducting research 

By the end of the inquiry task many students reported that selecting and narrowing down heaps of information for their AVD was difficult. The literacy skills skills of organising and understanding content well enough to summarise it and phrase it succinctly were needed to create the AVD which was a A3 sheet of paper with both information and illustrations. Typical comments were:
 Student A.  “I found putting the information into words and finding the right pictures quite difficult.
Student E.  “...picking out appropriate information from websites, creating the annotated visual display as well as working out which questions I should find research for”
Student I . “...difficult to know what exactly to put on my AVD and to find enough visuals so that my AVD was 60% visual and 40% text. This was due to me having too much into” Student “finding specific information about my topic”
Others found it difficult to come up with focus questions:
Student C.  “It was difficult to come up with the questions to start the assignment”.
Student B.  “To pick a topic”   
Student E.  “Working out which questions I should find research for”
Many students reported some of the information literacy skills both easy and difficult. This contradiction can be understood when examining excerpts such as the following:
Student A.  “If I’m interested, I will find reading and memorising some key points really easy”. This remark was categorised under “Understanding and applying information”, as it entails deriving meaning from information. The same student wrote “I find note taking really hard.” This was categorised as “difficulty with accessing information as it entailed difficulties with choosing the correct information from a wide array of information. It was also categorised under “Understanding and applying information”, as it entailed figuring out what was appropriate for the topic.
Student B made comments demonstrating self-insight and organisational skills under both Questions 4 and 5. For Question 4 she reported, “For me, I find it easy to take notes on paper because I can easily get distracted on the computer”. For Question 5 she put, “I find it hard to concentrate on the computer”.


Figure 12: Tasks deemed as difficult to do when conducting research

In the second and third questionnaire a question appeared asking students about their feelings regarding their research. The question was not asked in the first questionnaire.  Half way through the inquiry task only three student of the sample of 10 felt confident about their research and knew where they were heading. The rest of the students reported either feeling overwhelmed or confused. These sentiments were reflected by the cohort as whole although very few students reported feeling “frustrated”. This response is to be expected at this stage of an inquiry task. Once students have started researching they often find the enormous amount information available confusing and struggle to determine what is relevant, appropriate or true. It is at this stage that intervention from the instructional team is often required. Once they have started to master the topic and have presented their findings there is often an improvement in their feelings about the topic. This can be seen in the following graph comparing student feeling at the mid-way point to the end point of their inquiry task. 


Figure 13: Students' feelings about their research mid-way through the task as compared to the end of the task


As can be seen from the above graph all except one student either improved in affect or stayed the same. Only Student C remained unhappy about her work.

In Question 6 of Questionnaire 3 the students are asked “What did you learn in doing this research project?”
Although many of the students assumed this to be a repetition of Question 1 “Write down what you know about this topic” there were many others that reported gains in self-insight such as:
Student A.  “I learnt that I need to take good notes from information and then turn it into good, well-structures paragraphs. I also  learnt that time management is vital”.
Student C.  “I have to learn to manage my time before I start the project. I learnt that I need to complete all work when I can and as soon as I can. I learnt I need to prioritise”
Student D “researching for valid information is not extremely easy”
Student E.  “To start research early. Making sure to create a bibliography as I went. To not trust all websites. That some information is just opinion. Focusing on the assignment. Time management.”
Student F.  “ I need to work on not doing this at the last minute. I learnt that writing more notes and studying about my topic helps me present my topic better and makes it seem that I know a lot of stuff on litter”
Student H.  “ ...if I focus and concentrate hard I will complete a task, despite the due date. I also learnt that you have to have very good time management for a big research task like this as well as any other assignment or task”  Student I.  “I learnt to be organised and to make sure I keep up to date with my research and journal.”
Student J.  “I learnt about myself as a researcher that I am good at finding lots of general information, however sometimes I have trouble finding specific information about my topic. I also learnt that some websites are biased or have incorrect information”


Figure 14: Self reported skills attained by the inquiry task


As can be seen from the graph, responses that rated highly for this question were “Understanding and applying information” as well as “improvement strategies”; these were mentioned by over half of the students.  Organisational skills were also mentioned by a third of the students. These were identified as problem areas for many students in the previous 2 questionnaires. This shows that tutorials presented by the instructional team paid off and that students will be better equipped to excel at this type of inquiry task in future.

References

Kuhlthau, C., Maniotes, L., & Caspari, A. (2007). Guided Inquiry: Learning in the 21st Century. Westport, Connecticut: Libraries Unlimited


Todd, R. J., Kuhlthau, C. C., & Heinstrom, J. E. (2005). School library impact measure (SLIM): A toolkit and handbook for tracking and assessing student learning outcomes of guided inquiry through the school library. Center for International Scholarship in School Libraries, Rutgers University.












No comments:

Post a Comment