Promising Topics for Dissertations and Grant-Funded Research

 

Handbook and Other Materials l Asking the Right Questions (ARQ) l Training, Consulting, & External EvaluationFAQ

Return to Flashlight Evaluation Handbook Table of Contents

If one of the following topics interests you, feel free to use it!  Obviously, we'd appreciate knowing about your interest. If you'd like to know if someone else at a subscribing institution is doing something like this already, we'll help you find out.  For information, contact online@tltgroup.orgLater, if useful research tools or findings emerge from your work, perhaps we can make them available to our subscribers or in our public pages. 

This list of ideas for grant proposals and dissertations runs from "micro" level issues about teaching and learning to "macro" questions of institutional structure and national policy (at the bottom of the list).

  • Study strategies for using technology to make large courses more participatory, active, and engaging for diverse learners. When people think about transforming higher education, they are often rejecting an image of large lecture halls half-full of dosing students (and the other half not showing up at all). The National Center for Academic Transformation has made a good start in supporting redesign of such courses. It's time for a good study to create a taxonomy of such redesign strategies. What are the strengths of each? How do they relate to each other? What problems often occur when each such strategy is implemented?
  • Develop TLT case studies to help faculty learn to handle the rough issues that technology can raise.  When academic staff begin using technology to improve teaching and learning in their courses, they can easily experience frustration or even serious difficulty. Some of these problems are solvable (e.g., bring a spare copy of your slides, in case your first version is unexpectedly unavailable); know the help number to dial. But many such problems are not so easy to solve, especially when they arise from the teaching/learning process itself. What can go wrong when courses shift the balance away from lecture and recitation, plug-and-chug, toward active learning, collaborative learning, and creative work? 
    When using a TLT case study, faculty in a discussion group begin by reading a  "trigger case" which tells the story of a faculty member and how he or she came to face a common problem that can arise when instructors begin using technology to a. First individually and then as a group, the participating staff analyze the problem and discuss different ways to address it.
    Developing such TLT case studies is a) pragmatically useful for faculty development, and b) a distinctive approach to the scholarship of teaching and learning. (See this TLT case study, for example).
    Take a look at these first, brief TLT case studies. Try them out. We're sure you can do better, and we'll add  your cases to this Collection!
  • Are common uses of word processing helping colleges educate graduates who are more adept at developing and critiquing complex arguments?  The answer might be "yes" because (we hypothesize) rethinking one's work while rewriting can help students master a process of developing a complex, internally consistent argument. But, so far as we know, no one has done the study even though the data is out there, waiting to be analyzed. Click here to see more. 
  • Digital writing raises many interesting questions for study. For example, it's been claimed that, when students know that they can continue to develop a piece of writing after it's been submitted for a grade, a) at least some students will continue to do so and will benefit from that continuation, b) at least some students (more than the first group) will see the initial assignment as "real" rather than "schoolwork" and will therefore be more motivated to put thought and energy into the writing. This conjecture is closely linked to a conjecture about audience: that students will put more energy into writing when they know that an audience that means something to them will be seeing and reacting to the work. True?
  • How much "collaborative learning" is there in a typical course? How much "active learning?" How much faculty-student contact?  These abstract 'umbrella' terms cover a variety of student and faculty activities; they are also some of the more common ways in which faculty use technology to improve teaching and learning.  The Flashlight Current Student Inventory, Faculty Inventory, and Flashlight Online provide survey tools for estimating how much of this kind of activity goes on, and how helpful technology is for increasing or enriching it.  But these tools can be improved and augmented by other ways of observing teaching-learning activity.  Also lacking (so far) is any kind of norm to which individual courses, departments or institutions might be compared. These are big projects and perhaps worthy of a grant proposal.  You could tackle it on your own or partner with us...
  • Diagnostic tools for faculty.  This is a priority for Flashlight but you might take on part of it, alone or with us. Here's an example: items for gathering student feedback to faculty use of PowerPoint (working with Mount Royal College, we've already developed that)   
    What's next? The example about which we have talked the most is online collaboration among students. We've been talking with faculty for some time about barriers that can hinder at least a few students; the list is now over 60 barriers. In other words, in a class of 30 students, some may find no problems collaborating online, but others may face one or more of a rather large set of problems. That's the bad news. The other part of the bad news is that, without help, the faculty member can only see that participation is disappointingly low, but not why.   The good news is that, if the right person knew what problem each student was facing, almost all the problems are pretty easy to fix. Diagnostic tools: a survey and test for the institution to administer regularly, plus another survey or two for the faculty member to administer a week or two into the course.  These kinds of data might result in much better collaboration. That's the kind of tool we're talking about.  
  • Study how faculty learn to use technology in their teaching. For example, imagine a set of faculty who know relatively little about facilitating collaboration among students (as in a seminar or in team work on problems) and relatively little about using a system such as WebCT or Blackboard for that purpose.  What are some of the more efficient ways to help them learn? How does the availability of "learning objects" (e.g., from Merlot) affect faculty progress? For more on the nature of this problem, see the June 2002 issue of F-LIGHT.
  • Study competences faculty need to teach well online. The idea, in brief, is to identify two groups of faculty - one set of faculty that (by some measure) teaches superlatively and a set of other faculty who do not teach superlatively.  Compare how they each think about what they do? Are all the superlative teachers alike? how about the controls? If there are differences in how they think, do those differences lie in trainable skills? in deeply held values and assumptions?  For more on how such a study might be designed, click here.
  • Develop a measure of faculty approaches to teaching. The Schneider, Klemp, and Kastendiek (SKK) study described in the link above suggests there may be a dimension of approaches to teaching, characterized at one extreme by faculty (widely seen as excellent teachers) who believe all their students can become engaged, that students are different from one another, and that it's important to use continual trial and error to see what it takes to get each student engaged. At the other end of the scale are faculty (widely seen as normal teachers) who believe that only some students will become engaged, and that their job is to teach (explain) while it is the students' task to learn (study). "My job is to put it out there, their job is to learn it," a member of this normal group might say.
    The SKK model was derived from a small sample of faculty in a specialized type of program in 1980-81.  The research challenge is to create a way of measuring faculty approaches to teaching (survey? observation?) to see whether a) the model does cleanly describe today how different faculty approach teaching, and how many faculty fall into each category (if there are indeed just two categories). (See also the item immediately below on this page).                
  • Is there a relationship between how faculty use technology and what they believe about "assessment"If the "Balancing Act" measure described above can be developed, I predict this measure will be useful for predicting faculty uses of technology and faculty attitudes toward assessment.  (Click here to read more.)        
  • Do faculty approaches to teaching (as described above) change from 'normal' to 'excellent'? Are there techniques that universities can use to help more faculty become 'excellent'?  The SKK study above suggested that faculty widely regarded as 'excellent' believe that all students can become engaged, and try to help each student attain that, while 'normal' faculty believe that many students will never become engaged and do not focus on helping students as individuals. Over the years, how often do faculty shift from one belief set to the other? Do any university practices, or collegial practices, make it more likely that faculty will over time shift from a 'normal' approach to an 'excellent' approach?
  • (NEW!)  Under what circumstances have institutions teamed up to create and offer stable, successful, shared courses of study?  I'm talking about two or more institutions that each contribute some of the teaching in order to offer a set of courses for which any of their students can register. For example, five institutions collaborate to offer the OneMBA Executive Degree Program in Global Management. Tiny physics departments at a number of public institutions in Texas collaborate to offer a physics major open to all their students. The advantages of such programs are compelling, but the barriers are daunting.  Under what circumstances do such programs thrive? What does that suggest about future policies, technology platforms, and practices?  Click here for a longer description of this topic. 
  • Is your institution improving? evolving?  Some people suggest that the basic character of higher education is being transformed (e.g., "Access and/or Quality" essay; "vision and mission" of National Learning Infrastructure Initiative).  How might such transformational change be measured? This table summarizes a variety of research and evaluative questions - the top row contains questions raised by improvement, the second row by transformation, and the third row by efforts to guide or control that transformation.
  • Validating survey and interview estimates of activities.  In order to estimate the educational value of technology investments, we have to ask people what they've done with technology (e.g., how much time did faculty spend using e-mail to discuss assignments with students). We have to ask similar questions to estimate changes in educational costs resulting from technology investments (e.g., how much time did faculty spend using e-mail to discuss assignments with students). But are the responses to such survey or interview questions accurate enough to make policy? Are there better ways to frame the survey or interview questions? Are there other kinds of inquiry that could complement or replace the survey or interview? For a bit more on this topic, see this article in F-LIGHT.
  • Estimating the costs of current and alternative ways of supporting faculty, staff and student use of technology. Most institutions face some kind of support service crisis: their budgets for support staff fall far short of need.  Relatively little research has been done conceptualizing different methods of support and then estimating the monetary and human costs (and rewards) of each. Costs are so local in nature that it is probably more appropriate to do these studies within institution in order to guide local policies and practices. In the long term, however, new approaches to providing support on a large scale may emerge. For a bit more on this topic, see this article in F-LIGHT.
  • What kinds of virtual programs have the best chance in the future of gaining financial support from their alumni? This essay suggests that the building of current relationships around academic work (and its affective side) is key, and suggests a study that attends to teaching/learning practices and the details of the virtual workplace itself.
  • Research on assessment's role in the use of technology to help transform academic programs This essay sketches ways in which assessment (feedback) could be used to help guide and accelerate institutional improvements and outcomes, especially where technology is used to provide leverage. Part II of the paper describes some research questions whose answers could help institutions around the world make better use of such assessment.
  • What are some of the design issues and policy questions raised by infrastructure for integrated access?  "Virtual universities" that don't teach courses but do offer courses taught by others, along with advising.  Programs that bank credits, or offer credit by examination. Companies that proctor exams offered by online degree programs. These are just a few examples of infrastructure that mediates between online instructors/institutions and online learners. What are we learning about designs that do, and don't work?  What policy questions are raised about subsidies? regulation?  Click here to read more background information.

- Stephen C. Ehrmann, Ph.D., Director, The Flashlight Program

Return to Flashlight Evaluation Handbook Table of Contents

 

PO Box 5643
Takoma Park, Maryland 20913
Phone
: 301.270.8312/Fax: 301.270.8110  

To talk about our work
or our organization
contact:  Sally Gilbert

Search TLT Group.org

Contact us | Partners | TLTRs | FridayLive! | Consulting | 7 Principles | LTAs | TLT-SWG | Archives | Site Map |