Skip to main content

Home/ Educational Analytics/ Group items tagged improvement

Rss Feed Group items tagged

George Bradford

Measuring Teacher Effectiveness - DataQualityCampaign.Org - 0 views

  •  
    Measuring Teacher Effectiveness Significant State Data Capacity is Required to Measure and Improve Teacher Effectiveness  States Increasingly Focus on Improving Teacher Effectiveness: There is significant activity at the local, state, and federal levels to  measure and improve teacher effectiveness, with an unprecedented focus on the use of student achievement as a primary indicator of  effectiveness. > 23 states require that teacher evaluations include evidence of student learning in the form of student growth and/or value-added data (NCTQ, 2011). > 17 states and DC have adopted legislation or regulations that specifically require student achievement and/or student growth to "significantly" inform or be the primary criterion in teacher evaluations(NCTQ, 2011).  States Need Significant Data Capacity to Do This Work: These policy changes have significant data implications. > The linchpin of all these efforts is that states must reliably link students and teachers in ways that capture the complex connections that  exist in schools. > If such data is to be used for high stakes decisions-such as hiring, firing, and tenure-it must be accepted as valid, reliable, and fair. > Teacher effectiveness data can be leveraged to target professional development, inform staffing assignments, tailor classroom instruction,  reflect on practice, support research, and otherwise support teachers.  Federal Policies Are Accelerating State and Local Efforts: Federal policies increasingly support states' efforts to use student  achievement data to measure teacher effectiveness. > Various competitive grant funds, including the Race to the Top grants and the Teacher Incentive Fund, require states to implement teacher  and principal evaluation systems that take student data into account.  > States applying for NCLB waivers, including the 11 that submitted requests in November 2011, must commit to implementing teacher and  principal evaluation and support systems. > P
George Bradford

Networked Improvement Communities: Bryk lectures Bristol 2014 | Learning Emergence - 0 views

  •  
    "'Making Systems Work - whether in healthcare, education, climate change, or making a pathway out of poverty - is the great task of our generation as a whole' and at the heart of making systems work is the problem of complexity.  Prof Tony Bryk, President of the Carnegie Foundation for the Advancement of Teaching,  spent a week with people from the Learning Emergence network, leading a Master Class for practitioners, delivering two public lectures and participating in a consultation on Learning Analytics Hubs in Networked Improvement Communities  (background).  A key idea is that in order to engage in quality improvement in any system, we need to be able to 'see the system as a whole' and not just step in and meddle with one part of it."
George Bradford

Learning process analytics - EduTech Wiki - 1 views

  •  
    "Introduction In this discussion paper, we define learning process analytics as a collection of methods that allow teachers and learners to understand what is going on in a' 'learning scenario, i.e. what participants work(ed) on, how they interact(ed), what they produced(ed), what tools they use(ed), in which physical and virtual location, etc. Learning analytics is most often aimed at generating predictive models of general student behavior. So-called academic analytics even aims to improve the system. We are trying to find a solution to a somewhat different problem. In this paper we will focus on improving project-oriented learner-centered designs, i.e. a family of educational designs that include any or some of knowledge-building, writing-to-learn, project-based learning, inquiry learning, problem-based learning and so forth. We will first provide a short literature review of learning process analytics and related frameworks that can help improve the quality of educational scenarios. We will then describe a few project-oriented educational scenarios that are implemented in various programs at the University of Geneva. These examples illustrate the kind of learning scenarios we have in mind and help define the different types of analytics both learners and teachers need. Finally, we present a provisional list of analytics desiderata divided into "wanted tomorrow" and "nice to have in the future"."
George Bradford

SpringerLink - Abstract - Dr. Fox Rocks: Using Data-mining Techniques to Examine Studen... - 0 views

  •  
    Abstract Few traditions in higher education evoke more controversy, ambivalence, criticism, and, at the same time, support than student evaluation of instruction (SEI). Ostensibly, results from these end-of-course survey instruments serve two main functions: they provide instructors with formative input for improving their teaching, and they serve as the basis for summative profiles of professors' effectiveness through the eyes of their students. In the academy, instructor evaluations also can play out in the high-stakes environments of tenure, promotion, and merit salary increases, making this information particularly important to the professional lives of faculty members. At the research level, the volume of the literature for student ratings impresses even the most casual observer with well over 2,000 studies referenced in the Education Resources Information Center (ERIC) alone (Centra, 2003) and an untold number of additional studies published in educational, psychological, psychometric, and discipline-related journals. There have been numerous attempts at summarizing this work (Algozzine et al., 2004; Gump, 2007; Marsh & Roche, 1997; Pounder, 2007; Wachtel, 1998). Student ratings gained such notoriety that in November 1997 the American Psychologist devoted an entire issue to the topic (Greenwald, 1997). The issue included student ratings articles focusing on stability and reliability, validity, dimensionality, usefulness for improving teaching and learning, and sensitivity to biasing factors, such as the Dr. Fox phenomenon that describes eliciting high student ratings with strategies that reflect little or no relationship to effective teaching practice (Ware & Williams, 1975; Williams & Ware, 1976, 1977).
George Bradford

Learning Analytics + NICs for Systemic Educational Improvement | Learning Emergence - 0 views

  •  
    "Personal reflections on 2 workshops and a lecture with Tony Bryk (Carnegie Foundation for the Advancement of Teaching), hosted last week by Ruth Deakin Crick at University of Bristol. What follows after a brief introduction to the concept of NICs, are my thoughts on the intersection of NICs with Learning Analytics. I made a number of connection points between the features of the DEED+NIC approach, and learning analytics, which I'll highlight in green."
George Bradford

Assessment | University of Wisconsin-Madison - 0 views

  •  
    "Using Assessment for Academic Program Improvement Revised April 2009 "
George Bradford

QUT | Learning and Teaching Unit | REFRAME - 0 views

  •  
    REFRAME REFRAME is a university-wide project reconceptualising QUT's evaluation of learning and teaching. REFRAME is fundamentally reconsidering QUT's overall approach to evaluating learning and teaching. Our aim is to develop a sophisticated risk-based system to gather, analyse and respond to data along with a broader set of user-centered resources. The objective is to provide individuals and teams with the tools, support and reporting they need to meaningfully reflect upon, review and improve teaching, student learning and the curriculum. The approach will be informed by feedback from the university community, practices in other institutions and the literature, and will, as far as possible, be 'future-proofed' through awareness of emergent evaluation trends and tools. Central to REFRAME is the consideration of the purpose of evaluation and the features that a future approach should consider.
George Bradford

LOCO-Analyst - 0 views

  •  
    What is LOCO-Analyst? LOCO-Analyst is an educational tool aimed at providing teachers with feedback on the relevant aspects of the learning process taking place in a web-based learning environment, and thus helps them improve the content and the structure of their web-based courses. LOCO-Analyst aims at providing teachers with feedback regarding: *  all kinds of activities their students performed and/or took part in during the learning process, *  the usage and the comprehensibility of the learning content they had prepared and deployed in the LCMS, *  contextualized social interactions among students (i.e., social networking) in the virtual learning environment. This Web site provides some basic information about LOCO-Analyst, its functionalities and implementation. In addition, you can watch videos illustrating the tool's functionalities. You can also learn about the LOCO (Learning Object Context Ontologies) ontological framework that lies beneath the LOCO-Analyst tool and download the ontologies of this framework.
George Bradford

National Institute for Learning Outcomes Assessment - 0 views

  •  
    "Accrediting associations have expectations that call on institutions to collect and use evidence of student learning outcomes at the programmatic and institutional to confirm and improve student learning.  This section of the NILOA website lists both regional accrediting associations and specialized or programmatic accrediting organizations along with links to those groups."
George Bradford

re:Work - Guides - 0 views

  •  
    "Guides Practices, research, and ideas to improve your people processes."
George Bradford

Submissions for the Awards for Excellence in Learning Analytics « ascilite - 0 views

  •  
    "Awards for Excellence in Learning Analytics: Submissions The LA SIG operates an Awards program recognising excellence in the practical application of LA to enhance learning and teaching. A key driver for the Awards program is to create and share resources about effective LA practices. We want to give a voice to all who are working with LA to improve learning and teaching - whatever the scale of their endeavours.  To this end, all presentations from Award applicants are available below for viewing. (Submissions closed on 11 September).  We believe these presentations will form an important resource library around the use of LA in tertiary education across Australasia and New Zealand. You may view the awards submission criteria and other background information on the awards in the SIG Awards Program Information (PDF) and to find out more about LA-SIG activities, go here."
George Bradford

Assessment and Analytics in Institutional Transformation (EDUCAUSE Review) | EDUCAUSE - 0 views

  • At the University of Maryland, Baltimore County (UMBC), we believe that process is an important factor in creating cultural change. We thus approach transformational initiatives by using the same scholarly rigor that we expect of any researcher. This involves (1) reviewing the literature and prior work in the area, (2) identifying critical factors and variables, (3) collecting data associated with these critical factors, (4) using rigorous statistical analysis and modeling of the question and factors, (5) developing hypotheses to influence the critical factors, and (6) collecting data based on the changes and assessing the results.
  • among predominantly white higher education institutions in the United States, UMBC has become the leading producer of African-American bachelor’s degree recipients who go on to earn Ph.D.’s in STEM fields. The program has been recognized by the National Science Foundation and the National Academies as a national model.
  • UMBC has recently begun a major effort focused on the success of transfer students in STEM majors. This effort, with pilot funding from the Bill and Melinda Gates Foundation, will look at how universities can partner with community colleges to prepare their graduates to successfully complete a bachelor’s degree in a STEM field.
  • ...5 more annotations...
  • Too often, IT organizations try to help by providing an analytics “dashboard” designed by a vendor that doesn’t know the institution. As a result, the dashboard indicators don’t focus on those key factors most needed at the institution and quickly become window-dressing.
  • IT organizations can support assessment by showing how data in separate systems can become very useful when captured and correlated. For example, UMBC has spent considerable effort to develop a reporting system based on our learning management system (LMS) data. This effort, led from within the IT organization, has helped the institution find new insights into the way faculty and students are using the LMS and has helped us improve the services we offer. We are now working to integrate this data into our institutional data warehouse and are leveraging access to important demographic data to better assess student risk factors and develop interventions.
  • the purpose of learning analytics is “to observe and understand learning behaviors in order to enable appropriate interventions.
  • the 1st International Conference on Learning Analytics and Knowledge (LAK) was held in Banff, Alberta, Canada, in early 2011 (https://tekri.athabascau.ca/analytics/)
  • At UMBC, we are using analytics and assessment to shine a light on students’ performance and behavior and to support teaching effectiveness. What has made the use of analytics and assessment particularly effective on our campus has been the insistence that all groups—faculty, staff, and students—take ownership of the challenge involving student performance and persistence.
  •  
    Assessment and analytics, supported by information technology, can change institutional culture and drive the transformation in student retention, graduation, and success. U.S. higher education has an extraordinary record of accomplishment in preparing students for leadership, in serving as a wellspring of research and creative endeavor, and in providing public service. Despite this success, colleges and universities are facing an unprecedented set of challenges. To maintain the country's global preeminence, those of us in higher education are being called on to expand the number of students we educate, increase the proportion of students in science, technology, engineering, and mathematics (STEM), and address the pervasive and long-standing underrepresentation of minorities who earn college degrees-all at a time when budgets are being reduced and questions about institutional efficiency and effectiveness are being raised.
George Bradford

Seeking Evidence of Impact: Opportunities and Needs (EDUCAUSE Review) | EDUCAUSE - 0 views

  • Conversations with CIOs and other senior IT administrators reveal a keen interest in the results of evaluation in teaching and learning to guide fiscal, policy, and strategic decision-making. Yet those same conversations reveal that this need is not being met.
  • gain a wider and shared understanding of “evidence” and “impact” in teaching and learning
  • establish a community of practice
  • ...11 more annotations...
  • provide professional-development opportunities
  • explore successful institutional and political contexts
  • establish evidence-based practice
  • The most important reason is that in the absence of data, anecdote can become the primary basis for decision-making. Rarely does that work out very well.
  • autocatalytic evaluation process—one that builds its own synergy.
  • We live by three principles: uncollected data cannot be analyzed; the numbers are helped by a brief and coherent summary; and good graphs beat tables every time.
  • Reports and testimonies from faculty and students (57%) Measures of student and faculty satisfaction (50%) Measures of student mastery (learning outcomes) (41%) Changes in faculty teaching practice (35%) Measures of student and faculty engagement (32%)
  • The survey results also indicate a need for support in undertaking impact-evaluation projects.
  • Knowing where to begin to measure the impact of technology-based innovations in teaching and learning Knowing which measurement and evaluation techniques are most appropriate Knowing the most effective way to analyze evidence 
  • The challenge of persuasion is what ELI has been calling the last mile problem. There are two interrelated components to this issue: (1) influencing faculty members to improve instructional practices at the course level, and (2) providing evidence to help inform key strategic decisions at the institutional level.
  • Broadly summarized, our results reveal a disparity between the keen interest in research-based evaluation and the level of resources that are dedicated to it—prompting a grass-roots effort to support this work.
  •  
    The SEI program is working with the teaching and learning community to gather evidence of the impact of instructional innovations and current practices and to help evaluate the results. The calls for more accountability in higher education, the shrinking budgets that often force larger class sizes, and the pressures to increase degree-completion rates are all raising the stakes for colleges and universities today, especially with respect to the instructional enterprise. As resources shrink, teaching and learning is becoming the key point of accountability. The evaluation of instructional practice would thus seem to be an obvious response to such pressures, with institutions implementing systematic programs of evaluation in teaching and learning, especially of instructional innovations.
George Bradford

[!!!] Penetrating the Fog: Analytics in Learning and Education (EDUCAUSE Review) | EDUC... - 0 views

  • Continued growth in the amount of data creates an environment in which new or novel approaches are required to understand the patterns of value that exist within the data.
  • learning analytics is the measurement, collection, analysis and reporting of data about learners and their contexts, for purposes of understanding and optimising learning and the environments in which it occurs.
  • Academic analytics, in contrast, is the application of business intelligence in education and emphasizes analytics at institutional, regional, and international levels.
  • ...14 more annotations...
  • Course-level:
  • Educational data-mining
  • Intelligent curriculum
  • Adaptive content
  • the University of Maryland, Baltimore County (UMBC) Check My Activity tool, allows learners to “compare their own activity . . . against an anonymous summary of their course peers.
  • Mobile devices
  • social media monitoring tools (e.g., Radian6)
  • Analytics in education must be transformative, altering existing teaching, learning, and assessment processes, academic work, and administration.
    • George Bradford
       
      See Bradford - Brief vision of the semantic web as being used to support future learning: http://heybradfords.com/moonlight/research-resources/SemWeb_EducatorsVision 
    • George Bradford
       
      See Peter Goodyear's work on the Ecology of Sustainable e-Learning in Education.
  • How “real time” should analytics be in classroom settings?
  • Adaptive learning
  • EDUCAUSE Review, vol. 46, no. 5 (September/October 2011)
  • Penetrating the Fog: Analytics in Learning and Education
  •  
    Attempts to imagine the future of education often emphasize new technologies-ubiquitous computing devices, flexible classroom designs, and innovative visual displays. But the most dramatic factor shaping the future of higher education is something that we can't actually touch or see: big data and analytics. Basing decisions on data and evidence seems stunningly obvious, and indeed, research indicates that data-driven decision-making improves organizational output and productivity.1 For many leaders in higher education, however, experience and "gut instinct" have a stronger pull.
George Bradford

College Degrees, Designed by the Numbers - Technology - The Chronicle of Higher Education - 0 views

  • Arizona State's retention rate rose to 84 percent from 77 percent in recent years, a change that the provost credits largely to eAdvisor.
  • Mr. Lange and his colleagues had found that by the eighth day of class, they could predict, with 70-percent accuracy, whether a student would score a C or better. Mr. Lange built a system, rolled out in 2009, that sent professors frequently updated alerts about how well each student was predicted to do, based on course performance and online behavior.
  • Rio Salado knows from its database that students who hand in late assignments and don't log in frequently often fail or withdraw from a course. So the software is more likely to throw up a red flag for current students with those characteristics.
  • ...5 more annotations...
  • And in a cautionary tale about technical glitches, the college began sharing grade predictions with students last summer, hoping to encourage those lagging behind to step up, but had to shut the alerts down in the spring. Course revisions had skewed the calculations, and some predictions were found to be inaccurate. An internal analysis found no increase in the number of students dropping classes. An improved system is promised for the fall.
  • His software borrows a page from Netflix. It melds each student's transcript with thousands of past students' grades and standardized-test scores to make suggestions. When students log into the online portal, they see 10 "Course Suggestions for You," ranked on a five-star scale. For, say, a health-and-human-performance major, kinesiology might get five stars, as the next class needed for her major. Physics might also top the list, to satisfy a science requirement in the core curriculum.
  • Behind those recommendations is a complex algorithm, but the basics are simple enough. Degree requirements figure in the calculations. So do classes that can be used in many programs, like freshman writing. And the software bumps up courses for which a student might have a talent, by mining their records—grades, high-school grade-point average, ACT scores—and those of others who walked this path before.
  • The software sifts through a database of hundreds of thousands of grades other students have received. It analyzes the historical data to figure out how much weight to assign each piece of the health major's own academic record in forecasting how she will do in a particular course. Success in math is strongly predictive of success in physics, for example. So if her transcript and ACT score indicate a history of doing well in math, physics would probably be recommended over biology, though both satisfy the same core science requirement.
  • Every year, students in Tennessee lose their state scholarships because they fall a hair short of the GPA cutoff, Mr. Denley says, a financial swing that "massively changes their likelihood of graduating."
  •  
    July 18, 2012 College Degrees, Designed by the Numbers By Marc Parry Illustration by Randy Lyhus for The Chronicle Campuses are places of intuition and serendipity: A professor senses confusion on a student's face and repeats his point; a student majors in psychology after a roommate takes a course; two freshmen meet on the quad and eventually become husband and wife. Now imagine hard data substituting for happenstance. As Katye Allisone, a freshman at Arizona State University, hunkers down in a computer lab for an 8:35 a.m. math class, the Web-based course watches her back. Answers, scores, pace, click paths-it hoovers up information, like Google. But rather than personalizing search results, data shape Ms. Allisone's class according to her understanding of the material.
1 - 15 of 15
Showing 20 items per page