Evaluation design and metrics of success were determined using stakeholder involvement approaches; outcome indicators are assessed by triangulating primary and secondary quantitative and qualitative data sources either identified, or created by the collaborative members and evaluators. A quasi-experimental evaluation design is used to compare intervention group outcomes to comparison group outcomes; the comparison group is comprised of residents from a Louisville neighborhood containing comparable levels of crime, and other stress-inducing conditions
Trinidad Jackson 2016 AEA poster.pdf
The presentation from Jeremy Paley and Aimee White about creating a international shared measurement system for the Bill and Melinda Gates Foundation #sharedindicators #collaborativedesign #Non-ProfitsandFoundationsEvaluation #Collaborative,ParticipatoryandEmpowermentEval ...
Final CIMS AEA presentation Oct 2013.pdf
Different disciplines informed the study design, including ethnography, developmental psychology, educational research, and data analysis methods grounded in statistical theory. The results of this evaluation were highly useful in refining the program design, metrics, and quality of implementation to support positive student outcomes
AEA slides_BCCI_FINAL (AEA Library Post).pptx
Topics of this demonstration include how the evaluation team: collaborated with local partners on evaluation design; trained participants on methods they deemed useful; incorporated participatory methods from start (co-creating consent forms) to finish (reporting and use); and, used interactive methods, including story-circles and storyboards, to allow participants to interpret data and document their own recommendations
Eval2015-Reciprocity in Research.pptx
Slides from Session 1929 of the 2014 AEA Conference Panel presented by Tania Jarosewich, Kirk Knestis, Jeni Corn, and Rita O'Sullivan Adding "Value" to Evaluation in Education Settings: Opportunities for Evolving Roles of Evaluators in an Education Research and Development Paradigm ...
Adding Value to Evaluation in Education Settings.pdf
The mission of the Practice Research Network is to build an infrastructure linking investigators at The University of Maryland, Baltimore, Department of Psychiatry with the public mental health system through an innovative approach to nurturing the development of activities that reflect the...
5 attachments
Focus Search - An Evaluation of an Innovative, Collaborative Approach to Interfacing Research Systems with the Mental Health Community Deborah Piez, MS, LPC Dan Nieberding, MS, LCPC Sandra Sundeen, MS, RN CREATING THE PRACTICE RESEARCH NETWORK THE IDEA THE PLAYERS THE FUNDING – NIMH Research Infrastructure Support Project (IP - RISP) # 1R24MH082755 CREATING THE PRACTICE RESEARCH NETWORK STAFFING NUMBER QUALIFICATIONS QUALITIES INITIATION Goal: To develop collaborative relationships with research teams and the mental health communities PROVIDING: ACCESS TO RESEARCH EDUCATION AND INFORMATION RECIPROCITY CREATING A NEW PARADIGM FOR RECRUITMENT OF STUDY PARTICIPANTS Merging PRN innovative services into an existing operational design of study team approach Focusing on local mental health authorities as a gateway to groups of potential participants Integrating advocacy organizations Accomplishments So Far Introductions and partnerships with 24 local mental health authorities and the State mental health authority Collaborations developed with family, consumer, and other advocacy groups plus over 65 providers 11 studies represented by the PRN 587 PRN generated study referrals Behavioral Health Research web site created for Network of Care EVALUATING THE PRACTICE RESEARCH NETWORK Rationale for the evaluation Method selected Populations surveyed: Consumers/Family Members/Advocates Local Mental Health Authorities/Service Providers Investigators/Study Teams SURVEY DESIGN Three versions of the survey created Providers/Mental Health Authorities: 9 items Investigators/Study Teams: 9 items Consumers/Family Members/Advocates: 10 items Convenience Sample Providers/Mental Health Authorities: Survey Monkey Investigators/Study Teams: Survey Monkey Consumers/Family Members/Advocates: Survey Monkey or hard copy SURVEY RESPONDENTS Consumers/Family Members, Advocates: 120 Mental Health Authorities/Providers: 32 Research Investigators/Study Teams: 4 Total Responses: 156 Correlations How much has your organization benefited from the Practice_Research Network?
Presenter(s): Gina Svarovsky, Science Museum of Minnesota, gsvarovsky@smm.org Amy Grack Nelson, Science Museum of Minnesota, agnelson@smm.org Abstract: Evaluation projects involving evaluators distributed across the country can have several benefits, but also present a range of challenges....
2 attachments
Focus Search - Descriptors help others to locate your report and provide a tool to find similar reports by audience, venue, subject, study design and methods. 4.